Search results for: modifiable areal unit problem (MAUP)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9238

Search results for: modifiable areal unit problem (MAUP)

6838 Wobbled Laser Beam Welding for Macro-to Micro-Fabrication Process

Authors: Farzad Vakili-Farahani, Joern Lungershausen, Kilian Wasmer

Abstract:

Wobbled laser beam welding, fast oscillations of a tiny laser beam within a designed path (weld geometry) during the laser pulse illumination, opens new possibilities to improve the marco-to micro-manufacturing process. The present work introduces the wobbled laser beam welding as a robust welding strategy for improving macro-to micro-fabrication process, e.g., the laser processing for gap-bridging and packaging industry. The typical requisites and relevant equipment for the development of a wobbled laser processing unit are addressed, including a suitable laser source, light delivery system, optics, proper beam deflection system and the design geometry. In addition, experiments have been carried out on titanium plate to compare the results of wobbled laser welding with conventional pulsed laser welding. As compared to the pulsed laser welding, the wobbled laser welding offers a much greater fusion area (i.e. additional molten material) while minimizing the HAZ and provides a better confinement of the material microstructural changes.

Keywords: wobbled laser beam welding, wobbling function, beam oscillation, micro welding

Procedia PDF Downloads 328
6837 An Empirical Study of the Effect of Robot Programming Education on the Computational Thinking of Young Children: The Role of Flowcharts

Authors: Wei Sun, Yan Dong

Abstract:

There is an increasing interest in introducing computational thinking at an early age. Computational thinking, like mathematical thinking, engineering thinking, and scientific thinking, is a kind of analytical thinking. Learning computational thinking skills is not only to improve technological literacy, but also allows learners to equip with practicable skills such as problem-solving skills. As people realize the importance of computational thinking, the field of educational technology faces a problem: how to choose appropriate tools and activities to help students develop computational thinking skills. Robots are gradually becoming a popular teaching tool, as robots provide a tangible way for young children to access to technology, and controlling a robot through programming offers them opportunities to engage in developing computational thinking. This study explores whether the introduction of flowcharts into the robotics programming courses can help children convert natural language into a programming language more easily, and then to better cultivate their computational thinking skills. An experimental study was adopted with a sample of children ages six to seven (N = 16) participated, and a one-meter-tall humanoid robot was used as the teaching tool. Results show that children can master basic programming concepts through robotic courses. Children's computational thinking has been significantly improved. Besides, results suggest that flowcharts do have an impact on young children’s computational thinking skills development, but it only has a significant effect on the "sequencing" and "correspondence" skills. Overall, the study demonstrates that the humanoid robot and flowcharts have qualities that foster young children to learn programming and develop computational thinking skills.

Keywords: robotics, computational thinking, programming, young children, flow chart

Procedia PDF Downloads 146
6836 X-Ray Crystallographic, Hirshfeld Surface Analysis and Docking Study of Phthalyl Sulfacetamide

Authors: Sanjay M. Tailor, Urmila H. Patel

Abstract:

Phthalyl Sulfacetamide belongs to well-known member of antimicrobial sulfonamide family. It is a potent antitumor drug. Structural characteristics of 4-amino-N-(2quinoxalinyl) benzene-sulfonamides (Phthalyl Sulfacetamide), C14H12N4O2S has been studied by method of X-ray crystallography. The compound crystallizes in monoclinic space group P21/n with unit cell parameters a= 7.9841 Ǻ, b= 12.8208 Ǻ, c= 16.6607 Ǻ, α= 90˚, β= 93.23˚, γ= 90˚and Z=4. The X-ray based three-dimensional structure analysis has been carried out by direct methods and refined to an R-value of 0.0419. The crystal structure is stabilized by intermolecular N-H…N, N-H…O and π-π interactions. The Hirshfeld surfaces and consequently the fingerprint analysis have been performed to study the nature of interactions and their quantitative contributions towards the crystal packing. An analysis of Hirshfeld surfaces and fingerprint plots facilitates a comparison of intermolecular interactions, which are the key elements in building different supramolecular architectures. Docking is used for virtual screening for the prediction of the strongest binders based on various scoring functions. Docking studies are carried out on Phthalyl Sulfacetamide for better activity, which is important for the development of a new class of inhibitors.

Keywords: phthalyl sulfacetamide, crystal structure, hirshfeld surface analysis, docking

Procedia PDF Downloads 347
6835 Analysis of Automotive Sensor for Engine Knock System

Authors: Miroslav Gutten, Jozef Jurcik, Daniel Korenciak, Milan Sebok, Matej Kuceraa

Abstract:

This paper deals with the phenomenon of the undesirable detonation combustion in internal combustion engines. A control unit of the engine monitors these detonations using piezoelectric knock sensors. With the control of these sensors the detonations can be objectively measured just outside the car. If this component provides small amplitude of the output voltage it could happen that there would have been in the areas of the engine ignition combustion. The paper deals with the design of a simple device for the detection of this disorder. A construction of the testing device for the knock sensor suitable for diagnostics of knock combustion in internal combustion engines will be presented. The output signal of presented sensor will be described by Bessel functions. Using the first voltage extremes on the characteristics it is possible to create a reference for the evaluation of the polynomial residue. It should be taken into account that the velocity of sound in air is 330 m/s. This sound impinges on the walls of the combustion chamber and is detected by the sensor. The resonant frequency of the clicking of the motor is usually in the range from 5 kHz to 15 kHz. The sensor worked in the field to 37 kHz, which shall be taken into account on an own sensor resonance.

Keywords: diagnostics, knock sensor, measurement, testing device

Procedia PDF Downloads 447
6834 Continuous Functions Modeling with Artificial Neural Network: An Improvement Technique to Feed the Input-Output Mapping

Authors: A. Belayadi, A. Mougari, L. Ait-Gougam, F. Mekideche-Chafa

Abstract:

The artificial neural network is one of the interesting techniques that have been advantageously used to deal with modeling problems. In this study, the computing with artificial neural network (CANN) is proposed. The model is applied to modulate the information processing of one-dimensional task. We aim to integrate a new method which is based on a new coding approach of generating the input-output mapping. The latter is based on increasing the neuron unit in the last layer. Accordingly, to show the efficiency of the approach under study, a comparison is made between the proposed method of generating the input-output set and the conventional method. The results illustrated that the increasing of the neuron units, in the last layer, allows to find the optimal network’s parameters that fit with the mapping data. Moreover, it permits to decrease the training time, during the computation process, which avoids the use of computers with high memory usage.

Keywords: neural network computing, continuous functions generating the input-output mapping, decreasing the training time, machines with big memories

Procedia PDF Downloads 283
6833 Deep Learning for Qualitative and Quantitative Grain Quality Analysis Using Hyperspectral Imaging

Authors: Ole-Christian Galbo Engstrøm, Erik Schou Dreier, Birthe Møller Jespersen, Kim Steenstrup Pedersen

Abstract:

Grain quality analysis is a multi-parameterized problem that includes a variety of qualitative and quantitative parameters such as grain type classification, damage type classification, and nutrient regression. Currently, these parameters require human inspection, a multitude of instruments employing a variety of sensor technologies, and predictive model types or destructive and slow chemical analysis. This paper investigates the feasibility of applying near-infrared hyperspectral imaging (NIR-HSI) to grain quality analysis. For this study two datasets of NIR hyperspectral images in the wavelength range of 900 nm - 1700 nm have been used. Both datasets contain images of sparsely and densely packed grain kernels. The first dataset contains ~87,000 image crops of bulk wheat samples from 63 harvests where protein value has been determined by the FOSS Infratec NOVA which is the golden industry standard for protein content estimation in bulk samples of cereal grain. The second dataset consists of ~28,000 image crops of bulk grain kernels from seven different wheat varieties and a single rye variety. In the first dataset, protein regression analysis is the problem to solve while variety classification analysis is the problem to solve in the second dataset. Deep convolutional neural networks (CNNs) have the potential to utilize spatio-spectral correlations within a hyperspectral image to simultaneously estimate the qualitative and quantitative parameters. CNNs can autonomously derive meaningful representations of the input data reducing the need for advanced preprocessing techniques required for classical chemometric model types such as artificial neural networks (ANNs) and partial least-squares regression (PLS-R). A comparison between different CNN architectures utilizing 2D and 3D convolution is conducted. These results are compared to the performance of ANNs and PLS-R. Additionally, a variety of preprocessing techniques from image analysis and chemometrics are tested. These include centering, scaling, standard normal variate (SNV), Savitzky-Golay (SG) filtering, and detrending. The results indicate that the combination of NIR-HSI and CNNs has the potential to be the foundation for an automatic system unifying qualitative and quantitative grain quality analysis within a single sensor technology and predictive model type.

Keywords: deep learning, grain analysis, hyperspectral imaging, preprocessing techniques

Procedia PDF Downloads 99
6832 Intelligent Production Machine

Authors: A. Şahinoğlu, R. Gürbüz, A. Güllü, M. Karhan

Abstract:

This study in production machines, it is aimed that machine will automatically perceive cutting data and alter cutting parameters. The two most important parameters have to be checked in machine control unit are progress feed rate and speeds. These parameters are aimed to be controlled by sounds of machine. Optimum sound’s features introduced to computer. During process, real time data is received and converted by Matlab software. Data is converted into numerical values. According to them progress and speeds decreases/increases at a certain rate and thus optimum sound is acquired. Cutting process is made in respect of optimum cutting parameters. During chip remove progress, features of cutting tools, kind of cut material, cutting parameters and used machine; affects on various parameters. Instead of required parameters need to be measured such as temperature, vibration, and tool wear that emerged during cutting process; detailed analysis of the sound emerged during cutting process will provide detection of various data that included in the cutting process by the much more easy and economic way. The relation between cutting parameters and sound is being identified.

Keywords: cutting process, sound processing, intelligent late, sound analysis

Procedia PDF Downloads 334
6831 Community’s Role in Slum Development: A Case Study of the Kabul Old City

Authors: Habibi Said Mustafa, Hiroko Ono

Abstract:

Kabul, the capital of Afghanistan, has witnessed a major population growth in the last decades which caused a significant increase in the number of informal settlements. The residents are suffering from many problems such as poor infrastructure, shortage of public facilities, depriving of the sense of ownership and facing much environmental degradation. Even though majority of the residents are living in such condition, the government response has been quite weak. The government’s main planning strategy has been upgrading in the form of provision of facilities and improving elements of physical infrastructure, unfortunately most of the projects which had not community’s support, faced with lots of challenges such as people’s resistance or even in some cases the project rather than solving the problem, compounded them to a worse condition. A major reason as to why many projects have not been effective in achieving certain objectives in the past is the results of local people were not being involved. Community participation plays an important role in slum development projects. The development of an integrated urban development concept these days without the mobilization of participants and community seems difficult and impossible. Therefore this paper looks into factors and methods which can help in creating a participatory approach for developing the slums in Kabul city. Furthermore, it describes the result of a questionnaire which was conducted on a part of the Kabul Old City due to hear resident’s problem related to the slum upgrading and collect their opinions on this regard. The research also points out some factors which can severely hamper the successful implementation of a slum upgrading project if not adequately addressed.

Keywords: Community empowerment, Informal settlements, Job opportunities, People participation

Procedia PDF Downloads 172
6830 The Effect of Multimedia Use on Students’ Academic Achievement and Course-Oriented Self-Efficacy

Authors: Hasan Coruk, Recep Cakir

Abstract:

This study aimed at investigating the effect of multimedia containing ‘the structure and properties of matter’ unit on students’ academic achievement level and self-efficacy relating to science and technology course. The study used an experimental design with pre-test and post-test groups. The data collection tools were ‘Science and Technology Course Achievement Test’ and ‘Science and Technology Self-Efficacy Scale’. The sample of the study consisted of 8th grade students at a primary school in Tokat Province. The study was carried out with 42 students from two classes, 21 (8 males, 13 females) from experimental group and 21 (13 males and 8 females) from control group. The data were analyzed in SPSS.18 software. The findings of the study indicated that the use of multimedia increased the students’ academic achievement in science and technology course in comparison with traditional teaching methods. It was also determined that there was not a significant difference in students’ course-oriented self-efficacy levels regarding the two methods. Necessary and feasible suggestions were put forward for whom it concerns.

Keywords: multimedia learning, science and technology, the structure-properties of matter, self-efficacy, academic achievement

Procedia PDF Downloads 453
6829 Investigation of Shear Thickening Fluid Isolator with Vibration Isolation Performance

Authors: M. C. Yu, Z. L. Niu, L. G. Zhang, W. W. Cui, Y. L. Zhang

Abstract:

According to the theory of the vibration isolation for linear systems, linear damping can reduce the transmissibility at the resonant frequency, but inescapably increase the transmissibility of the isolation frequency region. To resolve this problem, nonlinear vibration isolation technology has recently received increasing attentions. Shear thickening fluid (STF) is a special colloidal material. When STF is subject to high shear rate, it rheological property changes from a flowable behavior into a rigid behavior, i.e., it presents shear thickening effect. STF isolator is a vibration isolator using STF as working material. Because of shear thickening effect, STF isolator is a variable-damped isolator. It exhibits small damping under high vibration frequency and strong damping at resonance frequency due to shearing rate increasing. So its special inherent character is very favorable for vibration isolation, especially for restraining resonance. In this paper, firstly, STF was prepared by dispersing nano-particles of silica into polyethylene glycol 200 fluid, followed by rheological properties test. After that, an STF isolator was designed. The vibration isolation system supported by STF isolator was modeled, and the numerical simulation was conducted to study the vibration isolation properties of STF. And finally, the effect factors on vibrations isolation performance was also researched quantitatively. The research suggests that owing to its variable damping, STF vibration isolator can effetely restrain resonance without bringing unfavorable effect at high frequency, which meets the need of ideal damping properties and resolves the problem of traditional isolators.

Keywords: shear thickening fluid, variable-damped isolator, vibration isolation, restrain resonance

Procedia PDF Downloads 179
6828 A Resource Optimization Strategy for CPU (Central Processing Unit) Intensive Applications

Authors: Junjie Peng, Jinbao Chen, Shuai Kong, Danxu Liu

Abstract:

On the basis of traditional resource allocation strategies, the usage of resources on physical servers in cloud data center is great uncertain. It will cause waste of resources if the assignment of tasks is not enough. On the contrary, it will cause overload if the assignment of tasks is too much. This is especially obvious when the applications are the same type because of its resource preferences. Considering CPU intensive application is one of the most common types of application in the cloud, we studied the optimization strategy for CPU intensive applications on the same server. We used resource preferences to analyze the case that multiple CPU intensive applications run simultaneously, and put forward a model which can predict the execution time for CPU intensive applications which run simultaneously. Based on the prediction model, we proposed the method to select the appropriate number of applications for a machine. Experiments show that the model can predict the execution time accurately for CPU intensive applications. To improve the execution efficiency of applications, we propose a scheduling model based on priority for CPU intensive applications. Extensive experiments verify the validity of the scheduling model.

Keywords: cloud computing, CPU intensive applications, resource optimization, strategy

Procedia PDF Downloads 278
6827 Determination of the Some IGF and IGFBP2 Polymorphisms and Their Association with Growth and Egg Traits in Atak-S Chickens

Authors: Huseyi̇n Das, Bülent Tarim, Sunay Demi̇r, Nurçi̇n Küçükkent, Sevi̇l Cengi̇z, Engi̇n Tülek, Veci̇hi̇ Aksakal

Abstract:

Atak-S laying hens are a high-performance strain obtained by crossing of the Rhode Island Red (RIR) X the Barred Plymouth Rock (BR) and are being produced in the Ankara Poultry Research Institute since 1997. Phenotypic and genetic improving studies are continued for this strain. In this study, 2 from IGF and 1 from IGFBP2, totally 3 different SNP polymorphisms were examined in 200 Atak-S chickens. Genotypes of SNPs were compared using ANOVA to body weight and egg number thorough 32 weeks of age, body weight at sexual maturity, age at sexual maturity and also egg quality traits such as egg shell breaking strength, shell thickness, Haugh unit, albumen index, yolk index, shape index. Only IGF(a) locus was in agreement with Hardy-Weinberg equilibrium, while, the other loci were not. As a result of the performance comparisons to the 3 SNP loci, it was determined that there has a significant association (P<0.05) between only TC genotypes of the IGF(b) locus and body weight at 32 weeks of age, but there was not any association to the other traits.

Keywords: Atak-S, Igf, Igfbp2, single nucleotide polymorphism

Procedia PDF Downloads 367
6826 A Programming Assessment Software Artefact Enhanced with the Help of Learners

Authors: Romeo A. Botes, Imelda Smit

Abstract:

The demands of an ever changing and complex higher education environment, along with the profile of modern learners challenge current approaches to assessment and feedback. More learners enter the education system every year. The younger generation expects immediate feedback. At the same time, feedback should be meaningful. The assessment of practical activities in programming poses a particular problem, since both lecturers and learners in the information and computer science discipline acknowledge that paper-based assessment for programming subjects lacks meaningful real-life testing. At the same time, feedback lacks promptness, consistency, comprehensiveness and individualisation. Most of these aspects may be addressed by modern, technology-assisted assessment. The focus of this paper is the continuous development of an artefact that is used to assist the lecturer in the assessment and feedback of practical programming activities in a senior database programming class. The artefact was developed using three Design Science Research cycles. The first implementation allowed one programming activity submission per assessment intervention. This pilot provided valuable insight into the obstacles regarding the implementation of this type of assessment tool. A second implementation improved the initial version to allow multiple programming activity submissions per assessment. The focus of this version is on providing scaffold feedback to the learner – allowing improvement with each subsequent submission. It also has a built-in capability to provide the lecturer with information regarding the key problem areas of each assessment intervention.

Keywords: programming, computer-aided assessment, technology-assisted assessment, programming assessment software, design science research, mixed-method

Procedia PDF Downloads 296
6825 From Medusa to #MeToo: Different Discourses on Sexual Violence with Particular Reference to the Situation in Serbia

Authors: Jelena Riznić

Abstract:

Sexual violence is a social fact that is both ubiquitous and invisible. From the myth of Medusa and Lucretia, through legends about sexual violence in war conflicts, to Hollywood films and other productions — sexual violence exists as a motive, implicitly or explicitly. Many Hollywood films contain a scene of rape, and the media is increasingly reporting on cases of sexual violence, often not following the guidelines for sensitized and ethical reporting. On the other hand, sexual violence remains an invisible phenomenon if we are talking from the perspective of the survivors. Only the wave of women's testimonies that flooded social networks after the #MeToo campaign in 2017 pointed to the prevalence and to the existing ideas about sexual violence that persist at the level of myths in society, but also through formal norms in the hearing of justice systems. The problem is also in the way rape is defined in the criminal codes of different countries, and all of this affects the reproduction of sexual violence. Precisely because it is a deeply intimate experience of violence, but also a structural problem; on the other hand, understanding sexual violence requires sociological imagination. Accordingly, the subject of this paper is the presentation and analysis of various discourses on sexual violence throughout history — pre/anti-feminist, feminist and criminal law, with particular reference to the situation in Serbia. The paper uses a critical review and comparative analysis of various sources on sexual violence, as well as an analysis of the impact of these sources on the modern legal framework that regulates sexual violence. Research has shown that despite feminist contributions, myths about sexual violence persist and influence the treatment of women who have survived violence in criminal systems and society in general.

Keywords: sexual violence, gender-based violence, MeToo campaign, feminism, Serbia

Procedia PDF Downloads 85
6824 Factors Relating to Motivation to Change Behaviors in Individuals Who Are Overweight

Authors: Teresa Wills, Geraldine Mccarthy, Nicola Cornally

Abstract:

Background: Obesity is an emerging healthcare epidemic affecting virtually all age and socio-economic groups and is one of the most serious and prevalent diseases of the 21st century. It is a public health challenge because of its prevalence, associated costs and health effects. The increasing prevalence of obesity has created a social perception that overweight body sizes are healthy and normal. This normalization of obesity within our society and the acceptance of higher body weights have led to individuals being unaware of the reality of their weight status and gravity of this situation thus impeding recognition of obesity. Given the escalating global health problem of obesity and its co-morbidities, the need to re-appraise its management is more compelling than ever. It is widely accepted that the causes of obesity are complex and multi-factorial. Engagement of individuals in weight management programmes is difficult if they do not perceive they have a problem with their weight. Recognition of the problem is a key component of obesity management and identifying the main predictors of behaviour is key to designing health behaviour interventions. Aim: The aim of the research was to determine factors relating to motivation to change behaviours in individuals who perceive themselves to be overweight. Method: The research design was quantitative, correlational and cross-sectional. The design was guided by the Health Belief Model. Data were collected online using a multi-section and multi-item questionnaire, developed from a review of the theoretical and empirical research. A sample of 202 men and women who perceived themselves to be overweight participated in the research. Descriptive and inferential statistical analyses were employed to describe relationships between variables. Findings: Following multivariate regression analysis, perceived barriers to weight loss and perceived benefits of weight loss were significant predictors of motivation to change behaviour. The perceived barriers to weight loss which were significant were psychological barriers to weight loss (p = < 0.019) and environmental barriers to physical activity (p= < 0.032).The greatest predictor of motivation to change behaviour was the perceived benefits of weight loss (p < 0.001). Perceived susceptibility to obesity and perceived severity of obesity did not emerge as significant predictors in this model. Total variance explained by the model was 33.5%. Conclusion: Perceived barriers to weight loss and perceived benefits of weight loss are important determinants of motivation to change behaviour. These findings have important implications for health professionals to help inform their practice and for the development of intervention programmes to prevent and control obesity.

Keywords: motivation to change behaviours, obesity, predictors of behavior, interventions, overweight

Procedia PDF Downloads 414
6823 Modeling Thermal Changes of Urban Blocks in Relation to the Landscape Structure and Configuration in Guilan Province

Authors: Roshanak Afrakhteh, Abdolrasoul Salman Mahini, Mahdi Motagh, Hamidreza Kamyab

Abstract:

Urban Heat Islands (UHIs) are distinctive urban areas characterized by densely populated central cores surrounded by less densely populated peripheral lands. These areas experience elevated temperatures, primarily due to impermeable surfaces and specific land use patterns. The consequences of these temperature variations are far-reaching, impacting the environment and society negatively, leading to increased energy consumption, air pollution, and public health concerns. This paper emphasizes the need for simplified approaches to comprehend UHI temperature dynamics and explains how urban development patterns contribute to land surface temperature variation. To illustrate this relationship, the study focuses on the Guilan Plain, utilizing techniques like principal component analysis and generalized additive models. The research centered on mapping land use and land surface temperature in the low-lying area of Guilan province. Satellite data from Landsat sensors for three different time periods (2002, 2012, and 2021) were employed. Using eCognition software, a spatial unit known as a "city block" was utilized through object-based analysis. The study also applied the normalized difference vegetation index (NDVI) method to estimate land surface radiance. Predictive variables for urban land surface temperature within residential city blocks were identified categorized as intrinsic (related to the block's structure) and neighboring (related to adjacent blocks) variables. Principal Component Analysis (PCA) was used to select significant variables, and a Generalized Additive Model (GAM) approach, implemented using R's mgcv package, modeled the relationship between urban land surface temperature and predictor variables.Notable findings included variations in urban temperature across different years attributed to environmental and climatic factors. Block size, shared boundary, mother polygon area, and perimeter-to-area ratio were identified as main variables for the generalized additive regression model. This model showed non-linear relationships, with block size, shared boundary, and mother polygon area positively correlated with temperature, while the perimeter-to-area ratio displayed a negative trend. The discussion highlights the challenges of predicting urban surface temperature and the significance of block size in determining urban temperature patterns. It also underscores the importance of spatial configuration and unit structure in shaping urban temperature patterns. In conclusion, this study contributes to the growing body of research on the connection between land use patterns and urban surface temperature. Block size, along with block dispersion and aggregation, emerged as key factors influencing urban surface temperature in residential areas. The proposed methodology enhances our understanding of parameter significance in shaping urban temperature patterns across various regions, particularly in Iran.

Keywords: urban heat island, land surface temperature, LST modeling, GAM, Gilan province

Procedia PDF Downloads 73
6822 Analysis of Erosion Quantity on Application of Conservation Techniques in Ci Liwung Hulu Watershed

Authors: Zaenal Mutaqin

Abstract:

The level of erosion that occurs in the upsteam watersheed will lead to limited infiltrattion, land degradation and river trivialisation and estuaries in the body. One of the watesheed that has been degraded caused by using land is the DA Ci Liwung Upstream. The high degradation that occurs in the DA Ci Liwung upstream is indicated by the hugher rate of erosion on the region, especially in the area of agriculture. In this case, agriculture cultivation intent to the agricultural land that has been applied conservation techniques. This study is applied to determine the quantity of erosion by reviewing Hidrologic Response Unit (HRU) in agricuktural cultivation land which is contained in DA Ci Liwung upstream by using the Soil and Water Assessmen Tool (SWAT). Conservation techniques applied are terracing, agroforestry and gulud terrace. It was concluded that agroforestry conservation techniques show the best value of erosion (lowest) compared with other conservation techniques with the contribution of erosion of 25.22 tonnes/ha/year. The results of the calibration between the discharge flow models with the observation that R²=0.9014 and NS=0.79 indicates that this model is acceptable and feasible applied to the Ci Liwung Hulu watershed.

Keywords: conservation, erosion, SWAT analysis, watersheed

Procedia PDF Downloads 292
6821 Dynamic Wetting and Solidification

Authors: Yulii D. Shikhmurzaev

Abstract:

The modelling of the non-isothermal free-surface flows coupled with the solidification process has become the topic of intensive research with the advent of additive manufacturing, where complex 3-dimensional structures are produced by successive deposition and solidification of microscopic droplets of different materials. The issue is that both the spreading of liquids over solids and the propagation of the solidification front into the fluid and along the solid substrate pose fundamental difficulties for their mathematical modelling. The first of these processes, known as ‘dynamic wetting’, leads to the well-known ‘moving contact-line problem’ where, as shown recently both experimentally and theoretically, the contact angle formed by the free surfac with the solid substrate is not a function of the contact-line speed but is rather a functional of the flow field. The modelling of the propagating solidification front requires generalization of the classical Stefan problem, which would be able to describe the onset of the process and the non-equilibrium regime of solidification. Furthermore, given that both dynamic wetting and solification occur concurrently and interactively, they should be described within the same conceptual framework. The present work addresses this formidable problem and presents a mathematical model capable of describing the key element of additive manufacturing in a self-consistent and singularity-free way. The model is illustrated simple examples highlighting its main features. The main idea of the work is that both dynamic wetting and solidification, as well as some other fluid flows, are particular cases in a general class of flows where interfaces form and/or disappear. This conceptual framework allows one to derive a mathematical model from first principles using the methods of irreversible thermodynamics. Crucially, the interfaces are not considered as zero-mass entities introduced using Gibbsian ‘dividing surface’ but the 2-dimensional surface phases produced by the continuum limit in which the thickness of what physically is an interfacial layer vanishes, and its properties are characterized by ‘surface’ parameters (surface tension, surface density, etc). This approach allows for the mass exchange between the surface and bulk phases, which is the essence of the interface formation. As shown numerically, the onset of solidification is preceded by the pure interface formation stage, whilst the Stefan regime is the final stage where the temperature at the solidification front asymptotically approaches the solidification temperature. The developed model can also be applied to the flow with the substrate melting as well as a complex flow where both types of phase transition take place.

Keywords: dynamic wetting, interface formation, phase transition, solidification

Procedia PDF Downloads 65
6820 Extended Intuitionistic Fuzzy VIKOR Method in Group Decision Making: The Case of Vendor Selection Decision

Authors: Nastaran Hajiheydari, Mohammad Soltani Delgosha

Abstract:

Vendor (supplier) selection is a group decision-making (GDM) process, in which, based on some predetermined criteria, the experts’ preferences are provided in order to rank and choose the most desirable suppliers. In the real business environment, our attitudes or our choices would be made in an uncertain and indecisive situation could not be expressed in a crisp framework. Intuitionistic fuzzy sets (IFSs) could handle such situations in the best way. VIKOR method was developed to solve multi-criteria decision-making (MCDM) problems. This method, which is used to determine the compromised feasible solution with respect to the conflicting criteria, introduces a multi-criteria ranking index based on the particular measure of 'closeness' to the 'ideal solution'. Until now, there has been a little investigation of VIKOR with IFS, therefore we extended the intuitionistic fuzzy (IF) VIKOR to solve vendor selection problem under IF GDM environment. The present study intends to develop an IF VIKOR method in a GDM situation. Therefore, a model is presented to calculate the criterion weights based on entropy measure. Then, the interval-valued intuitionistic fuzzy weighted geometric (IFWG) operator utilized to obtain the total decision matrix. In the next stage, an approach based on the positive idle intuitionistic fuzzy number (PIIFN) and negative idle intuitionistic fuzzy number (NIIFN) was developed. Finally, the application of the proposed method to solve a vendor selection problem illustrated.

Keywords: group decision making, intuitionistic fuzzy set, intuitionistic fuzzy entropy measure, vendor selection, VIKOR

Procedia PDF Downloads 156
6819 An Adjoint-Based Method to Compute Derivatives with Respect to Bed Boundary Positions in Resistivity Measurements

Authors: Mostafa Shahriari, Theophile Chaumont-Frelet, David Pardo

Abstract:

Resistivity measurements are used to characterize the Earth’s subsurface. They are categorized into two different groups: (a) those acquired on the Earth’s surface, for instance, controlled source electromagnetic (CSEM) and Magnetotellurics (MT), and (b) those recorded with borehole logging instruments such as Logging-While-Drilling (LWD) devices. LWD instruments are mostly used for geo-steering purposes, i.e., to adjust dip and azimuthal angles of a well trajectory to drill along a particular geological target. Modern LWD tools measure all nine components of the magnetic field corresponding to three orthogonal transmitter and receiver orientations. In order to map the Earth’s subsurface and perform geo-steering, we invert measurements using a gradient-based method that utilizes the derivatives of the recorded measurements with respect to the inversion variables. For resistivity measurements, these inversion variables are usually the constant resistivity value of each layer and the bed boundary positions. It is well-known how to compute derivatives with respect to the constant resistivity value of each layer using semi-analytic or numerical methods. However, similar formulas for computing the derivatives with respect to bed boundary positions are unavailable. The main contribution of this work is to provide an adjoint-based formulation for computing derivatives with respect to the bed boundary positions. The key idea to obtain the aforementioned adjoint state formulations for the derivatives is to separate the tangential and normal components of the field and treat them differently. This formulation allows us to compute the derivatives faster and more accurately than with traditional finite differences approximations. In the presentation, we shall first derive a formula for computing the derivatives with respect to the bed boundary positions for the potential equation. Then, we shall extend our formulation to 3D Maxwell’s equations. Finally, by considering a 1D domain and reducing the dimensionality of the problem, which is a common practice in the inversion of resistivity measurements, we shall derive a formulation to compute the derivatives of the measurements with respect to the bed boundary positions using a 1.5D variational formulation. Then, we shall illustrate the accuracy and convergence properties of our formulations by comparing numerical results with the analytical derivatives for the potential equation. For the 1.5D Maxwell’s system, we shall compare our numerical results based on the proposed adjoint-based formulation vs those obtained with a traditional finite difference approach. Numerical results shall show that our proposed adjoint-based technique produces enhanced accuracy solutions while its cost is negligible, as opposed to the finite difference approach that requires the solution of one additional problem per derivative.

Keywords: inverse problem, bed boundary positions, electromagnetism, potential equation

Procedia PDF Downloads 178
6818 New Restoration Reagent for Development of Erased Serial Number on Copper Metal Surface

Authors: Lav Kesharwani, Nalini Shankar, A. K. Gupta

Abstract:

A serial number is a unique code assigned for identification of a single unit. Serial number are present on many objects. In an attempt to hide the identity of the numbered item, the numbers are often obliterated or removed by mechanical methods. The present work was carried out with an objective to develop less toxic, less time consuming, more result oriented chemical etching reagent for restoration of serial number on the copper metal plate. Around nine different reagents were prepared using different combination of reagent along with standard reagent and it was applied over 50 erased samples of copper metal and compared it with the standard reagent for restoration of erased marks. After experiment, it was found that the prepared Etching reagent no. 3 (10 g FeCl3 + 20 ml glacial acetic acid + 100 ml distilled H2O) showed the best result for restoration of erased serial number on the copper metal plate .The reagent was also less toxic and less time consuming as compared to standard reagent (19 g FeCl3 + 6 ml cans. HCl + 100 ml distilled H2O).

Keywords: serial number restoration, copper plate, obliteration, chemical method

Procedia PDF Downloads 556
6817 Improvement of the 3D Finite Element Analysis of High Voltage Power Transformer Defects in Time Domain

Authors: M. Rashid Hussain, Shady S. Refaat

Abstract:

The high voltage power transformer is the most essential part of the electrical power utilities. Reliability on the transformers is the utmost concern, and any failure of the transformers can lead to catastrophic losses in electric power utility. The causes of transformer failure include insulation failure by partial discharge, core and tank failure, cooling unit failure, current transformer failure, etc. For the study of power transformer defects, finite element analysis (FEA) can provide valuable information on the severity of defects. FEA provides a more accurate representation of complex geometries because they consider thermal, electrical, and environmental influences on the insulation models to obtain basic characteristics of the insulation system during normal and partial discharge conditions. The purpose of this paper is the time domain analysis of defects 3D model of high voltage power transformer using FEA to study the electric field distribution at different points on the defects.

Keywords: power transformer, finite element analysis, dielectric response, partial discharge, insulation

Procedia PDF Downloads 157
6816 Propagation of Ultra-High Energy Cosmic Rays through Extragalactic Magnetic Fields: An Exploratory Study of the Distance Amplification from Rectilinear Propagation

Authors: Rubens P. Costa, Marcelo A. Leigui de Oliveira

Abstract:

The comprehension of features on the energy spectra, the chemical compositions, and the origins of Ultra-High Energy Cosmic Rays (UHECRs) - mainly atomic nuclei with energies above ~1.0 EeV (exa-electron volts) - are intrinsically linked to the problem of determining the magnitude of their deflections in cosmic magnetic fields on cosmological scales. In addition, as they propagate from the source to the observer, modifications are expected in their original energy spectra, anisotropy, and the chemical compositions due to interactions with low energy photons and matter. This means that any consistent interpretation of the nature and origin of UHECRs has to include the detailed knowledge of their propagation in a three-dimensional environment, taking into account the magnetic deflections and energy losses. The parameter space range for the magnetic fields in the universe is very large because the field strength and especially their orientation have big uncertainties. Particularly, the strength and morphology of the Extragalactic Magnetic Fields (EGMFs) remain largely unknown, because of the intrinsic difficulty of observing them. Monte Carlo simulations of charged particles traveling through a simulated magnetized universe is the straightforward way to study the influence of extragalactic magnetic fields on UHECRs propagation. However, this brings two major difficulties: an accurate numerical modeling of charged particles diffusion in magnetic fields, and an accurate numerical modeling of the magnetized Universe. Since magnetic fields do not cause energy losses, it is important to impose that the particle tracking method conserve the particle’s total energy and that the energy changes are results of the interactions with background photons only. Hence, special attention should be paid to computational effects. Additionally, because of the number of particles necessary to obtain a relevant statistical sample, the particle tracking method must be computationally efficient. In this work, we present an analysis of the propagation of ultra-high energy charged particles in the intergalactic medium. The EGMFs are considered to be coherent within cells of 1 Mpc (mega parsec) diameter, wherein they have uniform intensities of 1 nG (nano Gauss). Moreover, each cell has its field orientation randomly chosen, and a border region is defined such that at distances beyond 95% of the cell radius from the cell center smooth transitions have been applied in order to avoid discontinuities. The smooth transitions are simulated by weighting the magnetic field orientation by the particle's distance to the two nearby cells. The energy losses have been treated in the continuous approximation parameterizing the mean energy loss per unit path length by the energy loss length. We have shown, for a particle with the typical energy of interest the integration method performance in the relative error of Larmor radius, without energy losses and the relative error of energy. Additionally, we plotted the distance amplification from rectilinear propagation as a function of the traveled distance, particle's magnetic rigidity, without energy losses, and particle's energy, with energy losses, to study the influence of particle's species on these calculations. The results clearly show when it is necessary to use a full three-dimensional simulation.

Keywords: cosmic rays propagation, extragalactic magnetic fields, magnetic deflections, ultra-high energy

Procedia PDF Downloads 127
6815 A New Approach in a Problem of a Supersonic Panel Flutter

Authors: M. V. Belubekyan, S. R. Martirosyan

Abstract:

On the example of an elastic rectangular plate streamlined by a supersonic gas flow, we have investigated the phenomenon of divergence and of panel flatter of the overrunning of the gas flow at a free edge under assumption of the presence of concentrated inertial masses and moments at the free edge. We applied a new approach of finding of solution of these problems, which was developed based on the algorithm for an analytical solution finding. This algorithm is easy to use for theoretical studies for the wides circle of nonconservative problems of linear elastic stability. We have established the relation between the characteristics of natural vibrations of the plate and velocity of the streamlining gas flow, which enables one to draw some conclusions on the stability of disturbed motion of the plate depending on the parameters of the system plate-flow. Its solution shows that either the divergence or the localized divergence and the flutter instability are possible. The regions of the stability and instability in space of parameters of the problem are identified. We have investigated the dynamic behavior of the disturbed motion of the panel near the boundaries of region of the stability. The safe and dangerous boundaries of region of the stability are found. The transition through safe boundary of the region of the stability leads to the divergence or localized divergence arising in the vicinity of free edge of the rectangular plate. The transition through dangerous boundary of the region of the stability leads to the panel flutter. The deformations arising at the flutter are more dangerous to the skin of the modern aircrafts and rockets resulting to the loss of the strength and appearance of the fatigue cracks.

Keywords: stability, elastic plate, divergence, localized divergence, supersonic panels flutter

Procedia PDF Downloads 461
6814 Social Distancing as a Population Game in Networked Social Environments

Authors: Zhijun Wu

Abstract:

While social living is considered to be an indispensable part of human life in today's ever-connected world, social distancing has recently received much public attention on its importance since the outbreak of the coronavirus pandemic. In fact, social distancing has long been practiced in nature among solitary species and has been taken by humans as an effective way of stopping or slowing down the spread of infectious diseases. A social distancing problem is considered for how a population, when in the world with a network of social sites, decides to visit or stay at some sites while avoiding or closing down some others so that the social contacts across the network can be minimized. The problem is modeled as a population game, where every individual tries to find some network sites to visit or stay so that he/she can minimize all his/her social contacts. In the end, an optimal strategy can be found for everyone when the game reaches an equilibrium. The paper shows that a large class of equilibrium strategies can be obtained by selecting a set of social sites that forms a so-called maximal r-regular subnetwork. The latter includes many well-studied network types, which are easy to identify or construct and can be completely disconnected (with r = 0) for the most-strict isolation or allow certain degrees of connectivity (with r > 0) for more flexible distancing. The equilibrium conditions of these strategies are derived. Their rigidity and flexibility are analyzed on different types of r-regular subnetworks. It is proved that the strategies supported by maximal 0-regular subnetworks are strictly rigid, while those by general maximal r-regular subnetworks with r > 0 are flexible, though some can be weakly rigid. The proposed model can also be extended to weighted networks when different contact values are assigned to different network sites.

Keywords: social distancing, mitigation of spread of epidemics, populations games, networked social environments

Procedia PDF Downloads 133
6813 A Study on Fundamental Problems for Small and Medium Agricultural Machinery Industries in Central Region Area

Authors: P. Thepnarintra, S. Nikorn

Abstract:

Agricultural machinery industry plays an important role in the industrial development especially the production industry of the country. There has been continuing development responding to the higher demand of the production. However, the problem in agricultural machinery production still exists. Thus, the purpose of this research is to investigate problems on fundamental factors of industry based on the entrepreneurs’ point of view. The focus was on the small and medium size industry receiving a factory license typed number 0660 from the Department of Industrial Works. The investigation was on the comparison between the management of the small and medium size agricultural industry in 3 provinces in the central region of Thailand. Population in this study consisted of 189 company managers or managing directors, of which 101 were from the small size and 88 were from the medium size industry. The data were analyzed to find percentage, arithmetic mean, and standard deviation with independent sample T-test at the statistical significance .05. The results showed that the small and medium size agricultural machinery manufacturers in the central region of Thailand reported high problems in every aspect. When compared the problems on basic factors in running the business, it was found that there was no difference statistically at .05 in managing of the small and medium size agricultural machinery manufacturers. However, there was a statistically significant difference between the small and medium size agricultural machinery manufacturers on the aspect of policy and services of the government. The problems reported by the small and medium size agricultural machinery manufacturers were the services on public tap water and the problem on politic and stability of the country.

Keywords: agricultural machinery, manufacturers, problems, on running the business

Procedia PDF Downloads 292
6812 A Spatial Hypergraph Based Semi-Supervised Band Selection Method for Hyperspectral Imagery Semantic Interpretation

Authors: Akrem Sellami, Imed Riadh Farah

Abstract:

Hyperspectral imagery (HSI) typically provides a wealth of information captured in a wide range of the electromagnetic spectrum for each pixel in the image. Hence, a pixel in HSI is a high-dimensional vector of intensities with a large spectral range and a high spectral resolution. Therefore, the semantic interpretation is a challenging task of HSI analysis. We focused in this paper on object classification as HSI semantic interpretation. However, HSI classification still faces some issues, among which are the following: The spatial variability of spectral signatures, the high number of spectral bands, and the high cost of true sample labeling. Therefore, the high number of spectral bands and the low number of training samples pose the problem of the curse of dimensionality. In order to resolve this problem, we propose to introduce the process of dimensionality reduction trying to improve the classification of HSI. The presented approach is a semi-supervised band selection method based on spatial hypergraph embedding model to represent higher order relationships with different weights of the spatial neighbors corresponding to the centroid of pixel. This semi-supervised band selection has been developed to select useful bands for object classification. The presented approach is evaluated on AVIRIS and ROSIS HSIs and compared to other dimensionality reduction methods. The experimental results demonstrate the efficacy of our approach compared to many existing dimensionality reduction methods for HSI classification.

Keywords: dimensionality reduction, hyperspectral image, semantic interpretation, spatial hypergraph

Procedia PDF Downloads 306
6811 Analysis of Urban Housing Quality and Conditions within Kano Metropolis

Authors: Abdurraheem A. Yakub

Abstract:

Housing is one of the needs of mankind and is one of the best indicators of a person’s standard of living. This research was set out to analyze the housing qualities and conditions in Kano. Primary data was collected through both Personal observations where the researcher carried out an inspection of the study area prior to interview/implementation of questionnaires and took into consideration the type of housing units, construction materials and services available as well as the environmental condition of the study area. This was followed by an interview which was done through personal contact with the various people related to the study. In the course of doing that, questions were asked orally and notes were taken to record the responses. Thereafter, the Questionnaire was implemented which was earlier designed to elicit information from households in the study area using well-structured questions related to the type of facilities provided in the housing unit, types of houses and response with regard to quality of their houses and neighborhoods, tenure of house. The research work looked at the prevailing housing qualities and conditions and the state of the existing facilities and amenities within the environment and offered recommendations on policies and measures that could help improve the situation.

Keywords: housing provision, housing quality, housing standard, housing condition, housing affordability and housing facilities

Procedia PDF Downloads 345
6810 American Slang: Perception and Connotations – Issues of Translation

Authors: Lison Carlier

Abstract:

The English language that is taught in school or used in media nowadays is defined as 'standard English,' although unstandardized Englishes, or 'parallel' Englishes, are practiced throughout the world. The existence of these 'parallel' Englishes has challenged standardization by imposing its own specific vocabulary or grammar. These non-standard languages tend to be regarded as inferior and, therefore, pose a problem regarding their translation. In the USA, 'slanguage', or slang, is a good example of a 'parallel' language. It consists of a particular set of vocabulary, used mostly in speech, and rarely in writing. Qualified as vulgar, often reduced to an urban language spoken by young people from lower classes, slanguage – or the language that is often first spoken between youths – is still the most common language used in the English-speaking world. Moreover, it appears that the prime meaning of 'informal' (as in an informal language) – a language that is spoken with persons the speaker knows – has been put aside and replaced in the general mind by the idea of vulgarity and non-appropriateness, when in fact informality is a sign of intimacy, not of vulgarity. When it comes to translating American slang, the main problem a translator encounters is the image and the cultural background usually associated with this 'parallel' language. Indeed, one will have, unwillingly, a predisposition to categorize a speaker of a 'parallel' language as being part of a particular group of people. The way one sees a speaker using it is paramount, and needs to be transposed into the target language. This paper will conduct an analysis of American slang – its use, perception and the image it gives of its speakers – and its translation into French, using the novel Is Everyone Hanging Out Without Me? (and other concerns) by way of example. In her autobiography/personal essay book, comedy writer, actress and author Mindy Kaling speaks with a very familiar English, including slang, which participates in the construction of her own voice and style, and enables a deeper connection with her readers.

Keywords: translation, English, slang, French

Procedia PDF Downloads 318
6809 Protected Cultivation of Horticultural Crops: Increases Productivity per Unit of Area and Time

Authors: Deepak Loura

Abstract:

The most contemporary method of producing horticulture crops both qualitatively and quantitatively is protected cultivation, or greenhouse cultivation, which has gained widespread acceptance in recent decades. Protected farming, commonly referred to as controlled environment agriculture (CEA), is extremely productive, land- and water-wise, as well as environmentally friendly. The technology entails growing horticulture crops in a controlled environment where variables such as temperature, humidity, light, soil, water, fertilizer, etc. are adjusted to achieve optimal output and enable a consistent supply of them even during the off-season. Over the past ten years, protected cultivation of high-value crops and cut flowers has demonstrated remarkable potential. More and more agricultural and horticultural crop production systems are moving to protected environments as a result of the growing demand for high-quality products by global markets. By covering the crop, it is possible to control the macro- and microenvironments, enhancing plant performance and allowing for longer production times, earlier harvests, and higher yields of higher quality. These shielding features alter the environment of the plant while also offering protection from wind, rain, and insects. Protected farming opens up hitherto unexplored opportunities in agriculture as the liberalised economy and improved agricultural technologies advance. Typically, the revenues from fruit, vegetable, and flower crops are 4 to 8 times higher than those from other crops. If any of these high-value crops are cultivated in protected environments like greenhouses, net houses, tunnels, etc., this profit can be multiplied. Vegetable and cut flower post-harvest losses are extremely high (20–0%), however sheltered growing techniques and year-round cropping can greatly minimize post-harvest losses and enhance yield by 5–10 times. Seasonality and weather have a big impact on the production of vegetables and flowers. The variety of their products results in significant price and quality changes for vegetables. For the application of current technology in crop production, achieving a balance between year-round availability of vegetables and flowers with minimal environmental impact and remaining competitive is a significant problem. The future of agriculture will be protected since population growth is reducing the amount of land that may be held. Protected agriculture is a particularly profitable endeavor for tiny landholdings. Small greenhouses, net houses, nurseries, and low tunnel greenhouses can all be built by farmers to increase their income. Protected agriculture is also aided by the rise in biotic and abiotic stress factors. As a result of the greater productivity levels, these technologies are not only opening up opportunities for producers with larger landholdings, but also for those with smaller holdings. Protected cultivation can be thought of as a kind of precise, forward-thinking, parallel agriculture that covers almost all aspects of farming and is rather subject to additional inspection for technical applicability to circumstances, farmer economics, and market economics.

Keywords: protected cultivation, horticulture, greenhouse, vegetable, controlled environment agriculture

Procedia PDF Downloads 76