Search results for: classical methods
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 15856

Search results for: classical methods

15496 Temporal and Spacial Adaptation Strategies in Aerodynamic Simulation of Bluff Bodies Using Vortex Particle Methods

Authors: Dario Milani, Guido Morgenthal

Abstract:

Fluid dynamic computation of wind caused forces on bluff bodies e.g light flexible civil structures or high incidence of ground approaching airplane wings, is one of the major criteria governing their design. For such structures a significant dynamic response may result, requiring the usage of small scale devices as guide-vanes in bridge design to control these effects. The focus of this paper is on the numerical simulation of the bluff body problem involving multiscale phenomena induced by small scale devices. One of the solution methods for the CFD simulation that is relatively successful in this class of applications is the Vortex Particle Method (VPM). The method is based on a grid free Lagrangian formulation of the Navier-Stokes equations, where the velocity field is modeled by particles representing local vorticity. These vortices are being convected due to the free stream velocity as well as diffused. This representation yields the main advantages of low numerical diffusion, compact discretization as the vorticity is strongly localized, implicitly accounting for the free-space boundary conditions typical for this class of FSI problems, and a natural representation of the vortex creation process inherent in bluff body flows. When the particle resolution reaches the Kolmogorov dissipation length, the method becomes a Direct Numerical Simulation (DNS). However, it is crucial to note that any solution method aims at balancing the computational cost against the accuracy achievable. In the classical VPM method, if the fluid domain is discretized by Np particles, the computational cost is O(Np2). For the coupled FSI problem of interest, for example large structures such as long-span bridges, the aerodynamic behavior may be influenced or even dominated by small structural details such as barriers, handrails or fairings. For such geometrically complex and dimensionally large structures, resolving the complete domain with the conventional VPM particle discretization might become prohibitively expensive to compute even for moderate numbers of particles. It is possible to reduce this cost either by reducing the number of particles or by controlling its local distribution. It is also possible to increase the accuracy of the solution without increasing substantially the global computational cost by computing a correction of the particle-particle interaction in some regions of interest. In this paper different strategies are presented in order to extend the conventional VPM method to reduce the computational cost whilst resolving the required details of the flow. The methods include temporal sub stepping to increase the accuracy of the particles convection in certain regions as well as dynamically re-discretizing the particle map to locally control the global and the local amount of particles. Finally, these methods will be applied on a test case and the improvements in the efficiency as well as the accuracy of the proposed extension to the method are presented. The important benefits in terms of accuracy and computational cost of the combination of these methods will be thus presented as long as their relevant applications.

Keywords: adaptation, fluid dynamic, remeshing, substepping, vortex particle method

Procedia PDF Downloads 257
15495 A Computational Analysis of Flow and Acoustics around a Car Wing Mirror

Authors: Aidan J. Bowes, Reaz Hasan

Abstract:

The automotive industry is continually aiming to develop the aerodynamics of car body design. This may be for a variety of beneficial reasons such as to increase speed or fuel efficiency by reducing drag. However recently there has been a greater amount of focus on wind noise produced while driving. Designers in this industry seek a combination of both simplicity of approach and overall effectiveness. This combined with the growing availability of commercial CFD (Computational Fluid Dynamics) packages is likely to lead to an increase in the use of RANS (Reynolds Averaged Navier-Stokes) based CFD methods. This is due to these methods often being simpler than other CFD methods, having a lower demand on time and computing power. In this investigation the effectiveness of turbulent flow and acoustic noise prediction using RANS based methods has been assessed for different wing mirror geometries. Three different RANS based models were used, standard k-ε, realizable k-ε and k-ω SST. The merits and limitations of these methods are then discussed, by comparing with both experimental and numerical results found in literature. In general, flow prediction is fairly comparable to more complex LES (Large Eddy Simulation) based methods; in particular for the k-ω SST model. However acoustic noise prediction still leaves opportunities for more improvement using RANS based methods.

Keywords: acoustics, aerodynamics, RANS models, turbulent flow

Procedia PDF Downloads 443
15494 Portfolio Risk Management Using Quantum Annealing

Authors: Thomas Doutre, Emmanuel De Meric De Bellefon

Abstract:

This paper describes the application of local-search metaheuristic quantum annealing to portfolio opti- mization. Heuristic technics are particularly handy when Markowitz’ classical Mean-Variance problem is enriched with additional realistic constraints. Once tailored to the problem, computational experiments on real collected data have shown the superiority of quantum annealing over simulated annealing for this constrained optimization problem, taking advantages of quantum effects such as tunnelling.

Keywords: optimization, portfolio risk management, quantum annealing, metaheuristic

Procedia PDF Downloads 379
15493 Contrasted Mean and Median Models in Egyptian Stock Markets

Authors: Mai A. Ibrahim, Mohammed El-Beltagy, Motaz Khorshid

Abstract:

Emerging Markets return distributions have shown significance departure from normality were they are characterized by fatter tails relative to the normal distribution and exhibit levels of skewness and kurtosis that constitute a significant departure from normality. Therefore, the classical Markowitz Mean-Variance is not applicable for emerging markets since it assumes normally-distributed returns (with zero skewness and kurtosis) and a quadratic utility function. Moreover, the Markowitz mean-variance analysis can be used in cases of moderate non-normality and it still provides a good approximation of the expected utility, but it may be ineffective under large departure from normality. Higher moments models and median models have been suggested in the literature for asset allocation in this case. Higher moments models have been introduced to account for the insufficiency of the description of a portfolio by only its first two moments while the median model has been introduced as a robust statistic which is less affected by outliers than the mean. Tail risk measures such as Value-at Risk (VaR) and Conditional Value-at-Risk (CVaR) have been introduced instead of Variance to capture the effect of risk. In this research, higher moment models including the Mean-Variance-Skewness (MVS) and Mean-Variance-Skewness-Kurtosis (MVSK) are formulated as single-objective non-linear programming problems (NLP) and median models including the Median-Value at Risk (MedVaR) and Median-Mean Absolute Deviation (MedMAD) are formulated as a single-objective mixed-integer linear programming (MILP) problems. The higher moment models and median models are compared to some benchmark portfolios and tested on real financial data in the Egyptian main Index EGX30. The results show that all the median models outperform the higher moment models were they provide higher final wealth for the investor over the entire period of study. In addition, the results have confirmed the inapplicability of the classical Markowitz Mean-Variance to the Egyptian stock market as it resulted in very low realized profits.

Keywords: Egyptian stock exchange, emerging markets, higher moment models, median models, mixed-integer linear programming, non-linear programming

Procedia PDF Downloads 311
15492 Parallel Pipelined Conjugate Gradient Algorithm on Heterogeneous Platforms

Authors: Sergey Kopysov, Nikita Nedozhogin, Leonid Tonkov

Abstract:

The article presents a parallel iterative solver for large sparse linear systems which can be used on a heterogeneous platform. Traditionally, the problem of solving linear systems does not scale well on multi-CPU/multi-GPUs clusters. For example, most of the attempts to implement the classical conjugate gradient method were at best counted in the same amount of time as the problem was enlarged. The paper proposes the pipelined variant of the conjugate gradient method (PCG), a formulation that is potentially better suited for hybrid CPU/GPU computing since it requires only one synchronization point per one iteration instead of two for standard CG. The standard and pipelined CG methods need the vector entries generated by the current GPU and other GPUs for matrix-vector products. So the communication between GPUs becomes a major performance bottleneck on multi GPU cluster. The article presents an approach to minimize the communications between parallel parts of algorithms. Additionally, computation and communication can be overlapped to reduce the impact of data exchange. Using the pipelined version of the CG method with one synchronization point, the possibility of asynchronous calculations and communications, load balancing between the CPU and GPU for solving the large linear systems allows for scalability. The algorithm is implemented with the combined use of technologies: MPI, OpenMP, and CUDA. We show that almost optimum speed up on 8-CPU/2GPU may be reached (relatively to a one GPU execution). The parallelized solver achieves a speedup of up to 5.49 times on 16 NVIDIA Tesla GPUs, as compared to one GPU.

Keywords: conjugate gradient, GPU, parallel programming, pipelined algorithm

Procedia PDF Downloads 160
15491 Renovation of Pipeline in Residential Buildings by Polymeric Composites

Authors: Parastou Kharazmi

Abstract:

In this paper, rehabilitation methods for pipeline by advanced polymeric coating such as relining are reviewed. A number of diverse methods which are globally used are described and a brief summary of advances in technology, methods and materials is provided. The paper explains why it is claimed that sewerage rehabilitation with relining in residential buildings is environmentally friendly and economical, the importance of the quality control procedure is discussed and several quality tests are proposed.

Keywords: buildings, composite, material, renovation

Procedia PDF Downloads 265
15490 Cognitivism in Classical Japanese Art and Literature: The Cognitive Value of Haiku and Zen Painting

Authors: Benito Garcia-Valero

Abstract:

This paper analyses the cognitivist value of traditional Japanese theories about aesthetics, art, and literature. These reflections were developed several centuries before actual Cognitive Studies, which started in the seventies of the last century. A comparative methodology is employed to shed light on the similarities between traditional Japanese conceptions about art and current cognitivist principles. The Japanese texts to be compared are Zeami’s treatise on noh art, Okura Toraaki’s Waranbe-gusa on kabuki theatre, and several Buddhist canonical texts about wisdom and knowledge, like the Prajnaparamitahrdaya or Heart Sutra. Japanese contemporary critical sources on these works are also referred, like Nishida Kitaro’s reflections on Zen painting or Ichikawa Hiroshi’s analysis of body/mind dualism in Japanese physical practices. Their ideas are compared with cognitivist authors like George Lakoff, Mark Johnson, Mark Turner and Margaret Freeman. This comparative review reveals the anticipatory ideas of Japanese thinking on body/mind interrelationship, which agrees with cognitivist criticism against dualism, since both elucidate the physical grounds acting upon the formation of concepts and schemes during the production of knowledge. It also highlights the necessity of recovering ancient Japanese treatises on cognition to continue enlightening current research on art and literature. The artistic examples used to illustrate the theory are Sesshu’s Zen paintings and Basho’s classical haiku poetry. Zen painting is an excellent field to demonstrate how monk artists conceived human perception and guessed the active role of beholders during the contemplation of art. On the other hand, some haikus by Matsuo Basho aim at factoring subjectivity out from artistic praxis, which constitutes an ideal of illumination that cannot be achieved using art, due to the embodied nature of perception; a constraint consciously explored by the poet himself. These ideas consolidate the conclusions drawn today by cognitivism about the interrelation between subject and object and the concept of intersubjectivity.

Keywords: cognitivism, dualism, haiku, Zen painting

Procedia PDF Downloads 140
15489 Fine-Grained Sentiment Analysis: Recent Progress

Authors: Jie Liu, Xudong Luo, Pingping Lin, Yifan Fan

Abstract:

Facebook, Twitter, Weibo, and other social media and significant e-commerce sites generate a massive amount of online texts, which can be used to analyse people’s opinions or sentiments for better decision-making. So, sentiment analysis, especially fine-grained sentiment analysis, is a very active research topic. In this paper, we survey various methods for fine-grained sentiment analysis, including traditional sentiment lexicon-based methods, machine learning-based methods, and deep learning-based methods in aspect/target/attribute-based sentiment analysis tasks. Besides, we discuss their advantages and problems worthy of careful studies in the future.

Keywords: sentiment analysis, fine-grained, machine learning, deep learning

Procedia PDF Downloads 256
15488 Constructions of Linear and Robust Codes Based on Wavelet Decompositions

Authors: Alla Levina, Sergey Taranov

Abstract:

The classical approach to the providing noise immunity and integrity of information that process in computing devices and communication channels is to use linear codes. Linear codes have fast and efficient algorithms of encoding and decoding information, but this codes concentrate their detect and correct abilities in certain error configurations. To protect against any configuration of errors at predetermined probability can robust codes. This is accomplished by the use of perfect nonlinear and almost perfect nonlinear functions to calculate the code redundancy. The paper presents the error-correcting coding scheme using biorthogonal wavelet transform. Wavelet transform applied in various fields of science. Some of the wavelet applications are cleaning of signal from noise, data compression, spectral analysis of the signal components. The article suggests methods for constructing linear codes based on wavelet decomposition. For developed constructions we build generator and check matrix that contain the scaling function coefficients of wavelet. Based on linear wavelet codes we develop robust codes that provide uniform protection against all errors. In article we propose two constructions of robust code. The first class of robust code is based on multiplicative inverse in finite field. In the second robust code construction the redundancy part is a cube of information part. Also, this paper investigates the characteristics of proposed robust and linear codes.

Keywords: robust code, linear code, wavelet decomposition, scaling function, error masking probability

Procedia PDF Downloads 485
15487 Tapping into Debt: The Effect of Contactless Payment Methods on Overdraft Fee Occurrence

Authors: Merle Van Den Akker, Neil Stewart, Andrea Isoni

Abstract:

Contactless methods of payment referred to as tap&go, have become increasingly popular globally. However, little is known about the consequences of this payment method on spending, spending habits, personal finance management, and debt accumulation. The literature on other payment methods such as credit cards suggests that, through increased ease and reduced friction, the pain of paying in these methods is reduced, leading to higher and more frequent spending, resulting in higher debt accumulation. Within this research, we use a dataset of 300 million transactions of 165.000 individuals to see whether the onset of using contactless methods of payment increases the occurrence of overdraft fees. Using the R package MatchIt, we find, when matching people on initial overdraft occurrence and salary, that people who do start using contactless incur a significantly higher number of overdraft fees, as compared to those who do not start using contactless in the same year. Having accounted for income, opting-in, and time-of-year effects, these results show that contactless methods of payment fall within the scope of earlier theories on credit cards, such as the pain of paying, meaning that this payment method leads to increasing difficulties managing personal finance.

Keywords: contactless, debt accumulation, overdraft fees, payment methods, spending

Procedia PDF Downloads 117
15486 Integration of Resistivity and Seismic Refraction Using Combine Inversion for Ancient River Findings at Sungai Batu, Lembah Bujang, Malaysia

Authors: Rais Yusoh, Rosli Saad, Mokhtar Saidin, Fauzi Andika, Sabiu Bala Muhammad

Abstract:

Resistivity and seismic refraction profiling have become a common method in pre-investigations for visualizing subsurface structure. The integration of the methods could reduce an interpretation ambiguity. Both methods have their individual software packages for data inversion, but potential to combine certain geophysical methods are restricted; however, the research algorithms that have this functionality was existed and are evaluated personally. The interpretation of subsurface were improve by combining inversion data from both methods by influence each other models using closure coupling; thus, by implementing both methods to support each other which could improve the subsurface interpretation. These methods were applied on a field dataset from a pre-investigation for archeology in finding the ancient river. There were no major changes in the inverted model by combining data inversion for this archetype which probably due to complex geology. The combine data analysis provides an additional technique for interpretation such as an alluvium, which can have strong influence on the ancient river findings.

Keywords: ancient river, combine inversion, resistivity, seismic refraction

Procedia PDF Downloads 326
15485 Understanding Consumer Behaviors by Using Neuromarketing Tools and Methods

Authors: Tabrej Khan

Abstract:

Neuromarketing can refer to the commercial application of neuroscience technologies and insights to drive business further. On the other side, consumer neuroscience can be seen as the academic use of neuroscience to better understand marketing effects on consumer behavior. Consumer Neuroscience and Neuromarketing is a multidisciplinary effort between economics, psychology, and neuroscience and information technology. Traditional methods are using survey, interviews, focus group people are overtly and consciously reporting on their experience and thoughts. The unconscious side of customer behavior is largely unmeasured in the traditional methods. Neuroscience has a potential to understand the unconscious part. Through this paper, we are going to present specific results of selected tools and methods that are used to understand consumer behaviors.

Keywords: neuromarketing, neuroscience, consumer behaviors, tools

Procedia PDF Downloads 393
15484 Teaching Academic Vocabulary: A Recent and Old Approach

Authors: Sara Fine-Meltzer

Abstract:

An obvious, but ill-addressed hindrance to reading comprehension in academic English is poor vocabulary. Unfortunately, dealing with the problem is usually delayed until university entrance. It is the contention of this paper that the chore should be confronted much earlier and by using a very old-fashioned method. This presentation is accompanied by vocabulary lists for advanced level university students with explanations concerning the content and justification for the 500-word lists: how they change over time in accordance with evolving styles of academic writing. There are also sample quizzes and methods to ensure that the words are “absorbed” over time. There is a discussion of other vocabulary acquisition methods and conclusions drawn from the drawbacks of such methods. The paper concludes with the rationale for beginning the study of “academic” vocabulary earlier than is generally acceptable.

Keywords: academic vocabulary, old-fashioned methods, quizzes, vocabulary lists

Procedia PDF Downloads 119
15483 Reading Literacy and Methods of Improving Reading

Authors: Iva Košek Bartošová, Andrea Jokešová, Eva Kozlová, Helena Matějová

Abstract:

The paper presents results of a research team from Faculty of Education, University of Hradec Králové in the Czech Republic. It introduces with the most reading methods used in the 1st classes of a primary school and presents results of a pilot research focused on mastering reading techniques and the quality of reading comprehension of pupils in the first half of a school year during training in teaching reading by an analytic-synthetic method and by a genetic method. These methods of practicing reading skills are the most used ones in the Czech Republic. During the school year 2015/16 there has been a measurement made of two groups of pupils of the 1st year and monitoring of quantitative and qualitative parameters of reading pupils’ outputs by several methods. Both of these methods are based on different theoretical basis and each of them has a specific educational and methodical procedure. This contribution represents results during a piloting project and draws pilot conclusions which will be verified in the subsequent broader research at the end of the school year of the first class of primary school.

Keywords: analytic-synthetic method of reading, genetic method of reading, reading comprehension, reading literacy, reading methods, reading speed

Procedia PDF Downloads 255
15482 Using Audio-Visual Aids and Computer-Assisted Language Instruction (CALI) to Overcome Learning Difficulties of Listening in Students of Special Needs

Authors: Sadeq Al Yaari, Muhammad Alkhunayn, Ayman Al Yaari, Montaha Al Yaari, Adham Al Yaari, Sajedah Al Yaari, Fatehi Eissa

Abstract:

Background & Aims: Audio-visual aids and computer-aided language instruction (CALI) have been documented to improve receptive skills, namely listening skills, in normal students. The increased listening has been attributed to the understanding of other interlocutors' speech, but recent experiments have suggested that audio-visual aids and CALI should be tested against the listening of students of special needs to see the effects of the former in the latter. This investigation described the effect of audio-visual aids and CALI on the performance of these students. Methods: Pre-and-posttests were administered to 40 students of special needs of both sexes at al-Malādh school for students of special needs aged between 8 and 18 years old. A comparison was held between this group of students and another similar group (control group). Whereas the former group underwent a listening course using audio-visual aids and CALI, the latter studied the same course with the same speech language therapist (SLT) with the classical method. The outcomes of the two tests for the two groups were qualitatively and quantitatively analyzed. Results: Significant improvement in the performance was found in the first group (treatment group) (posttest= 72.45% vs. pre-test= 25.55%) in comparison to the second (control) (posttest= 25.55% vs. pre-test= 23.72%). In comparison to the males’ scores, the scores of females are higher (1487 scores vs. 1411 scores). Suggested results support the necessity of the use of audio-visual aids and CALI in teaching listening at the schools of students of special needs.

Keywords: listening, receptive skills, audio-visual aids, CALI, special needs

Procedia PDF Downloads 42
15481 Fuzzy Time Series- Markov Chain Method for Corn and Soybean Price Forecasting in North Carolina Markets

Authors: Selin Guney, Andres Riquelme

Abstract:

Among the main purposes of optimal and efficient forecasts of agricultural commodity prices is to guide the firms to advance the economic decision making process such as planning business operations and marketing decisions. Governments are also the beneficiaries and suppliers of agricultural price forecasts. They use this information to establish a proper agricultural policy, and hence, the forecasts affect social welfare and systematic errors in forecasts could lead to a misallocation of scarce resources. Various empirical approaches have been applied to forecast commodity prices that have used different methodologies. Most commonly-used approaches to forecast commodity sectors depend on classical time series models that assume values of the response variables are precise which is quite often not true in reality. Recently, this literature has mostly evolved to a consideration of fuzzy time series models that provide more flexibility in terms of the classical time series models assumptions such as stationarity, and large sample size requirement. Besides, fuzzy modeling approach allows decision making with estimated values under incomplete information or uncertainty. A number of fuzzy time series models have been developed and implemented over the last decades; however, most of them are not appropriate for forecasting repeated and nonconsecutive transitions in the data. The modeling scheme used in this paper eliminates this problem by introducing Markov modeling approach that takes into account both the repeated and nonconsecutive transitions. Also, the determination of length of interval is crucial in terms of the accuracy of forecasts. The problem of determining the length of interval arbitrarily is overcome and a methodology to determine the proper length of interval based on the distribution or mean of the first differences of series to improve forecast accuracy is proposed. The specific purpose of this paper is to propose and investigate the potential of a new forecasting model that integrates methodologies for determining the proper length of interval based on the distribution or mean of the first differences of series and Fuzzy Time Series- Markov Chain model. Moreover, the accuracy of the forecasting performance of proposed integrated model is compared to different univariate time series models and the superiority of proposed method over competing methods in respect of modelling and forecasting on the basis of forecast evaluation criteria is demonstrated. The application is to daily corn and soybean prices observed at three commercially important North Carolina markets; Candor, Cofield and Roaring River for corn and Fayetteville, Cofield and Greenville City for soybeans respectively. One main conclusion from this paper is that using fuzzy logic improves the forecast performance and accuracy; the effectiveness and potential benefits of the proposed model is confirmed with small selection criteria value such MAPE. The paper concludes with a discussion of the implications of integrating fuzzy logic and nonarbitrary determination of length of interval for the reliability and accuracy of price forecasts. The empirical results represent a significant contribution to our understanding of the applicability of fuzzy modeling in commodity price forecasts.

Keywords: commodity, forecast, fuzzy, Markov

Procedia PDF Downloads 216
15480 Broadband Ultrasonic and Rheological Characterization of Liquids Using Longitudinal Waves

Authors: M. Abderrahmane Mograne, Didier Laux, Jean-Yves Ferrandis

Abstract:

Rheological characterizations of complex liquids like polymer solutions present an important scientific interest for a lot of researchers in many fields as biology, food industry, chemistry. In order to establish master curves (elastic moduli vs frequency) which can give information about microstructure, classical rheometers or viscometers (such as Couette systems) are used. For broadband characterization of the sample, temperature is modified in a very large range leading to equivalent frequency modifications applying the Time Temperature Superposition principle. For many liquids undergoing phase transitions, this approach is not applicable. That is the reason, why the development of broadband spectroscopic methods around room temperature becomes a major concern. In literature many solutions have been proposed but, to our knowledge, there is no experimental bench giving the whole rheological characterization for frequencies about a few Hz (Hertz) to many MHz (Mega Hertz). Consequently, our goal is to investigate in a nondestructive way in very broadband frequency (A few Hz – Hundreds of MHz) rheological properties using longitudinal ultrasonic waves (L waves), a unique experimental bench and a specific container for the liquid: a test tube. More specifically, we aim to estimate the three viscosities (longitudinal, shear and bulk) and the complex elastic moduli (M*, G* and K*) respectively longitudinal, shear and bulk moduli. We have decided to use only L waves conditioned in two ways: bulk L wave in the liquid or guided L waves in the tube test walls. In this paper, we will present first results for very low frequencies using the ultrasonic tracking of a falling ball in the test tube. This will lead to the estimation of shear viscosity from a few mPa.s to a few Pa.s (Pascal second). Corrections due to the small dimensions of the tube will be applied and discussed regarding the size of the falling ball. Then the use of bulk L wave’s propagation in the liquid and the development of a specific signal processing in order to assess longitudinal velocity and attenuation will conduct to the longitudinal viscosity evaluation in the MHz frequency range. At last, the first results concerning the propagation, the generation and the processing of guided compressional waves in the test tube walls will be discussed. All these approaches and results will be compared to standard methods available and already validated in our lab.

Keywords: nondestructive measurement for liquid, piezoelectric transducer, ultrasonic longitudinal waves, viscosities

Procedia PDF Downloads 264
15479 Measuring E-Learning Effectiveness Using a Three-Way Comparison

Authors: Matthew Montebello

Abstract:

The way e-learning effectiveness has been notoriously measured within an academic setting is by comparing the e-learning medium to the traditional face-to-face teaching methodology. In this paper, a simple yet innovative comparison methodology is introduced, whereby the effectiveness of next generation e-learning systems are assessed in contrast not only to the face-to-face mode, but also to the classical e-learning modality. Ethical and logistical issues are also discussed, as this three-way approach to compare teaching methodologies was applied and documented in a real empirical study within a higher education institution.

Keywords: e-learning effectiveness, higher education, teaching modality comparison

Procedia PDF Downloads 381
15478 Parallel Multisplitting Methods for Differential Systems

Authors: Malika El Kyal, Ahmed Machmoum

Abstract:

We prove the superlinear convergence of asynchronous multi-splitting methods applied to differential equations. This study is based on the technique of nested sets. It permits to specify kind of the convergence in the asynchronous mode.The main characteristic of an asynchronous mode is that the local algorithm not have to wait at predetermined messages to become available. We allow some processors to communicate more frequently than others, and we allow the communication delays to be substantial and unpredictable. Note that synchronous algorithms in the computer science sense are particular cases of our formulation of asynchronous one.

Keywords: parallel methods, asynchronous mode, multisplitting, ODE

Procedia PDF Downloads 521
15477 Programming without Code: An Approach and Environment to Conditions-On-Data Programming

Authors: Philippe Larvet

Abstract:

This paper presents the concept of an object-based programming language where tests (if... then... else) and control structures (while, repeat, for...) disappear and are replaced by conditions on data. According to the object paradigm, by using this concept, data are still embedded inside objects, as variable-value couples, but object methods are expressed into the form of logical propositions (‘conditions on data’ or COD).For instance : variable1 = value1 AND variable2 > value2 => variable3 = value3. Implementing this approach, a central inference engine turns and examines objects one after another, collecting all CODs of each object. CODs are considered as rules in a rule-based system: the left part of each proposition (left side of the ‘=>‘ sign) is the premise and the right part is the conclusion. So, premises are evaluated and conclusions are fired. Conclusions modify the variable-value couples of the object and the engine goes to examine the next object. The paper develops the principles of writing CODs instead of complex algorithms. Through samples, the paper also presents several hints for implementing a simple mechanism able to process this ‘COD language’. The proposed approach can be used within the context of simulation, process control, industrial systems validation, etc. By writing simple and rigorous conditions on data, instead of using classical and long-to-learn languages, engineers and specialists can easily simulate and validate the functioning of complex systems.

Keywords: conditions on data, logical proposition, programming without code, object-oriented programming, system simulation, system validation

Procedia PDF Downloads 216
15476 Evaluation of Wheat Sowing and Fertilizer Application Methods in Wheat Weeds Management

Authors: Ebrahim Izadi-Darbandi

Abstract:

In order to investigation the effects of sowing methods, nitrogen and phosphorus application methods in wheat weeds management, an experiment was performed as split plot, based on randomized completely block design with three replications at Research Farm, Faculty of Agriculture, Ferdowsi University of Mashhad, in 2010. Treatments included, wheat sowing methods (single-row with 30 cm distance and twine row on 50 cm width ridges) as main plots and nitrogen and phosphorus application methods (Broadcast and Band) as sub plots. In this experiment, phosphorus and nitrogen sources for fertilization were super phosphate triple (150 kg ha-1) applied before wheat sowing and incorporated with soil and urea (200 kg ha-1) respectively, applied in 2 phases (pre-plant 50%) and near wheat shooting (50%). Results showed that the effect of fertilizers application methods and wheat sowing methods were significant (p≤0.01) on wheat yield increasing and reducing weed-wheat competition. Wheat twine row sowing method, reduced weeds biomass for 25% compared wheat single-row sowing method and increased wheat seed yield and biomass for 60% and 30% respectively. Phosphorus and nitrogen band application reduced weeds biomass for 46% and 53% respectively and increased wheat seed yield for 22% and 33% compared to their broadcast application. The effects of wheat sowing method plus phosphorus and nitrogen application methods interactions, showed that the fertilizers band application and wheat twine-row sowing method were the best methods in wheat yield improvement and reducing wheat-weeds interaction. These results shows that modifying of fertilization methods and wheat sowing method can have important role in fertilizers use efficiency and improving of weeds managements.

Keywords: competition, wheat yield, fertilizer management, biomass

Procedia PDF Downloads 364
15475 The 'Ineffectiveness' of Teaching Research Methods in Moroccan Higher Education: A Qualitative Study

Authors: Ahmed Chouari

Abstract:

Although research methods has been an integral part of the curriculum in Moroccan higher education for decades, it seems that the research methods teaching pedagogy that teachers use suffers from a serious absence of a body of literature in the field. Also, the various challenges that both teachers and students of research methods face have received little interest by researchers in comparison to other fields such as applied linguistics. Therefore, the main aim of this study is to remedy to this situation by exploring one of the major issues in teaching research methods – that is, the phenomenon of students’ dissatisfaction with the research methods course in higher education in Morocco. The aim is also to understand students’ attitudes and perceptions on how to make the research methods course more effective in the future. Three qualitative research questions were used: (1) To what extent are graduate students satisfied with the pedagogies used by the teachers of the research methods course in Moroccan higher education? (2) To what extent are graduate students satisfied with the approach used in assessing research methods in Moroccan higher education? (3) What are students’ perceptions on how to make the research methods course more effective in Moroccan higher education? In this study, a qualitative content analysis was adopted to analyze students’ views and perspectives about the major factors behind their dissatisfaction with the course at the School of Arts and Humanities – University of Moulay Ismail. A semi-structured interview was used to collect data from 14 respondents from two different Master programs. The results show that there is a general consensus among the respondents about the major factors behind the ineffectiveness of the course. These factors include theory-practice gap, heavy reliance on theoretical knowledge at the expense of procedural knowledge, and ineffectiveness of some teachers. The findings also reveal that teaching research methods in Morocco requires more time, better equipment, and more competent teachers. Above all, the findings indicate that today there is an urgent need in Morocco to shift from teacher-centered approaches to learner-centered approaches in teaching the research methods course. These findings, thus, contribute to the existing literature by unraveling the factors that impede the learning process, and by suggesting a set of strategies that can make course more effective.

Keywords: competencies, learner-centered teaching, research methods, student autonomy, pedagogy

Procedia PDF Downloads 261
15474 Vaccination against Hepatitis B in Tunisian Health Care Workers

Authors: Asma Ammar, Nabiha Bouafia , Asma BenCheikh, Mohamed Mahjoub, Olfa Ezzi, Wadiaa Bannour, Radhia Helali, Mansour Njah

Abstract:

Background: The objective of the present study was to identify factors associated with vaccination against Hepatitis B virus (HBV) among healthcare workers (HWs) in the University Hospital Center (UHC) Farhat Hached Sousse, Tunisia. Methods: We conducted a descriptive cross-sectional study all licensed physicians (n= 206) and a representative sample of paramedical staff (n= 372) exercising at UHC Hached Sousse (Tunisia) during two months (January and February 2014). Data were collected using a self-administered and pre-tested questionnaire, which composed by 21 questions. In order to determinate factors associated with vaccination against hepatitis B among HWs, this questionnaire was based on the Health Belief Model, one of the most classical behavior theories. Logistic regression with the stepwise method of Hosmer and Lemeshow was used to identify the determinants of the use of vaccination against HBV. Results: The response rates were 79.8%. Fifty two percent believe that HBV is frequent in our healthcare units and 60.6% consider it a severe infection. The prevalence of HWs vaccination was 39%, 95% CI [34.49%; 43.5%]. In multivariate analysis, determinants of the use of vaccination against HBV among HWs were young age (p=10-4), male gender (p = 0. 006), high or very high importance accorded to health (p = 0.035), perception membership in a risk group for HBV infection (p = 0.038) and very favorable or favorable opinion about vaccination against HVB (p=10-4). Conclusion: The results of our study should be considered in any strategy for preventing VHB infection in HWs. In the mean time, coverage with standard vaccines should be improved also by supplying complete information on the risks of VHB infection and on the safety and efficacy of vaccination.

Keywords: Hepatitis B virus, healthcare workers, prevalence, vaccination

Procedia PDF Downloads 344
15473 Coherent Optical Tomography Imaging of Epidermal Hyperplasia in Vivo in a Mouse Model of Oxazolone Induced Atopic Dermatitis

Authors: Eric Lacoste

Abstract:

Laboratory animals are currently widely used as a model of human pathologies in dermatology such as atopic dermatitis (AD). These models provide a better understanding of the pathophysiology of this complex and multifactorial disease, the discovery of potential new therapeutic targets and the testing of the efficacy of new therapeutics. However, confirmation of the correct development of AD is mainly based on histology from skin biopsies requiring invasive surgery or euthanasia of the animals, plus slicing and staining protocols. However, there are currently accessible imaging technologies such as Optical Coherence Tomography (OCT), which allows non-invasive visualization of the main histological structures of the skin (like stratum corneum, epidermis, and dermis) and assessment of the dynamics of the pathology or efficacy of new treatments. Briefly, female immunocompetent hairless mice (SKH1 strain) were sensitized and challenged topically on back and ears for about 4 weeks. Back skin and ears thickness were measured using calliper at 3 occasions per week in complement to a macroscopic evaluation of atopic dermatitis lesions on back: erythema, scaling and excoriations scoring. In addition, OCT was performed on the back and ears of animals. OCT allows a virtual in-depth section (tomography) of the imaged organ to be made using a laser, a camera and image processing software allowing fast, non-contact and non-denaturing acquisitions of the explored tissues. To perform the imaging sessions, the animals were anesthetized with isoflurane, placed on a support under the OCT for a total examination time of 5 to 10 minutes. The results show a good correlation of the OCT technique with classical HES histology for skin lesions structures such as hyperkeratosis, epidermal hyperplasia, and dermis thickness. This OCT imaging technique can, therefore, be used in live animals at different times for longitudinal evaluation by repeated measurements of lesions in the same animals, in addition to the classical histological evaluation. Furthermore, this original imaging technique speeds up research protocols, reduces the number of animals and refines the use of the laboratory animal.

Keywords: atopic dermatitis, mouse model, oxzolone model, histology, imaging

Procedia PDF Downloads 128
15472 Lean Product Development and Sustainability: A Systematic Literature Review

Authors: João P. E. De Souza, Rob Dekkers

Abstract:

Whereas lean product development aims at maximising customer value whilst optimising product and process design, the question arises whether this approach includes sustainability. A systematic literature review reveals that methods associated with this conceptualisation of product development are suitable for including sustainability, but that the criteria for the triple-bottom line need to be included when using these methods; this is particularly the case for social aspects. Thus, the main finding is that not new methods should be developed, but that existing methods should be more inclusive towards all aspects of sustainability and product life-cycle thinking.

Keywords: lean product development, product life-cycle, sustainability, systematic literature review, triple bottom-line

Procedia PDF Downloads 161
15471 Punishment In Athenian Forensic Oratory

Authors: Eleni Volonaki

Abstract:

In Athenian forensic speeches, the argumentation on punishment of the wrongdoers constitutes a fundamental ideal of exacting justice in court. The present paper explores the variation of approaches to punishment as a means of reformation, revenge, correction, education, example, chance to restoration of justice. As it will be shown, all these approaches reflect the social and political ideology of Athenian justice in the classical period and enhances the role of the courts and the importance of rhetoric in the process of decision-making. Punishment entails a wide range of penalties but also of ideological principles related to the Athenian constitution of democracy.

Keywords: punishment, athenian forensic speeches, justice, athenian democracy

Procedia PDF Downloads 184
15470 Project and Module Based Teaching and Learning

Authors: Jingyu Hou

Abstract:

This paper proposes a new teaching and learning approach-project and Module Based Teaching and Learning (PMBTL). The PMBTL approach incorporates the merits of project/problem based and module based learning methods, and overcomes the limitations of these methods. The correlation between teaching, learning, practice, and assessment is emphasized in this approach, and new methods have been proposed accordingly. The distinct features of these new methods differentiate the PMBTL approach from conventional teaching approaches. Evaluation of this approach on practical teaching and learning activities demonstrates the effectiveness and stability of the approach in improving the performance and quality of teaching and learning. The approach proposed in this paper is also intuitive to the design of other teaching units.

Keywords: computer science education, project and module based, software engineering, module based teaching and learning

Procedia PDF Downloads 487
15469 Testing Immunochemical Method for the Bacteriological Diagnosis of Bovine Tuberculosis

Authors: Assiya Madenovna Borsynbayeva, Kairat Altynbekovich Turgenbayev, Nikolay Petrovich Ivanov

Abstract:

In this article presents the results of rapid diagnostics of tuberculosis in comparison with classical bacteriological method. The proposed method of rapid diagnosis of tuberculosis than bacteriological method allows shortening the time of diagnosis to 7 days, to visualize the growth of mycobacteria in the semi-liquid medium and differentiate the type of mycobacterium. Fast definition of Mycobacterium tuberculosis and its derivatives in the culture medium is a new and promising direction in the diagnosis of tuberculosis.

Keywords: animal diagnosis of tuberculosis, bacteriological diagnostics, antigen, specific antibodies, immunological reaction

Procedia PDF Downloads 340
15468 Cloud-Based Dynamic Routing with Feedback in Formal Methods

Authors: Jawid Ahmad Baktash, Mursal Dawodi, Tomokazu Nagata

Abstract:

With the rapid growth of Cloud Computing, Formal Methods became a good choice for the refinement of message specification and verification for Dynamic Routing in Cloud Computing. Cloud-based Dynamic Routing is becoming increasingly popular. We propose feedback in Formal Methods for Dynamic Routing and Cloud Computing; the model and topologies show how to send messages from index zero to all others formally. The responsibility of proper verification becomes crucial with Dynamic Routing in the cloud. Formal Methods can play an essential role in the routing and development of Networks, and the testing of distributed systems. Event-B is a formal technique that consists of describing the problem rigorously and introduces solutions or details in the refinement steps. Event-B is a variant of B, designed for developing distributed systems and message passing of the dynamic routing. In Event-B and formal methods, the events consist of guarded actions occurring spontaneously rather than being invoked.

Keywords: cloud, dynamic routing, formal method, Pro-B, event-B

Procedia PDF Downloads 419
15467 An Alternative Way to Mapping Cone

Authors: Yousuf Alkhezi

Abstract:

Since most of the literature on algebra does not make much deal with the special case of mapping cone. This paper is an alternative way to examine the special tensor product and mapping cone. Also, we show that the isomorphism that implies the mapping cone commutes with the tensor product for the ordinary tensor product no longer holds for the pinched tensor product. However, we show there is a morphism. We will introduce an alternative way of mapping cone. We are looking for more properties which is our future project. Also, we want to apply these new properties in some application. Many results and examples with classical algorithms will be provided.

Keywords: complex, tensor product, pinched tensore product, mapping cone

Procedia PDF Downloads 124