Search results for: classical predicate calculus
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1094

Search results for: classical predicate calculus

554 Mecano-Reliability Approach Applied to a Water Storage Tank Placed on Ground

Authors: Amar Aliche, Hocine Hammoum, Karima Bouzelha, Arezki Ben Abderrahmane

Abstract:

Traditionally, the dimensioning of storage tanks is conducted with a deterministic approach based on partial coefficients of safety. These coefficients are applied to take into account the uncertainties related to hazards on properties of materials used and applied loads. However, the use of these safety factors in the design process does not assure an optimal and reliable solution and can sometimes lead to a lack of robustness of the structure. The reliability theory based on a probabilistic formulation of constructions safety can respond in an adapted manner. It allows constructing a modelling in which uncertain data are represented by random variables, and therefore allows a better appreciation of safety margins with confidence indicators. The work presented in this paper consists of a mecano-reliability analysis of a concrete storage tank placed on ground. The classical method of Monte Carlo simulation is used to evaluate the failure probability of concrete tank by considering the seismic acceleration as random variable.

Keywords: reliability approach, storage tanks, monte carlo simulation, seismic acceleration

Procedia PDF Downloads 307
553 Comparative Assessment of ISSR and RAPD Markers among Egyptian Jojoba Shrubs

Authors: Abdelsabour G. A. Khaled, Galal A.R. El-Sherbeny, Ahmed M. Hassanein, Gameel M. G. Aly

Abstract:

Classical methods of identification, based on agronomical characterization, are not always the most accurate way due to the instability of these characteristics under the influence of the different environments. In order to estimate the genetic diversity, molecular markers provided excellent tools. In this study, Genetic variation of nine Egyptian jojoba shrubs was tested using ISSR (inter simple sequences repeats), RAPD (random amplified polymorphic DNA) markers and based on the morphological characterization. The average of the percentage of polymorphism (%P) ranged between 58.17% and 74.07% for ISSR and RAPD markers, respectively. The range of genetic similarity percents among shrubs based on ISSR and RAPD markers were from 82.9 to 97.9% and from 85.5 to 97.8%, respectively. The average of PIC (polymorphism information content) values were 0.19 (ISSR) and 0.24 (RAPD). In the present study, RAPD markers were more efficient than the ISSR markers. Where the RAPD technique exhibited higher marker index (MI) average (1.26) compared to ISSR one (1.11). There was an insignificant correlation between the ISSR and RAPD data (0.076, P > 0.05). The dendrogram constructed by the combined RAPD and ISSR data gave a relatively different clustering pattern.

Keywords: correlation, molecular markers, polymorphism, marker index

Procedia PDF Downloads 478
552 Holistic Development of Children through Performing Classical Art Forms: A Study in Tamil Nadu, India

Authors: Meera Rajeev Kumar

Abstract:

An overall social, emotional, and cultural development in a child is what a parent expects. There is no point in comparing the generations of 70’s or 80’s with that of the children of today as the trends are changing drastically. Technology has enabled them to become smart as well as over smart in one way or the other. Children today are quite ignorant of today’s values or ethics and are imbibing different cultures around them and ultimately confused on what to follow. The researcher has gained experience in transmitting or imparting the traditional culture through performing arts. It is understood that the children undergo a transformation from what they knew to what the truth is, through learning and experience. Through performing arts, the child develops an emotional, quick learning, abundant creativity, and ultimately self-realisation on what is right and wrong. The child also gains good organising skills, good decision making skills, therefore summing up to a holistic development. The sample study is 50, and a random sampling technique is adopted to differentiate between a normal child and a child learning an art. The study is conducted in Tamil Nadu, in India.

Keywords: creativity, cultural, emotional, empower

Procedia PDF Downloads 202
551 Topological Analysis of Hydrogen Bonds in Pyruvic Acid-Water Mixtures

Authors: Ferid Hammami

Abstract:

The molecular geometries of the possible conformations of pyruvic acid-water complexes (PA-(H₂O)ₙ = 1- 4) have been fully optimized at DFT/B3LYP/6-311G ++ (d, p) levels of calculation. Among several optimized molecular clusters, the most stable molecular arrangements obtained when one, two, three, and four water molecules are hydrogen-bonded to a central pyruvic acid molecule are presented in this paper. Apposite topological and geometrical parameters are considered as primary indicators of H-bond strength. Atoms in molecules (AIM) analysis shows that pyruvic acid can form a ring structure with water, and the molecular structures are stabilized by both strong O-H...O and C-H...O hydrogen bonds. In large clusters, classical O-H...O hydrogen bonds still exist between water molecules, and a cage-like structure is built around some parts of the central molecule of pyruvic acid. The electrostatic potential energy map (MEP) and the HOMO-LUMO molecular orbital (highest occupied molecular orbital-lowest unoccupied molecular orbital) analysis has been performed for all considered complexes.

Keywords: pyruvic acid, PA-water complex, hydrogen bonding, DFT, AIM, MEP, HOMO-LUMO

Procedia PDF Downloads 214
550 A Unique Multi-Class Support Vector Machine Algorithm Using MapReduce

Authors: Aditi Viswanathan, Shree Ranjani, Aruna Govada

Abstract:

With data sizes constantly expanding, and with classical machine learning algorithms that analyze such data requiring larger and larger amounts of computation time and storage space, the need to distribute computation and memory requirements among several computers has become apparent. Although substantial work has been done in developing distributed binary SVM algorithms and multi-class SVM algorithms individually, the field of multi-class distributed SVMs remains largely unexplored. This research seeks to develop an algorithm that implements the Support Vector Machine over a multi-class data set and is efficient in a distributed environment. For this, we recursively choose the best binary split of a set of classes using a greedy technique. Much like the divide and conquer approach. Our algorithm has shown better computation time during the testing phase than the traditional sequential SVM methods (One vs. One, One vs. Rest) and out-performs them as the size of the data set grows. This approach also classifies the data with higher accuracy than the traditional multi-class algorithms.

Keywords: distributed algorithm, MapReduce, multi-class, support vector machine

Procedia PDF Downloads 401
549 Uncertain Time-Cost Trade off Problems of Construction Projects Using Fuzzy Set Theory

Authors: V. S. S. Kumar, B. Vikram

Abstract:

The development of effective decision support tools that adopted in the construction industry is vital in the world we live in today, since it can lead to substantial cost reduction and efficient resource consumption. Solving the time-cost trade off problems and its related variants is at the heart of scientific research for optimizing construction planning problems. In general, the classical optimization techniques have difficulties in dealing with TCT problems. One of the main reasons of their failure is that they can easily be entrapped in local minima. This paper presents an investigation on the application of meta-heuristic techniques to two particular variants of the time-cost trade of analysis, the time-cost trade off problem (TCT), and time-cost trade off optimization problem (TCO). In first problem, the total project cost should be minimized, and in the second problem, the total project cost and total project duration should be minimized simultaneously. Finally it is expected that, the optimization models developed in this paper will contribute significantly for efficient planning and management of construction project.

Keywords: fuzzy sets, uncertainty, optimization, time cost trade off problems

Procedia PDF Downloads 356
548 Innovative Technologies of Distant Spectral Temperature Control

Authors: Leonid Zhukov, Dmytro Petrenko

Abstract:

Optical thermometry has no alternative in many cases of industrial most effective continuous temperature control. Classical optical thermometry technologies can be used on available for pyrometers controlled objects with stable radiation characteristics and transmissivity of the intermediate medium. Without using temperature corrections, it is possible in the case of a “black” body for energy pyrometry and the cases of “black” and “grey” bodies for spectral ratio pyrometry or with using corrections – for any colored bodies. Consequently, with increasing the number of operating waves, optical thermometry possibilities to reduce methodical errors significantly expand. That is why, in recent 25-30 years, research works have been reoriented on more perfect spectral (multicolor) thermometry technologies. There are two physical material substances, i.e., substance (controlled object) and electromagnetic field (thermal radiation), to be operated in optical thermometry. Heat is transferred by radiation; therefore, radiation has the energy, entropy, and temperature. Optical thermometry was originating simultaneously with the developing of thermal radiation theory when the concept and the term "radiation temperature" was not used, and therefore concepts and terms "conditional temperatures" or "pseudo temperature" of controlled objects were introduced. They do not correspond to the physical sense and definitions of temperature in thermodynamics, molecular-kinetic theory, and statistical physics. Launched by the scientific thermometric society, discussion about the possibilities of temperature measurements of objects, including colored bodies, using the temperatures of their radiation is not finished. Are the information about controlled objects transferred by their radiation enough for temperature measurements? The positive and negative answers on this fundamental question divided experts into two opposite camps. Recent achievements of spectral thermometry develop events in her favour and don’t leave any hope for skeptics. This article presents the results of investigations and developments in the field of spectral thermometry carried out by the authors in the Department of Thermometry and Physics-Chemical Investigations. The authors have many-year’s of experience in the field of modern optical thermometry technologies. Innovative technologies of optical continuous temperature control have been developed: symmetric-wave, two-color compensative, and based on obtained nonlinearity equation of spectral emissivity distribution linear, two-range, and parabolic. Тhe technologies are based on direct measurements of physically substantiated and proposed by Prof. L. Zhukov, radiation temperatures with the next calculation of the controlled object temperature using this radiation temperatures and corresponding mathematical models. Тhe technologies significantly increase metrological characteristics of continuous contactless and light-guide temperature control in energy, metallurgical, ceramic, glassy, and other productions. For example, under the same conditions, the methodical errors of proposed technologies are less than the errors of known spectral and classical technologies in 2 and 3-13 times, respectively. Innovative technologies provide quality products obtaining at the lowest possible resource-including energy costs. More than 600 publications have been published on the completed developments, including more than 100 domestic patents, as well as 34 patents in Australia, Bulgaria, Germany, France, Canada, the USA, Sweden, and Japan. The developments have been implemented in the enterprises of USA, as well as Western Europe and Asia, including Germany and Japan.

Keywords: emissivity, radiation temperature, object temperature, spectral thermometry

Procedia PDF Downloads 98
547 Perspectives of Computational Modeling in Sanskrit Lexicons

Authors: Baldev Ram Khandoliyan, Ram Kishor

Abstract:

India has a classical tradition of Sanskrit Lexicons. Research work has been done on the study of Indian lexicography. India has seen amazing strides in Information and Communication Technology (ICT) applications for Indian languages in general and for Sanskrit in particular. Since Machine Translation from Sanskrit to other Indian languages is often the desired goal, traditional Sanskrit lexicography has attracted a lot of attention from the ICT and Computational Linguistics community. From Nighaŋţu and Nirukta to Amarakośa and Medinīkośa, Sanskrit owns a rich history of lexicography. As these kośas do not follow the same typology or standard in the selection and arrangement of the words and the information related to them, several types of Kośa-styles have emerged in this tradition. The model of a grammar given by Aṣṭādhyāyī is well appreciated by Indian and western linguists and grammarians. But the different models provided by lexicographic tradition also have importance. The general usefulness of Sanskrit traditional Kośas is well discussed by some scholars. That is most of the matter made available in the text. Some also have discussed the good arrangement of lexica. This paper aims to discuss some more use of the different models of Sanskrit lexicography especially focusing on its computational modeling and its use in different computational operations.

Keywords: computational lexicography, Sanskrit Lexicons, nighanṭu, kośa, Amarkosa

Procedia PDF Downloads 164
546 Representativity Based Wasserstein Active Regression

Authors: Benjamin Bobbia, Matthias Picard

Abstract:

In recent years active learning methodologies based on the representativity of the data seems more promising to limit overfitting. The presented query methodology for regression using the Wasserstein distance measuring the representativity of our labelled dataset compared to the global distribution. In this work a crucial use of GroupSort Neural Networks is made therewith to draw a double advantage. The Wasserstein distance can be exactly expressed in terms of such neural networks. Moreover, one can provide explicit bounds for their size and depth together with rates of convergence. However, heterogeneity of the dataset is also considered by weighting the Wasserstein distance with the error of approximation at the previous step of active learning. Such an approach leads to a reduction of overfitting and high prediction performance after few steps of query. After having detailed the methodology and algorithm, an empirical study is presented in order to investigate the range of our hyperparameters. The performances of this method are compared, in terms of numbers of query needed, with other classical and recent query methods on several UCI datasets.

Keywords: active learning, Lipschitz regularization, neural networks, optimal transport, regression

Procedia PDF Downloads 80
545 A Positive Neuroscience Perspective for Child Development and Special Education

Authors: Amedeo D'Angiulli, Kylie Schibli

Abstract:

Traditionally, children’s brain development research has emphasized the limitative aspects of disability and impairment, electing as an explanatory model the classical clinical notions of brain lesion or functional deficit. In contrast, Positive Educational Neuroscience (PEN) is a new approach that emphasizes strengths and human flourishing related to the brain by exploring how learning practices have the potential to enhance neurocognitive flexibility through neuroplastic overcompensation. This mini-review provides an overview of PEN and shows how it links to the concept of neurocognitive flexibility. We provide examples of how the present concept of neurocognitive flexibility can be applied to special education by exploring examples of neuroplasticity in the learning domain, including: (1) learning to draw in congenitally totally blind children, and (2) music training in children from disadvantaged neighborhoods. PEN encourages educators to focus on children’s strengths by recognizing the brain’s capacity for positive change and to incorporate activities that support children’s individual development.

Keywords: neurocognitive development, positive educational neuroscience, sociocultural approach, special education

Procedia PDF Downloads 241
544 Syllogistic Reasoning with 108 Inference Rules While Case Quantities Change

Authors: Mikhail Zarechnev, Bora I. Kumova

Abstract:

A syllogism is a deductive inference scheme used to derive a conclusion from a set of premises. In a categorical syllogisms, there are only two premises and every premise and conclusion is given in form of a quantified relationship between two objects. The different order of objects in premises give classification known as figures. We have shown that the ordered combinations of 3 generalized quantifiers with certain figure provide in total of 108 syllogistic moods which can be considered as different inference rules. The classical syllogistic system allows to model human thought and reasoning with syllogistic structures always attracted the attention of cognitive scientists. Since automated reasoning is considered as part of learning subsystem of AI agents, syllogistic system can be applied for this approach. Another application of syllogistic system is related to inference mechanisms on the Semantic Web applications. In this paper we proposed the mathematical model and algorithm for syllogistic reasoning. Also the model of iterative syllogistic reasoning in case of continuous flows of incoming data based on case–based reasoning and possible applications of proposed system were discussed.

Keywords: categorical syllogism, case-based reasoning, cognitive architecture, inference on the semantic web, syllogistic reasoning

Procedia PDF Downloads 411
543 Finite Element Simulation for Preliminary Study on Microorganism Detection System

Authors: Muhammad Rosli Abdullah, Noor Hasmiza Harun

Abstract:

A microorganism detection system has a potential to be used with the advancement in a biosensor development. The detection system requires an optical sensing system, microfluidic device and biological reagent. Although, the biosensors are available in the market, a label free and a lab-on-chip approach will promote a flexible solution. As a preliminary study of microorganism detection, three mechanisms such as Total Internal Reflection (TIR), Micro Fluidic Channel (MFC) and magnetic-electric field propagation were study and simulated. The objective are to identify the TIR angle, MFC parabolic flow and the wavelength for the microorganism detection. The simulation result indicates that evanescent wave is achieved when TIR angle > 42°, the corner and centre of a parabolic velocity are 0.02 m/s and 0.06 m/s respectively, and a higher energy distribution of a perfect electromagnetic scattering with dipole resonance radiation occurs at 500 nm. This simulation is beneficial to determine the components of the microorganism detection system that does not rely on classical microbiological, immunological and genetic methods which are laborious, time-consuming procedures and confined to specialized laboratories with expensive instrumentation equipment.

Keywords: microorganism, microfluidic, total internal reflection, lab on chip

Procedia PDF Downloads 277
542 Local Boundary Analysis for Generative Theory of Tonal Music: From the Aspect of Classic Music Melody Analysis

Authors: Po-Chun Wang, Yan-Ru Lai, Sophia I. C. Lin, Alvin W. Y. Su

Abstract:

The Generative Theory of Tonal Music (GTTM) provides systematic approaches to recognizing local boundaries of music. The rules have been implemented in some automated melody segmentation algorithms. Besides, there are also deep learning methods with GTTM features applied to boundary detection tasks. However, these studies might face constraints such as a lack of or inconsistent label data. The GTTM database is currently the most widely used GTTM database, which includes manually labeled GTTM rules and local boundaries. Even so, we found some problems with these labels. They are sometimes discrepancies with GTTM rules. In addition, since it is labeled at different times by multiple musicians, they are not within the same scope in some cases. Therefore, in this paper, we examine this database with musicians from the aspect of classical music and relabel the scores. The relabeled database - GTTM Database v2.0 - will be released for academic research usage. Despite the experimental and statistical results showing that the relabeled database is more consistent, the improvement in boundary detection is not substantial. It seems that we need more clues than GTTM rules for boundary detection in the future.

Keywords: dataset, GTTM, local boundary, neural network

Procedia PDF Downloads 145
541 Social Data-Based Users Profiles' Enrichment

Authors: Amel Hannech, Mehdi Adda, Hamid Mcheick

Abstract:

In this paper, we propose a generic model of user profile integrating several elements that may positively impact the research process. We exploit the classical behavior of users and integrate a delimitation process of their research activities into several research sessions enriched with contextual and temporal information, which allows reflecting the current interests of these users in every period of time and infer data freshness. We argue that the annotation of resources gives more transparency on users' needs. It also strengthens social links among resources and users, and can so increase the scope of the user profile. Based on this idea, we integrate the social tagging practice in order to exploit the social users' behavior to enrich their profiles. These profiles are then integrated into a recommendation system in order to predict the interesting personalized items of users allowing to assist them in their researches and further enrich their profiles. In this recommendation, we provide users new research experiences.

Keywords: user profiles, topical ontology, contextual information, folksonomies, tags' clusters, data freshness, association rules, data recommendation

Procedia PDF Downloads 264
540 Fuzzy Time Series- Markov Chain Method for Corn and Soybean Price Forecasting in North Carolina Markets

Authors: Selin Guney, Andres Riquelme

Abstract:

Among the main purposes of optimal and efficient forecasts of agricultural commodity prices is to guide the firms to advance the economic decision making process such as planning business operations and marketing decisions. Governments are also the beneficiaries and suppliers of agricultural price forecasts. They use this information to establish a proper agricultural policy, and hence, the forecasts affect social welfare and systematic errors in forecasts could lead to a misallocation of scarce resources. Various empirical approaches have been applied to forecast commodity prices that have used different methodologies. Most commonly-used approaches to forecast commodity sectors depend on classical time series models that assume values of the response variables are precise which is quite often not true in reality. Recently, this literature has mostly evolved to a consideration of fuzzy time series models that provide more flexibility in terms of the classical time series models assumptions such as stationarity, and large sample size requirement. Besides, fuzzy modeling approach allows decision making with estimated values under incomplete information or uncertainty. A number of fuzzy time series models have been developed and implemented over the last decades; however, most of them are not appropriate for forecasting repeated and nonconsecutive transitions in the data. The modeling scheme used in this paper eliminates this problem by introducing Markov modeling approach that takes into account both the repeated and nonconsecutive transitions. Also, the determination of length of interval is crucial in terms of the accuracy of forecasts. The problem of determining the length of interval arbitrarily is overcome and a methodology to determine the proper length of interval based on the distribution or mean of the first differences of series to improve forecast accuracy is proposed. The specific purpose of this paper is to propose and investigate the potential of a new forecasting model that integrates methodologies for determining the proper length of interval based on the distribution or mean of the first differences of series and Fuzzy Time Series- Markov Chain model. Moreover, the accuracy of the forecasting performance of proposed integrated model is compared to different univariate time series models and the superiority of proposed method over competing methods in respect of modelling and forecasting on the basis of forecast evaluation criteria is demonstrated. The application is to daily corn and soybean prices observed at three commercially important North Carolina markets; Candor, Cofield and Roaring River for corn and Fayetteville, Cofield and Greenville City for soybeans respectively. One main conclusion from this paper is that using fuzzy logic improves the forecast performance and accuracy; the effectiveness and potential benefits of the proposed model is confirmed with small selection criteria value such MAPE. The paper concludes with a discussion of the implications of integrating fuzzy logic and nonarbitrary determination of length of interval for the reliability and accuracy of price forecasts. The empirical results represent a significant contribution to our understanding of the applicability of fuzzy modeling in commodity price forecasts.

Keywords: commodity, forecast, fuzzy, Markov

Procedia PDF Downloads 217
539 Model and Neural Control of the Depth of Anesthesia during Surgery

Authors: Javier Fernandez, Mayte Medina, Rafael Fernandez de Canete, Nuria Alcain, Juan Carlos Ramos-Diaz

Abstract:

At present, the experimentation of anesthetic drugs on patients requires a regulation protocol, and the response of each patient to several doses of entry drug must be well known. Therefore, the development of pharmacological dose control systems is a promising field of research in anesthesiology. In this paper, it has been developed a non-linear compartmental the pharmacokinetic-pharmacodynamical model which describes the anesthesia depth effect in a sufficiently reliable way over a set of patients with the depth effect quantified by the Bi-Spectral Index. Afterwards, an Artificial Neural Network (ANN) predictive controller has been designed based on the depth of anesthesia model so as to keep the patient in the optimum condition while he undergoes surgical treatment. For the purpose of quantifying the efficiency of the neural predictive controller, a classical proportional-integral-derivative controller has also been developed to compare both strategies. Results show the superior performance of predictive neural controller during BiSpectral Index reference tracking.

Keywords: anesthesia, bi-spectral index, neural network control, pharmacokinetic-pharmacodynamical model

Procedia PDF Downloads 336
538 [Keynote Talk]: Existence of Random Fixed Point Theorem for Contractive Mappings

Authors: D. S. Palimkar

Abstract:

Random fixed point theory has received much attention in recent years, and it is needed for the study of various classes of random equations. The study of random fixed point theorems was initiated by the Prague school of probabilistic in the 1950s. The existence and uniqueness of fixed points for the self-maps of a metric space by altering distances between the points with the use of a control function is an interesting aspect in the classical fixed point theory. In a new category of fixed point problems for a single self-map with the help of a control function that alters the distance between two points in a metric space which they called an altering distance function. In this paper, we prove the results of existence of random common fixed point and its uniqueness for a pair of random mappings under weakly contractive condition for generalizing alter distance function in polish spaces using Random Common Fixed Point Theorem for Generalized Weakly Contractions.

Keywords: Polish space, random common fixed point theorem, weakly contractive mapping, altering function

Procedia PDF Downloads 273
537 The Intercultural Communicative Competence (ICC) Perspective in the Film Classroom

Authors: Yan Zhang

Abstract:

With the development of commercial movies, more and more instructors are drawn to adapt film pedagogy to teach history and culture. By challenging traditional standards of classroom culture, instruction through film represents an intersection of modernity and adaptability which is no longer optional but essential to maintaining educational accessibility. First, this presentation describes special features of the film that can be used in the classroom and help students acquire intercultural communicative competence (ICC) and achieve the learning goal. Second, the author brings forward the 5 A STAIRCASE model (Acknowledge-Adjust-Acculturate-Act-Assess) to explore how students acquire international communicative competence. Third, this article presents the intersections between new digital environments and classroom practice, such as how films can contribute to combining classical and contemporary Chinese cultures seamlessly and how film pedagogy can be an effective way to get students to engage in deeper critical thinking by exposing them to visuals, music, language, and styling which do not exist in traditional learning formats. Last, the student’s final video project will be exemplified at the end, demonstrating how to engage students in the analysis and experience of history and culture.

Keywords: intercultural education, curriculum, media, history

Procedia PDF Downloads 74
536 Post-modernist Tragi-Comedy: A Study of Tom Stoppard’s “Rosencrantz and Guildenstern Are Dead”

Authors: Azza Taha Zaki

Abstract:

The death of tragedy is probably the most distinctive literary controversy of the twentieth century. There is common critical consent that tragedy in the classical sense of the word is no longer possible. Thinkers, philosophers, and critics such as Nietzsche, Durrenmatt, and George Steiner have all agreed that the decline of the genre in the modern age is due to the total lack of a unified world image and the absence of a shared vision in a fragmented and ideologically diversified world. The production of Rosencrantz and Guildenstern are Dead in 1967 marked the rise of the genre of tragi-comedy as a more appropriate reflection of the spirit of the age. At the hands of such great dramatists as Tom Stoppard (1937- ), the revived genre was not used as an extra comic element to give some comic relief to an otherwise tragic text, but it was given a postmodernist touch to serve the interpretation of the dilemma of man in the postmodernist world. This paper will study features of postmodernist tragi-comedy in Rosencrantz and Guildenstern are Dead as one of the most important plays in modern British theatre and investigate Stoppard’s vision of man and life as influenced by postmodernist thought and philosophy.

Keywords: British, drama, postmodernist, Stoppard, tragi-comedy

Procedia PDF Downloads 186
535 The Experimental Measurement of the LiBr Concentration of a Solar Absorption Machine

Authors: N. Hatraf, L. Merabti, Z. Neffah, W. Taane

Abstract:

The excessive consumption of fossil energies (electrical energy) during summer caused by the technological development involves more and more climate warming. In order to reduce the worst impact of gas emissions produced from classical air conditioning, heat driven solar absorption chiller is pretty promising; it consists on using solar as motive energy which is clean and environmentally friendly to provide cold. Solar absorption machine is composed by four components using Lithium Bromide /water as a refrigerating couple. LiBr- water is the most promising in chiller applications due to high safety, high volatility ratio, high affinity, high stability and its high latent heat. The lithium bromide solution is constitute by the salt lithium bromide which absorbs water under certain conditions of pressure and temperature however if the concentration of the solution is high in the absorption chillers; which exceed 70%, the solution will crystallize. The main aim of this article is to study the phenomena of the crystallization and to evaluate how the dependence between the electric conductivity and the concentration which should be controlled.

Keywords: absorption, crystallization, experimental results, lithium bromide solution

Procedia PDF Downloads 310
534 Replacing MOSFETs with Single Electron Transistors (SET) to Reduce Power Consumption of an Inverter Circuit

Authors: Ahmed Shariful Alam, Abu Hena M. Mustafa Kamal, M. Abdul Rahman, M. Nasmus Sakib Khan Shabbir, Atiqul Islam

Abstract:

According to the rules of quantum mechanics there is a non-vanishing probability of for an electron to tunnel through a thin insulating barrier or a thin capacitor which is not possible according to the laws of classical physics. Tunneling of electron through a thin insulating barrier or tunnel junction is a random event and the magnitude of current flowing due to the tunneling of electron is very low. As the current flowing through a Single Electron Transistor (SET) is the result of electron tunneling through tunnel junctions of its source and drain the supply voltage requirement is also very low. As a result, the power consumption across a Single Electron Transistor is ultra-low in comparison to that of a MOSFET. In this paper simulations have been done with PSPICE for an inverter built with both SETs and MOSFETs. 35mV supply voltage was used for a SET built inverter circuit and the supply voltage used for a CMOS inverter was 3.5V.

Keywords: ITRS, enhancement type MOSFET, island, DC analysis, transient analysis, power consumption, background charge co-tunneling

Procedia PDF Downloads 526
533 Multisymplectic Geometry and Noether Symmetries for the Field Theories and the Relativistic Mechanics

Authors: H. Loumi-Fergane, A. Belaidi

Abstract:

The problem of symmetries in field theory has been analyzed using geometric frameworks, such as the multisymplectic models by using in particular the multivector field formalism. In this paper, we expand the vector fields associated to infinitesimal symmetries which give rise to invariant quantities as Noether currents for classical field theories and relativistic mechanic using the multisymplectic geometry where the Poincaré-Cartan form has thus been greatly simplified using the Second Order Partial Differential Equation (SOPDE) for multi-vector fields verifying Euler equations. These symmetries have been classified naturally according to the construction of the fiber bundle used.  In this work, unlike other works using the analytical method, our geometric model has allowed us firstly to distinguish the angular moments of the gauge field obtained during different transformations while these moments are gathered in a single expression and are obtained during a rotation in the Minkowsky space. Secondly, no conditions are imposed on the Lagrangian of the mechanics with respect to its dependence in time and in qi, the currents obtained naturally from the transformations are respectively the energy and the momentum of the system.

Keywords: conservation laws, field theories, multisymplectic geometry, relativistic mechanics

Procedia PDF Downloads 206
532 Poetic Music by the Poet, Commander of the Faithful, Muhammad Bello: Prosodical Study

Authors: Sirajo Muhammad Sokoto

Abstract:

The Commander of the Faithful, Muhammad Bello, is considered one of the most distinguished scholars and poetic geniuses who is famous for reciting poetry in the classical vertical style. He is also represented by pre-Islamic poets such as Imru’ al-Qays and Alqamah and among the Islamists such as Hassan bin Thabit, Amr bin Abi Rabi’ah, and others. The poet drew from the seas of the Arabic language and its styles at the hands of His father, Sheikh Othman Bin Fodio, and his uncle, Sheikh Abdullah Bin Fodio, are both things that made Muhammad Bello conversant with the Arabic language until he was able to write poetry in a beautiful format and good style. The Commander of the Faithful, Muhammad Bello, did not deviate from what the Arabs know of poetic elements, such as taking into account its meanings and music; Muhammadu Bello has used every Bahr of prosody and its technicals in many of his poems. This article prepares the reader for the efforts made by the poet Muhammad Bello in composing poems on poetic seas, taking into account musical tones for different purposes according to his desire. The article will also discuss the poet’s talent, skill, and eloquence.

Keywords: music, Muhammad Bello, poetry, performances

Procedia PDF Downloads 76
531 Decoding Kinematic Characteristics of Finger Movement from Electrocorticography Using Classical Methods and Deep Convolutional Neural Networks

Authors: Ksenia Volkova, Artur Petrosyan, Ignatii Dubyshkin, Alexei Ossadtchi

Abstract:

Brain-computer interfaces are a growing research field producing many implementations that find use in different fields and are used for research and practical purposes. Despite the popularity of the implementations using non-invasive neuroimaging methods, radical improvement of the state channel bandwidth and, thus, decoding accuracy is only possible by using invasive techniques. Electrocorticography (ECoG) is a minimally invasive neuroimaging method that provides highly informative brain activity signals, effective analysis of which requires the use of machine learning methods that are able to learn representations of complex patterns. Deep learning is a family of machine learning algorithms that allow learning representations of data with multiple levels of abstraction. This study explores the potential of deep learning approaches for ECoG processing, decoding movement intentions and the perception of proprioceptive information. To obtain synchronous recording of kinematic movement characteristics and corresponding electrical brain activity, a series of experiments were carried out, during which subjects performed finger movements at their own pace. Finger movements were recorded with a three-axis accelerometer, while ECoG was synchronously registered from the electrode strips that were implanted over the contralateral sensorimotor cortex. Then, multichannel ECoG signals were used to track finger movement trajectory characterized by accelerometer signal. This process was carried out both causally and non-causally, using different position of the ECoG data segment with respect to the accelerometer data stream. The recorded data was split into training and testing sets, containing continuous non-overlapping fragments of the multichannel ECoG. A deep convolutional neural network was implemented and trained, using 1-second segments of ECoG data from the training dataset as input. To assess the decoding accuracy, correlation coefficient r between the output of the model and the accelerometer readings was computed. After optimization of hyperparameters and training, the deep learning model allowed reasonably accurate causal decoding of finger movement with correlation coefficient r = 0.8. In contrast, the classical Wiener-filter like approach was able to achieve only 0.56 in the causal decoding mode. In the noncausal case, the traditional approach reached the accuracy of r = 0.69, which may be due to the presence of additional proprioceptive information. This result demonstrates that the deep neural network was able to effectively find a representation of the complex top-down information related to the actual movement rather than proprioception. The sensitivity analysis shows physiologically plausible pictures of the extent to which individual features (channel, wavelet subband) are utilized during the decoding procedure. In conclusion, the results of this study have demonstrated that a combination of a minimally invasive neuroimaging technique such as ECoG and advanced machine learning approaches allows decoding motion with high accuracy. Such setup provides means for control of devices with a large number of degrees of freedom as well as exploratory studies of the complex neural processes underlying movement execution.

Keywords: brain-computer interface, deep learning, ECoG, movement decoding, sensorimotor cortex

Procedia PDF Downloads 177
530 Analysis of Complex Business Negotiations: Contributions from Agency-Theory

Authors: Jan Van Uden

Abstract:

The paper reviews classical agency-theory and its contributions to the analysis of complex business negotiations and gives an approach for the modification of the basic agency-model in order to examine the negotiation specific dimensions of agency-problems. By illustrating fundamental potentials for the modification of agency-theory in context of business negotiations the paper highlights recent empirical research that investigates agent-based negotiations and inter-team constellations. A general theoretical analysis of complex negotiation would be based on a two-level approach. First, the modification of the basic agency-model in order to illustrate the organizational context of business negotiations (i.e., multi-agent issues, common-agencies, multi-period models and the concept of bounded rationality). Second, the application of the modified agency-model on complex business negotiations to identify agency-problems and relating areas of risk in the negotiation process. The paper is placed on the first level of analysis – the modification. The method builds on the one hand on insights from behavior decision research (BRD) and on the other hand on findings from agency-theory as normative directives to the modification of the basic model. Through neoclassical assumptions concerning the fundamental aspects of agency-relationships in business negotiations (i.e., asymmetric information, self-interest, risk preferences and conflict of interests), agency-theory helps to draw solutions on stated worst-case-scenarios taken from the daily negotiation routine. As agency-theory is the only universal approach able to identify trade-offs between certain aspects of economic cooperation, insights obtained provide a deeper understanding of the forces that shape business negotiation complexity. The need for a modification of the basic model is illustrated by highlighting selected issues of business negotiations from agency-theory perspective: Negotiation Teams require a multi-agent approach under the condition that often decision-makers as superior-agents are part of the team. The diversity of competences and decision-making authority is a phenomenon that overrides the assumptions of classical agency-theory and varies greatly in context of certain forms of business negotiations. Further, the basic model is bound to dyadic relationships preceded by the delegation of decision-making authority and builds on a contractual created (vertical) hierarchy. As a result, horizontal dynamics within the negotiation team playing an important role for negotiation success are therefore not considered in the investigation of agency-problems. Also, the trade-off between short-term relationships within the negotiation sphere and the long-term relationships of the corporate sphere calls for a multi-period perspective taking into account the sphere-specific governance-mechanisms already established (i.e., reward and monitoring systems). Within the analysis, the implementation of bounded rationality is closely related to findings from BRD to assess the impact of negotiation behavior on underlying principal-agent-relationships. As empirical findings show, the disclosure and reservation of information to the agent affect his negotiation behavior as well as final negotiation outcomes. Last, in context of business negotiations, asymmetric information is often intended by decision-makers acting as superior-agents or principals which calls for a bilateral risk-approach to agency-relations.

Keywords: business negotiations, agency-theory, negotiation analysis, interteam negotiations

Procedia PDF Downloads 139
529 Path Integrals and Effective Field Theory of Large Scale Structure

Authors: Revant Nayar

Abstract:

In this work, we recast the equations describing large scale structure, and by extension all nonlinear fluids, in the path integral formalism. We first calculate the well known two and three point functions using Schwinger Keldysh formalism used commonly to perturbatively solve path integrals in non- equilibrium systems. Then we include EFT corrections due to pressure, viscosity, and noise as effects on the time-dependent propagator. We are able to express results for arbitrary two and three point correlation functions in LSS in terms of differential operators acting on a triple K master intergral. We also, for the first time, get analytical results for more general initial conditions deviating from the usual power law P∝kⁿ by introducing a mass scale in the initial conditions. This robust field theoretic formalism empowers us with tools from strongly coupled QFT to study the strongly non-linear regime of LSS and turbulent fluid dynamics such as OPE and holographic duals. These could be used to capture fully the strongly non-linear dynamics of fluids and move towards solving the open problem of classical turbulence.

Keywords: quantum field theory, cosmology, effective field theory, renormallisation

Procedia PDF Downloads 135
528 Spatially Downscaling Land Surface Temperature with a Non-Linear Model

Authors: Kai Liu

Abstract:

Remote sensing-derived land surface temperature (LST) can provide an indication of the temporal and spatial patterns of surface evapotranspiration (ET). However, the spatial resolution achieved by existing commonly satellite products is ~1 km, which remains too coarse for ET estimations. This paper proposed a model that can disaggregate coarse resolution MODIS LST at 1 km scale to fine spatial resolutions at the scale of 250 m. Our approach attempted to weaken the impacts of soil moisture and growing statues on LST variations. The proposed model spatially disaggregates the coarse thermal data by using a non-linear model involving Bowen ratio, normalized difference vegetation index (NDVI) and photochemical reflectance index (PRI). This LST disaggregation model was tested on two heterogeneous landscapes in central Iowa, USA and Heihe River, China, during the growing seasons. Statistical results demonstrated that our model achieved better than the two classical methods (DisTrad and TsHARP). Furthermore, using the surface energy balance model, it was observed that the estimated ETs using the disaggregated LST from our model were more accurate than those using the disaggregated LST from DisTrad and TsHARP.

Keywords: Bowen ration, downscaling, evapotranspiration, land surface temperature

Procedia PDF Downloads 329
527 Ground State Phases in Two-Mode Quantum Rabi Models

Authors: Suren Chilingaryan

Abstract:

We study two models describing a single two-level system coupled to two boson field modes in either a parallel or orthogonal setup. Both models may be feasible for experimental realization through Raman adiabatic driving in cavity QED. We study their ground state configurations; that is, we find the quantum precursors of the corresponding semi-classical phase transitions. We found that the ground state configurations of both models present the same critical coupling as the quantum Rabi model. Around this critical coupling, the ground state goes from the so-called normal configuration with no excitation, the qubit in the ground state and the fields in the quantum vacuum state, to a ground state with excitations, the qubit in a superposition of ground and excited state, while the fields are not in the vacuum anymore, for the first model. The second model shows a more complex ground state configuration landscape where we find the normal configuration mentioned above, two single-mode configurations, where just one of the fields and the qubit are excited, and a dual-mode configuration, where both fields and the qubit are excited.

Keywords: quantum optics, quantum phase transition, cavity QED, circuit QED

Procedia PDF Downloads 367
526 The Use of Artificial Intelligence to Harmonization in the Lawmaking Process

Authors: Supriyadi, Andi Intan Purnamasari, Aminuddin Kasim, Sulbadana, Mohammad Reza

Abstract:

The development of the Industrial Revolution Era 4.0 brought a significant influence in the administration of countries in all parts of the world, including Indonesia, not only in the administration and economic sectors but the ways and methods of forming laws should also be adjusted. Until now, the process of making laws carried out by the Parliament with the Government still uses the classical method. The law-making process still uses manual methods, such as typing harmonization of regulations, so that it is not uncommon for errors to occur, such as writing errors, copying articles and so on, things that require a high level of accuracy and relying on inventory and harmonization carried out manually by humans. However, this method often creates several problems due to errors and inaccuracies on the part of officers who harmonize laws after discussion and approval; this has a very serious impact on the system of law formation in Indonesia. The use of artificial intelligence in the process of forming laws seems to be justified and becomes the answer in order to minimize the disharmony of various laws and regulations. This research is normative research using the Legislative Approach and the Conceptual Approach. This research focuses on the question of how to use Artificial Intelligence for Harmonization in the Lawmaking Process.

Keywords: artificial intelligence, harmonization, laws, intelligence

Procedia PDF Downloads 161
525 Monetary Policy and Economic Growth in West African Business Cycles: Markov Switching Approach

Authors: Omolade Adeleke, Jonathan Olusegun Famoroti

Abstract:

This study empirically examined the monetary policy and economic growth in the classical cycles in 8 member countries of the West African Economic and Monetary Union (WAEMU), using the Markov switching model for the Two-phase Regime, covering the period 1980Q1 to 2020Q4. Our estimates suggest that these countries demonstrate to have similar business cycles, and the economies stay more in an expansion regime than a recession regime. The result further shows that the union has an average duration period of 3.1 and 15.9 quarters for contraction and expansion periods, respectively. The business cycle duration, on average, suggests 19 quarters, varying from country to country. Therefore, the formulation of policies that can enhance aggregate demand by member countries in the union is an antidote for recession and is necessary to drive the economy into equilibrium. Also, a low-interest rate and reduced inflation rate would ginger long-run economic growth.

Keywords: monetary policy, business cycle, economic growth, Markov switching

Procedia PDF Downloads 75