Search results for: uses and functions of music
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3029

Search results for: uses and functions of music

2219 Bending the Consciousnesses: Uncovering Environmental Issues Through Circuit Bending

Authors: Enrico Dorigatti

Abstract:

The growing pile of hazardous e-waste produced especially by those developed and wealthy countries gets relentlessly bigger, composed of the EEDs (Electric and Electronic Device) that are often thrown away although still well functioning, mainly due to (programmed) obsolescence. As a consequence, e-waste has taken, over the last years, the shape of a frightful, uncontrollable, and unstoppable phenomenon, mainly fuelled by market policies aiming to maximize sales—and thus profits—at any cost. Against it, governments and organizations put some efforts in developing ambitious frameworks and policies aiming to regulate, in some cases, the whole lifecycle of EEDs—from the design to the recycling. Incidentally, however, such regulations sometimes make the disposal of the devices economically unprofitable, which often translates into growing illegal e-waste trafficking—an activity usually undertaken by criminal organizations. It seems that nothing, at least in the near future, can stop the phenomenon of e-waste production and accumulation. But while, from a practical standpoint, a solution seems hard to find, much can be done regarding people's education, which translates into informing and promoting good practices such as reusing and repurposing. This research argues that circuit bending—an activity rooted in neo-materialist philosophy and post-digital aesthetic, and based on repurposing EEDs into novel music instruments and sound generators—could have a great potential in this. In particular, it asserts that circuit bending could expose ecological, environmental, and social criticalities related to the current market policies and economic model. Not only thanks to its practical side (e.g., sourcing and repurposing devices) but also to the artistic one (e.g., employing bent instruments for ecological-aware installations, performances). Currently, relevant literature and debate lack interest and information about the ecological aspects and implications of the practical and artistic sides of circuit bending. This research, therefore, although still at an early stage, aims to fill in this gap by investigating, on the one side, the ecologic potential of circuit bending and, on the other side, its capacity of sensitizing people, through artistic practice, about e-waste-related issues. The methodology will articulate in three main steps. Firstly, field research will be undertaken—with the purpose of understanding where and how to source, in an ecologic and sustainable way, (discarded) EEDs for circuit bending. Secondly, artistic installations and performances will be organized—to sensitize the audience about environmental concerns through sound art and music derived from bent instruments. Data, such as audiences' feedback, will be collected at this stage. The last step will consist in realising workshops to spread an ecologically-aware circuit bending practice. Additionally, all the data and findings collected will be made available and disseminated as resources.

Keywords: circuit bending, ecology, sound art, sustainability

Procedia PDF Downloads 171
2218 Analyzing the Construction of Collective Memories by History Movies/TV Programs: Case Study of Masters in the Forbidden City

Authors: Lulu Wang, Yongjun Xu, Xiaoyang Qiao

Abstract:

The Forbidden City is well known for being full of Chinese cultural and historical relics. However, the Masters in the Forbidden City, a documentary film, doesn’t just dwell on the stories of the past. Instead, it focuses on ordinary people—the restorers of the relics and antiquities, which has caught the sight of Chinese audiences. From this popular documentary film, a new way can be considered, that is to show the relics, antiquities and painting with a character of modern humanities by films and TV programs. Of course, it can’t just like a simple explanation from tour guides in museums. It should be a perfect combination of scenes, heritages, stories, storytellers and background music. All we want to do is trying to dig up the humanity behind the heritages and then create a virtual scene for the audience to have emotional resonance from the humanity. It is believed that there are two problems. One is that compared with the entertainment shows, why people prefer to see the boring restoration work. The other is that what the interaction is between those history documentary films, the heritages, the audiences and collective memory. This paper mainly used the methods of text analysis and data analysis. The audiences’ comment texts were collected from all kinds of popular video sites. Through analyzing those texts, there was a word cloud chart about people preferring to use what kind of words to comment the film. Then the usage rate of all comments words was calculated. After that, there was a Radar Chart to show the rank results. Eventually, each of them was given an emotional value classification according their comment tone and content. Based on the above analysis results, an interaction model among the audience, history films/TV programs and the collective memory can be summarized. According to the word cloud chart, people prefer to use such words to comment, including moving, history, love, family, celebrity, tone... From those emotional words, we can see Chinese audience felt so proud and shared the sense of Collective Identity, so they leave such comments: To our great motherland! Chinese traditional culture is really profound! It is found that in the construction of collective memory symbology, the films formed an imaginary system by organizing a ‘personalized audience’. The audience is not just a recipient of information, but a participant of the documentary films and a cooperator of collective memory. At the same time, it is believed that the traditional background music, the spectacular present scenes and the tone of the storytellers/hosts are also important, so it is suggested that the museums could try to cooperate with the producers of movie and TV program to create a vivid scene for the people. Maybe it’s a more artistic way for heritages to be open to all the world.

Keywords: audience, heritages, history movies, TV programs

Procedia PDF Downloads 161
2217 Forecasting Market Share of Electric Vehicles in Taiwan Using Conjoint Models and Monte Carlo Simulation

Authors: Li-hsing Shih, Wei-Jen Hsu

Abstract:

Recently, the sale of electrical vehicles (EVs) has increased dramatically due to maturing technology development and decreasing cost. Governments of many countries have made regulations and policies in favor of EVs due to their long-term commitment to net zero carbon emissions. However, due to uncertain factors such as the future price of EVs, forecasting the future market share of EVs is a challenging subject for both the auto industry and local government. This study tries to forecast the market share of EVs using conjoint models and Monte Carlo simulation. The research is conducted in three phases. (1) A conjoint model is established to represent the customer preference structure on purchasing vehicles while five product attributes of both EV and internal combustion engine vehicles (ICEV) are selected. A questionnaire survey is conducted to collect responses from Taiwanese consumers and estimate the part-worth utility functions of all respondents. The resulting part-worth utility functions can be used to estimate the market share, assuming each respondent will purchase the product with the highest total utility. For example, attribute values of an ICEV and a competing EV are given respectively, two total utilities of the two vehicles of a respondent are calculated and then knowing his/her choice. Once the choices of all respondents are known, an estimate of market share can be obtained. (2) Among the attributes, future price is the key attribute that dominates consumers’ choice. This study adopts the assumption of a learning curve to predict the future price of EVs. Based on the learning curve method and past price data of EVs, a regression model is established and the probability distribution function of the price of EVs in 2030 is obtained. (3) Since the future price is a random variable from the results of phase 2, a Monte Carlo simulation is then conducted to simulate the choices of all respondents by using their part-worth utility functions. For instance, using one thousand generated future prices of an EV together with other forecasted attribute values of the EV and an ICEV, one thousand market shares can be obtained with a Monte Carlo simulation. The resulting probability distribution of the market share of EVs provides more information than a fixed number forecast, reflecting the uncertain nature of the future development of EVs. The research results can help the auto industry and local government make more appropriate decisions and future action plans.

Keywords: conjoint model, electrical vehicle, learning curve, Monte Carlo simulation

Procedia PDF Downloads 69
2216 Nonparametric Truncated Spline Regression Model on the Data of Human Development Index in Indonesia

Authors: Kornelius Ronald Demu, Dewi Retno Sari Saputro, Purnami Widyaningsih

Abstract:

Human Development Index (HDI) is a standard measurement for a country's human development. Several factors may have influenced it, such as life expectancy, gross domestic product (GDP) based on the province's annual expenditure, the number of poor people, and the percentage of an illiterate people. The scatter plot between HDI and the influenced factors show that the plot does not follow a specific pattern or form. Therefore, the HDI's data in Indonesia can be applied with a nonparametric regression model. The estimation of the regression curve in the nonparametric regression model is flexible because it follows the shape of the data pattern. One of the nonparametric regression's method is a truncated spline. Truncated spline regression is one of the nonparametric approach, which is a modification of the segmented polynomial functions. The estimator of a truncated spline regression model was affected by the selection of the optimal knots point. Knot points is a focus point of spline truncated functions. The optimal knots point was determined by the minimum value of generalized cross validation (GCV). In this article were applied the data of Human Development Index with a truncated spline nonparametric regression model. The results of this research were obtained the best-truncated spline regression model to the HDI's data in Indonesia with the combination of optimal knots point 5-5-5-4. Life expectancy and the percentage of an illiterate people were the significant factors depend to the HDI in Indonesia. The coefficient of determination is 94.54%. This means the regression model is good enough to applied on the data of HDI in Indonesia.

Keywords: generalized cross validation (GCV), Human Development Index (HDI), knots point, nonparametric regression, truncated spline

Procedia PDF Downloads 339
2215 A Study of Algebraic Structure Involving Banach Space through Q-Analogue

Authors: Abdul Hakim Khan

Abstract:

The aim of the present paper is to study the Banach Space and Combinatorial Algebraic Structure of R. It is further aimed to study algebraic structure of set of all q-extension of classical formula and function for 0 < q < 1.

Keywords: integral functions, q-extensions, q numbers of metric space, algebraic structure of r and banach space

Procedia PDF Downloads 579
2214 Integration of STEM Education in Quebec, Canada – Challenges and Opportunities

Authors: B. El Fadil, R. Najar

Abstract:

STEM education is promoted by many scholars and curricula around the world, but it is not yet well established in the province of Quebec in Canada. In addition, effective instructional STEM activities and design methods are required to ensure that students and teachers' needs are being met. One potential method is the Engineering Design Process (EDP), a methodology that emphasizes the importance of creativity and collaboration in problem-solving strategies. This article reports on a case study that focused on using the EDP to develop instructional materials by means of making a technological artifact to teach mathematical variables and functions at the secondary level. The five iterative stages of the EDP (design, make, test, infer, and iterate) were integrated into the development of the course materials. Data was collected from different sources: pre- and post-questionnaires, as well as a working document dealing with pupils' understanding based on designing, making, testing, and simulating. Twenty-four grade seven (13 years old) students in Northern Quebec participated in the study. The findings of this study indicate that STEM activities have a positive impact not only on students' engagement in classroom activities but also on learning new mathematical concepts. Furthermore, STEM-focused activities have a significant effect on problem-solving skills development in an interdisciplinary approach. Based on the study's results, we can conclude, inter alia, that teachers should integrate STEM activities into their teaching practices to increase learning outcomes and attach more importance to STEM-focused activities to develop students' reflective thinking and hands-on skills.

Keywords: engineering design process, motivation, stem, integration, variables, functions

Procedia PDF Downloads 88
2213 [Keynote Talk]: Applying p-Balanced Energy Technique to Solve Liouville-Type Problems in Calculus

Authors: Lina Wu, Ye Li, Jia Liu

Abstract:

We are interested in solving Liouville-type problems to explore constancy properties for maps or differential forms on Riemannian manifolds. Geometric structures on manifolds, the existence of constancy properties for maps or differential forms, and energy growth for maps or differential forms are intertwined. In this article, we concentrate on discovery of solutions to Liouville-type problems where manifolds are Euclidean spaces (i.e. flat Riemannian manifolds) and maps become real-valued functions. Liouville-type results of vanishing properties for functions are obtained. The original work in our research findings is to extend the q-energy for a function from finite in Lq space to infinite in non-Lq space by applying p-balanced technique where q = p = 2. Calculation skills such as Hölder's Inequality and Tests for Series have been used to evaluate limits and integrations for function energy. Calculation ideas and computational techniques for solving Liouville-type problems shown in this article, which are utilized in Euclidean spaces, can be universalized as a successful algorithm, which works for both maps and differential forms on Riemannian manifolds. This innovative algorithm has a far-reaching impact on research work of solving Liouville-type problems in the general settings involved with infinite energy. The p-balanced technique in this algorithm provides a clue to success on the road of q-energy extension from finite to infinite.

Keywords: differential forms, holder inequality, Liouville-type problems, p-balanced growth, p-harmonic maps, q-energy growth, tests for series

Procedia PDF Downloads 235
2212 R Statistical Software Applied in Reliability Analysis: Case Study of Diesel Generator Fans

Authors: Jelena Vucicevic

Abstract:

Reliability analysis represents a very important task in different areas of work. In any industry, this is crucial for maintenance, efficiency, safety and monetary costs. There are ways to calculate reliability, unreliability, failure density and failure rate. This paper will try to introduce another way of calculating reliability by using R statistical software. R is a free software environment for statistical computing and graphics. It compiles and runs on a wide variety of UNIX platforms, Windows and MacOS. The R programming environment is a widely used open source system for statistical analysis and statistical programming. It includes thousands of functions for the implementation of both standard and new statistical methods. R does not limit user only to operation related only to these functions. This program has many benefits over other similar programs: it is free and, as an open source, constantly updated; it has built-in help system; the R language is easy to extend with user-written functions. The significance of the work is calculation of time to failure or reliability in a new way, using statistic. Another advantage of this calculation is that there is no need for technical details and it can be implemented in any part for which we need to know time to fail in order to have appropriate maintenance, but also to maximize usage and minimize costs. In this case, calculations have been made on diesel generator fans but the same principle can be applied to any other part. The data for this paper came from a field engineering study of the time to failure of diesel generator fans. The ultimate goal was to decide whether or not to replace the working fans with a higher quality fan to prevent future failures. Seventy generators were studied. For each one, the number of hours of running time from its first being put into service until fan failure or until the end of the study (whichever came first) was recorded. Dataset consists of two variables: hours and status. Hours show the time of each fan working and status shows the event: 1- failed, 0- censored data. Censored data represent cases when we cannot track the specific case, so it could fail or success. Gaining the result by using R was easy and quick. The program will take into consideration censored data and include this into the results. This is not so easy in hand calculation. For the purpose of the paper results from R program have been compared to hand calculations in two different cases: censored data taken as a failure and censored data taken as a success. In all three cases, results are significantly different. If user decides to use the R for further calculations, it will give more precise results with work on censored data than the hand calculation.

Keywords: censored data, R statistical software, reliability analysis, time to failure

Procedia PDF Downloads 401
2211 Musical Instrument Recognition in Polyphonic Audio Through Convolutional Neural Networks and Spectrograms

Authors: Rujia Chen, Akbar Ghobakhlou, Ajit Narayanan

Abstract:

This study investigates the task of identifying musical instruments in polyphonic compositions using Convolutional Neural Networks (CNNs) from spectrogram inputs, focusing on binary classification. The model showed promising results, with an accuracy of 97% on solo instrument recognition. When applied to polyphonic combinations of 1 to 10 instruments, the overall accuracy was 64%, reflecting the increasing challenge with larger ensembles. These findings contribute to the field of Music Information Retrieval (MIR) by highlighting the potential and limitations of current approaches in handling complex musical arrangements. Future work aims to include a broader range of musical sounds, including electronic and synthetic sounds, to improve the model's robustness and applicability in real-time MIR systems.

Keywords: binary classifier, CNN, spectrogram, instrument

Procedia PDF Downloads 78
2210 Hydrological Modelling of Geological Behaviours in Environmental Planning for Urban Areas

Authors: Sheetal Sharma

Abstract:

Runoff,decreasing water levels and recharge in urban areas have been a complex issue now a days pointing defective urban design and increasing demography as cause. Very less has been discussed or analysed for water sensitive Urban Master Plans or local area plans. Land use planning deals with land transformation from natural areas into developed ones, which lead to changes in natural environment. Elaborated knowledge of relationship between the existing patterns of land use-land cover and recharge with respect to prevailing soil below is less as compared to speed of development. The parameters of incompatibility between urban functions and the functions of the natural environment are becoming various. Changes in land patterns due to built up, pavements, roads and similar land cover affects surface water flow seriously. It also changes permeability and absorption characteristics of the soil. Urban planners need to know natural processes along with modern means and best technologies available,as there is a huge gap between basic knowledge of natural processes and its requirement for balanced development planning leading to minimum impact on water recharge. The present paper analyzes the variations in land use land cover and their impacts on surface flows and sub-surface recharge in study area. The methodology adopted was to analyse the changes in land use and land cover using GIS and Civil 3d auto cad. The variations were used in  computer modeling using Storm-water Management Model to find out the runoff for various soil groups and resulting recharge observing water levels in POW data for last 40 years of the study area. Results were anlayzed again to find best correlations for sustainable recharge in urban areas.

Keywords: geology, runoff, urban planning, land use-land cover

Procedia PDF Downloads 317
2209 Life Prediction Method of Lithium-Ion Battery Based on Grey Support Vector Machines

Authors: Xiaogang Li, Jieqiong Miao

Abstract:

As for the problem of the grey forecasting model prediction accuracy is low, an improved grey prediction model is put forward. Firstly, use trigonometric function transform the original data sequence in order to improve the smoothness of data , this model called SGM( smoothness of grey prediction model), then combine the improved grey model with support vector machine , and put forward the grey support vector machine model (SGM - SVM).Before the establishment of the model, we use trigonometric functions and accumulation generation operation preprocessing data in order to enhance the smoothness of the data and weaken the randomness of the data, then use support vector machine (SVM) to establish a prediction model for pre-processed data and select model parameters using genetic algorithms to obtain the optimum value of the global search. Finally, restore data through the "regressive generate" operation to get forecasting data. In order to prove that the SGM-SVM model is superior to other models, we select the battery life data from calce. The presented model is used to predict life of battery and the predicted result was compared with that of grey model and support vector machines.For a more intuitive comparison of the three models, this paper presents root mean square error of this three different models .The results show that the effect of grey support vector machine (SGM-SVM) to predict life is optimal, and the root mean square error is only 3.18%. Keywords: grey forecasting model, trigonometric function, support vector machine, genetic algorithms, root mean square error

Keywords: Grey prediction model, trigonometric functions, support vector machines, genetic algorithms, root mean square error

Procedia PDF Downloads 461
2208 ReS, Resonant String Shell: Development of an Acoustic Shell for Outdoor Chamber Music Concerts

Authors: Serafino Di Rosario

Abstract:

ReS is a sustainable hand-built temporary acoustic shell, developed since 2011 and built during the architectural workshop at Villa Pennisi in Musica in Acireale, Sicily, each year since 2012. The design concept aims to provide a portable structure by reducing the on-site construction problems and the skills required by the builders together with maximizing the acoustic performance for the audience and the musicians. The shell is built using only wood, recycled for the most part, and can be built and dismantled by non-specialized workers in just three days. This paper describes the research process, which spans over four years and presents the final results in form of acoustic simulations performed by acoustic modeling software and real world measurements. ReS is developed by the ReS team who has been presented with the Peter Lord Award in 2015 by the Institute of Acoustics in the UK.

Keywords: acoustic shell, outdoor natural amplification, computational design, room acoustics

Procedia PDF Downloads 228
2207 Simulation-Based Control Module for Offshore Single Point Mooring System

Authors: Daehyun Baek, Seungmin Lee, Minju Kim Jangik Park, Hyeong-Soon Moon

Abstract:

SPM (Single Point Mooring) is one of the mooring buoy facilities installed on a coast near oil and gas terminal which is not able to berth FPSO or large oil tankers under the condition of high draft due to geometrical limitation. Loading and unloading of crude oil and gas through a subsea pipeline can be carried out between the mooring buoy, ships and onshore facilities. SPM is an offshore-standalone system which has to withstand the harsh marine environment with harsh conditions such as high wind, current and so on. Therefore, SPM is required to have high stability, reliability and durability. Also, SPM is comprised to be integrated systems which consist of power management, high pressure valve control, sophisticated hardware/software and a long distance communication system. In order to secure required functions of SPM system, a simulation model for the integrated system of SPM using MATLAB Simulink and State flow tool has been developed. The developed model consists of configuration of hydraulic system for opening and closing of PLEM (Pipeline End Manifold) valves and control system logic. To verify functions of the model, an integrated simulation model for overall systems of SPM was also developed by considering handshaking variables between individual systems. In addition to the dynamic model, a self-diagnostic function to determine failure of the system was configured, which enables the SPM system itself to alert users about the failure once a failure signal comes to arise. Controlling and monitoring the SPM system is able to be done by a HMI system which is capable of managing the SPM system remotely, which was carried out by building a communication environment between the SPM system and the HMI system.

Keywords: HMI system, mooring buoy, simulink simulation model, single point mooring, stateflow

Procedia PDF Downloads 417
2206 Feasibility of an Extreme Wind Risk Assessment Software for Industrial Applications

Authors: Francesco Pandolfi, Georgios Baltzopoulos, Iunio Iervolino

Abstract:

The impact of extreme winds on industrial assets and the built environment is gaining increasing attention from stakeholders, including the corporate insurance industry. This has led to a progressively more in-depth study of building vulnerability and fragility to wind. Wind vulnerability models are used in probabilistic risk assessment to relate a loss metric to an intensity measure of the natural event, usually a gust or a mean wind speed. In fact, vulnerability models can be integrated with the wind hazard, which consists of associating a probability to each intensity level in a time interval (e.g., by means of return periods) to provide an assessment of future losses due to extreme wind. This has also given impulse to the world- and regional-scale wind hazard studies.Another approach often adopted for the probabilistic description of building vulnerability to the wind is the use of fragility functions, which provide the conditional probability that selected building components will exceed certain damage states, given wind intensity. In fact, in wind engineering literature, it is more common to find structural system- or component-level fragility functions rather than wind vulnerability models for an entire building. Loss assessment based on component fragilities requires some logical combination rules that define the building’s damage state given the damage state of each component and the availability of a consequence model that provides the losses associated with each damage state. When risk calculations are based on numerical simulation of a structure’s behavior during extreme wind scenarios, the interaction of component fragilities is intertwined with the computational procedure. However, simulation-based approaches are usually computationally demanding and case-specific. In this context, the present work introduces the ExtReMe wind risk assESsment prototype Software, ERMESS, which is being developed at the University of Naples Federico II. ERMESS is a wind risk assessment tool for insurance applications to industrial facilities, collecting a wide assortment of available wind vulnerability models and fragility functions to facilitate their incorporation into risk calculations based on in-built or user-defined wind hazard data. This software implements an alternative method for building-specific risk assessment based on existing component-level fragility functions and on a number of simplifying assumptions for their interactions. The applicability of this alternative procedure is explored by means of an illustrative proof-of-concept example, which considers four main building components, namely: the roof covering, roof structure, envelope wall and envelope openings. The application shows that, despite the simplifying assumptions, the procedure can yield risk evaluations that are comparable to those obtained via more rigorous building-level simulation-based methods, at least in the considered example. The advantage of this approach is shown to lie in the fact that a database of building component fragility curves can be put to use for the development of new wind vulnerability models to cover building typologies not yet adequately covered by existing works and whose rigorous development is usually beyond the budget of portfolio-related industrial applications.

Keywords: component wind fragility, probabilistic risk assessment, vulnerability model, wind-induced losses

Procedia PDF Downloads 181
2205 Dominican Representation in Introductory Level Spanish Textbooks in the United States

Authors: Sheridan Wigginton

Abstract:

This research project investigates the representation of Dominicans and Dominican culture in nine widely-used Spanish textbooks in universities and colleges in the United States. The project uses the “culture” standard established by the American Council on the Teaching of Foreign Languages to examine the texts. The organization, commonly referred to by its acronym ACTFL, describes products as books, tools, foods, laws, music, and games; practices as patterns of social interactions; and perspectives as meanings, attitudes, values, and ideas. The content analysis of the texts will also specifically include visual analysis of the physical representation of the people depicted in Dominican-themed culture activities to more clearly integrate issues of color and national identity into the discussion.

Keywords: blackness, culture, Dominican republic, foreign language education, national identity, Spanish, textbooks

Procedia PDF Downloads 89
2204 Multiple Intelligences to Improve Pronunciation

Authors: Jean Pierre Ribeiro Daquila

Abstract:

This paper aims to analyze the use of the Theory of Multiple Intelligences as a tool to facilitate students’ learning. This theory, proposed by the American psychologist and educator Howard Gardner, was first established in 1983 and advocates that human beings possess eight intelligence and not only one, as defended by psychologists prior to his theory. These intelligence are bodily-kinesthetic intelligence, musical, linguistic, logical-mathematical, spatial, interpersonal, intrapersonal, and naturalist. This paper will focus on bodily-kinesthetic intelligence. Spatial and bodily-kinesthetic intelligences are sensed by athletes, dancers, and others who use their bodies in ways that exceed normal abilities. These are intelligences that are closely related. A quarterback or a ballet dancer needs to have both an awareness of body motions and abilities as well as a sense of the space involved in the action. Nevertheless, there are many reasons which make classical ballet dance more integrated with other intelligences. Ballet dancers make it look effortless as they move across the stage, from the lifts to the toe points; therefore, there is acting both in the performance of the repertoire and in hiding the pain or physical stress. The ballet dancer has to have great mathematical intelligence to perform a fast allegro; for instance, each movement has to be executed in a specific millisecond. Flamenco dancers need to rely as well on their mathematic abilities, as the footwork requires the ability to make half, two, three, four or even six movements in just one beat. However, the precision of the arm movements is freer than in ballet dance; for this reason, ballet dancers need to be more holistically aware of their movements; therefore, our experiment will test whether this greater attention required by ballet dancers makes them acquire better results in the training sessions when compared to flamenco dancers. An experiment will be carried out in this study by training ballet dancers through dance (four years of experience dancing minimum – experimental group 1); a group of flamenco dancers (four years of experience dancing minimum – experimental group 2). Both experimental groups will be trained in two different domains – phonetics and chemistry – to examine whether there is a significant improvement in these areas compared to the control group (a group of regular students who will receive the same training through a traditional method). However, this paper will focus on phonetic training. Experimental group 1 will be trained with the aid of classical music plus bodily work. Experimental group 2 will be trained with flamenco rhythm and kinesthetic work. We would like to highlight that this study takes dance as an example of a possible area of strength; nonetheless, other types of arts can and should be used to support students, such as drama, creative writing, music and others. The main aim of this work is to suggest that other intelligences, in the case of this study, bodily-kinesthetic, can be used to help improve pronunciation.

Keywords: multiple intelligences, pronunciation, effective pronunciation trainings, short drills, musical intelligence, bodily-kinesthetic intelligence

Procedia PDF Downloads 96
2203 Functional Aspects of Carbonic Anhydrase

Authors: Bashistha Kumar Kanth, Seung Pil Pack

Abstract:

Carbonic anhydrase is ubiquitously distributed in organisms, and is fundamental to many eukaryotic biological processes such as photosynthesis, respiration, CO2 and ion transport, calcification and acid–base balance. However, CA occurs across the spectrum of prokaryotic metabolism in both the archaea and bacteria domains and many individual species contain more than one class. In this review, various roles of CA involved in cellular mechanism are presented to find out the CA functions applicable for industrial use.

Keywords: carbonic anhydrase, mechanism, CO2 sequestration, respiration

Procedia PDF Downloads 491
2202 Strengthening Regulation and Supervision of Microfinance Sector for Development in Ethiopia

Authors: Megersa Dugasa Fite

Abstract:

This paper analyses regulatory and supervisory issues in the Ethiopian micro finance sector, which caters to the needs of those who have been excluded from the formal financial sector. Micro-finance has received increased importance in development because of its grand goal to give credits to the poor to raise their economic and social well-being and improve the quality of lives. The micro-finance at present has been moving towards a credit-plus period through covering savings and insurance functions. It thus helps in reducing the rate of financial exclusion and social segregation, alleviating poverty and, consequently, stimulating development. The Ethiopian micro finance policy has been generally positive and developmental but major regulatory and supervisory limitations such as the absolute prohibition of NGOs to participate in micro credit functions, higher risks for depositors of micro-finance institutions, lack of credit information services with research and development, the unmet demand, and risks of market failures due to over-regulation are disappointing. Therefore, to remove the limited reach and high degree of problems typical in the informal means of financial intermediation plus to deal with the failure of formal banks to provide basic financial services to a significant portion of the country’s population, more needs to be done on micro finance. Certain key regulatory and supervisory revisions hence need to be taken to strengthen the Ethiopian micro finance sector so that it can practically provide majority poor access to a range of high quality financial services that help them work their way out of poverty and the incapacity it imposes.

Keywords: micro-finance, micro-finance regulation and supervision, micro-finance institutions, financial access, social segregation, poverty alleviation, development, Ethiopia

Procedia PDF Downloads 395
2201 Decision Support System Based On GIS and MCDM to Identify Land Suitability for Agriculture

Authors: Abdelkader Mendas

Abstract:

The integration of MultiCriteria Decision Making (MCDM) approaches in a Geographical Information System (GIS) provides a powerful spatial decision support system which offers the opportunity to efficiently produce the land suitability maps for agriculture. Indeed, GIS is a powerful tool for analyzing spatial data and establishing a process for decision support. Because of their spatial aggregation functions, MCDM methods can facilitate decision making in situations where several solutions are available, various criteria have to be taken into account and decision-makers are in conflict. The parameters and the classification system used in this work are inspired from the FAO (Food and Agriculture Organization) approach dedicated to a sustainable agriculture. A spatial decision support system has been developed for establishing the land suitability map for agriculture. It incorporates the multicriteria analysis method ELECTRE Tri (ELimitation Et Choix Traduisant la REalité) in a GIS within the GIS program package environment. The main purpose of this research is to propose a conceptual and methodological framework for the combination of GIS and multicriteria methods in a single coherent system that takes into account the whole process from the acquisition of spatially referenced data to decision-making. In this context, a spatial decision support system for developing land suitability maps for agriculture has been developed. The algorithm of ELECTRE Tri is incorporated into a GIS environment and added to the other analysis functions of GIS. This approach has been tested on an area in Algeria. A land suitability map for durum wheat has been produced. Through the obtained results, it appears that ELECTRE Tri method, integrated into a GIS, is better suited to the problem of land suitability for agriculture. The coherence of the obtained maps confirms the system effectiveness.

Keywords: multicriteria decision analysis, decision support system, geographical information system, land suitability for agriculture

Procedia PDF Downloads 638
2200 Spatiotemporal Variation Characteristics of Soil pH around the Balikesir City, Turkey

Authors: Çağan Alevkayali, Şermin Tağil

Abstract:

Determination of soil pH surface distribution in urban areas is substantial for sustainable development. Changes on soil properties occur due to functions on performed in agriculture, industry and other urban functions. Soil pH is important to effect on soil productivity which based on sensitive and complex relation between plant and soil. Furthermore, the spatial variability of soil reaction is necessary to measure the effects of urbanization. The objective of this study was to explore the spatial variation of soil pH quality and the influence factors of human land use on soil Ph around Balikesir City using data for 2015 and Geographic Information Systems (GIS). For this, soil samples were taken from 40 different locations, and collected with the method of "Systematic Random" from the pits at 0-20 cm depths, because anthropologic sourced pollutants accumulate on upper layers of soil. The study area was divided into a grid system with 750 x 750 m. GPS was used to determine sampling locations, and Inverse Distance Weighting (IDW) interpolation technique was used to analyze the spatial distribution of pH in the study area and to predict the variable values of un-exampled places with the help from the values of exampled places. Natural soil acidity and alkalinity depend on interaction between climate, vegetation, and soil geological properties. However, analyzing soil pH is important to indirectly evaluate soil pollution caused by urbanization and industrialization. The result of this study showed that soil pH around the Balikesir City was neutral, in generally, with values were between 6.5 and 7.0. On the other hand, some slight changes were demonstrated around open dump areas and the small industrial sites. The results obtained from this study can be indicator of important soil problems and this data can be used by ecologists, planners and managers to protect soil supplies around the Balikesir City.

Keywords: Balikesir, IDW, GIS, spatial variability, soil pH, urbanization

Procedia PDF Downloads 322
2199 Fast Switching Mechanism for Multicasting Failure in OpenFlow Networks

Authors: Alaa Allakany, Koji Okamura

Abstract:

Multicast technology is an efficient and scalable technology for data distribution in order to optimize network resources. However, in the IP network, the responsibility for management of multicast groups is distributed among network routers, which causes some limitations such as delays in processing group events, high bandwidth consumption and redundant tree calculation. Software Defined Networking (SDN) represented by OpenFlow presented as a solution for many problems, in SDN the control plane and data plane are separated by shifting the control and management to a remote centralized controller, and the routers are used as a forwarder only. In this paper we will proposed fast switching mechanism for solving the problem of link failure in multicast tree based on Tabu Search heuristic algorithm and modifying the functions of OpenFlow switch to fasts switch to the pack up sub tree rather than sending to the controller. In this work we will implement multicasting OpenFlow controller, this centralized controller is a core part in our multicasting approach, which is responsible for 1- constructing the multicast tree, 2- handling the multicast group events and multicast state maintenance. And finally modifying OpenFlow switch functions for fasts switch to pack up paths. Forwarders, forward the multicast packet based on multicast routing entries which were generated by the centralized controller. Tabu search will be used as heuristic algorithm for construction near optimum multicast tree and maintain multicast tree to still near optimum in case of join or leave any members from multicast group (group events).

Keywords: multicast tree, software define networks, tabu search, OpenFlow

Procedia PDF Downloads 263
2198 Selection Criteria in the Spanish Secondary Education Content and Language Integrated Learning (CLIL) Programmes and Their Effect on Code-Switching in CLIL Methodology

Authors: Dembele Dembele, Philippe

Abstract:

Several Second Language Acquisition (SLA) studies have stressed the benefits of Content and Language Integrated Learning (CLIL) and shown how CLIL students outperformed their non-CLIL counterparts in many L2 skills. However, numerous experimental CLIL programs seem to have mainly targeted above-average and rather highly motivated language learners. The need to understand the impact of the student’s language proficiency on code-switching in CLIL instruction motivated this study. Therefore, determining the implications of the students’ low-language proficiency for CLIL methodology, as well as the frequency with which CLIL teachers use the main pedagogical functions of code-switching, seemed crucial for a Spanish CLIL instruction on a large scale. In the mixed-method approach adopted, ten face-to-face interviews were conducted in nine Valencian public secondary education schools, while over 30 CLIL teachers also contributed with their experience in two online survey questionnaires. The results showed the crucial role language proficiency plays in the Valencian CLIL/Plurilingual selection criteria. The presence of a substantial number of low-language proficient students in CLIL groups, which in turn implied important methodological consequences, was another finding of the study. Indeed, though the pedagogical use of L1 was confirmed as an extended practice among CLIL teachers, more than half of the participants perceived that code-switching impaired attaining their CLIL lesson objectives. Therein, the dissertation highlights the need for more extensive empirical research on how code-switching could prove beneficial in CLIL instruction involving low-language proficient students while maintaining the maximum possible exposure to the target language.

Keywords: CLIL methodology, low language proficiency, code switching, selection criteria, code-switching functions

Procedia PDF Downloads 81
2197 Effectiveness of Raga Desi Todi on Depression, Anxiety and Stress Among Adults

Authors: Sushila Pareek, Divya Shekhawat

Abstract:

Music has been shown as a therapeutic agent in depression, anxiety, and stress. A pilot study was carried out to see the therapeutic effects of Indian classical instrumental Raga Todi on depression, anxiety, and stress. 50 individuals diagnosed with depression, anxiety, and stress with DSM-V were taken for the study. Subjects were randomly divided into two groups: the experimental group and the control group. The experimental group received the instrumental raga Todi whereas the other control group didn't receive any intervention. DASS-21 was used on the baseline and after the intervention to measure depression, anxiety, and stress. The result indicates that anxiety, stress, and depression level was reduced after listening to the raga desi Todi. It was concluded that raga desi Todi is an effective intervention for reducing depression, anxiety, and stress.

Keywords: raga, anxiety, stress, depression, DASS-21, mental health

Procedia PDF Downloads 144
2196 Cognitive Rehabilitation in Schizophrenia: A Review of the Indian Scenario

Authors: Garima Joshi, Pratap Sharan, V. Sreenivas, Nand Kumar, Kameshwar Prasad, Ashima N. Wadhawan

Abstract:

Schizophrenia is a debilitating disorder and is marked by cognitive impairment, which deleteriously impacts the social and professional functioning along with the quality of life of the patients and the caregivers. Often the cognitive symptoms are in their prodromal state and worsen as the illness progresses; they have proven to have a good predictive value for the prognosis of the illness. It has been shown that intensive cognitive rehabilitation (CR) leads to improvements in the healthy as well as cognitively-impaired subjects. As the majority of population in India falls in the lower to middle socio-economic status and have low education levels, using the existing packages, a majority of which are developed in the West, for cognitive rehabilitation becomes difficult. The use of technology is also restricted due to the high costs involved and the limited availability and familiarity with computers and other devices, which pose as an impedance for continued therapy. Cognitive rehabilitation in India uses a plethora of retraining methods for the patients with schizophrenia targeting the functions of attention, information processing, executive functions, learning and memory, and comprehension along with Social Cognition. Psychologists often have to follow an integrative therapy approach involving social skills training, family therapy and psychoeducation in order to maintain the gains from the cognitive rehabilitation in the long run. This paper reviews the methodologies and cognitive retaining programs used in India. It attempts to elucidate the evolution and development of methodologies used, from traditional paper-pencil based retraining to more sophisticated neuroscience-informed techniques in cognitive rehabilitation of deficits in schizophrenia as home-based or supervised and guided programs for cognitive rehabilitation.

Keywords: schizophrenia, cognitive rehabilitation, neuropsychological interventions, integrated approached to rehabilitation

Procedia PDF Downloads 363
2195 Enhancing Embedded System Efficiency with Digital Signal Processing Cores

Authors: Anil H. Dhanawade, Akshay S., Harshal M. Lakesar

Abstract:

This paper presents a comprehensive analysis of the performance advantages offered by DSP (Digital Signal Processing) cores compared to traditional MCU (Microcontroller Unit) cores in the execution of various functions critical to real-time applications. The focus is on the integration of DSP functionalities, specifically in the context of motor control applications such as Field-Oriented Control (FOC), trigonometric calculations, back-EMF estimation, digital filtering, and high-resolution PWM generation. Through comparative analysis, it is demonstrated that DSP cores significantly enhance processing efficiency, achieving faster execution times for complex mathematical operations essential for precise torque and speed control. The study highlights the capabilities of DSP cores, including single-cycle Multiply-Accumulate (MAC) operations and optimized hardware for trigonometric functions, which collectively reduce latency and improve real-time performance. In contrast, MCU cores, while capable of performing similar tasks, typically exhibit longer execution times due to reliance on software-based solutions and lack of dedicated hardware acceleration. The findings underscore the critical role of DSP cores in applications requiring high-speed processing and low-latency response, making them indispensable in the automotive, industrial, and robotics sectors. This work serves as a reference for future developments in embedded systems, emphasizing the importance of architecture choice in achieving optimal performance in demanding computational tasks.

Keywords: CPU core, DSP, assembly code, motor control

Procedia PDF Downloads 16
2194 Analysis of Adipose Tissue-Derived Mesenchymal Stem Cells under Atherosclerosis Microenvironment

Authors: Do Khanh Vy, Vuong Cat Khanh, Osamu Ohneda

Abstract:

During atherosclerosis (AS) progression, perivascular adipose tissue-derived mesenchymal stem cells (PVAT-MSCs) are exposed to the hypoxic environment due to the oxygenic deprivation which might influence the adipose tissue-derived mesenchymal stem cells (AT-MSCs) function. Additionally, it has been reported that the angiogenic ability of subcutaneous AT-MSCs (SAT-MSCs) was impaired in the AS patients. However, up to now, the effects of AS on the characteristics and function of PVAT-MSCs have not been clarified yet. In the present study, we analyzed the AS microenvironment effects on the characteristics and function of AT-MSCs. We found that there was no significant difference in cellular morphology and differentiation ability between SAT-MSCs and PVAT-MSCs in AS patients. However, the proliferation of AS-derived PVAT-MSCs was less than those of AS-derived SAT-MSCs. Importantly, the migration of AS-derived PVAT-MSCs was faster than AS-derived SAT-MSCs. Of note, AS-derived PVAT-MSCs showed the upregulation of SDF1, which is related to the homing, and VEGF, which is related to the angiogenesis compared to those of AS-derived SAT-MSCs. Consistent with these results, AS-derived PVAT-MSCs showed the higher ability to recruit EPCs and ECs than AS-derived SAT-MSCs. In addition, EPCs and ECs which cultured in the presence of AS-derived PVAT-MSC conditioned medium showed the higher angiogenic function of the tube formation compared to those cultured in AS-derived SAT-MSC conditioned medium. This result suggests that the higher paracrine effects of AS-derived PVAT-MSCs support the angiogenic function of the target cells. Our data showed the different characteristics and functions of AT-MSCs derived from different sources of tissues. Under the AS microenvironment, it seems that the characteristics and functions of PVAT-MSCs might reflect the progression of AS. Further study will be necessary to clarify the mechanism in the future.

Keywords: atherosclerosis, mesenchymal stem cells, perivascular adipose tissue, subcutaneous adipose tissue

Procedia PDF Downloads 161
2193 Analysis of Truck Drivers’ Distraction on Crash Risk

Authors: Samuel Nderitu Muchiri, Tracy Wangechi Maina

Abstract:

Truck drivers face a myriad of challenges in their profession. Enhancements in logistics effectiveness can be pivotal in propelling economic developments. The specific objective of the study was to assess the influence of driver distraction on crash risk. The study is significant as it elucidates best practices that truck drivers can embrace in an effort to enhance road safety. These include amalgamating behaviors that enable drivers to fruitfully execute multifaceted functions such as finding and following routes, evading collisions, monitoring speed, adhering to road regulations, and evaluating vehicle systems’ conditions. The analysis involved an empirical review of ten previous studies related to the research topic. The articles revealed that driver distraction plays a substantial role in road accidents and other crucial road security incidents across the globe. Africa depends immensely on the freight transport sector to facilitate supply chain operations. Several studies indicate that drivers who operate primarily on rural roads, such as those found in Sub-Saharan Africa, have an increased propensity to engage in distracted activities such as cell phone usage while driving. The findings also identified the need for digitalization in truck driving operations, including carrier management techniques such as fatigue management, artificial intelligence, and automating functions like cell phone usage controls. The recommendations can aid policymakers and commercial truck carriers in deepening their understanding of driver distraction and enforcing mitigations to foster road safety.

Keywords: truck drivers, distraction, digitalization, crash risk, road safety

Procedia PDF Downloads 49
2192 Mechanisms and Regulation of the Bi-directional Motility of Mitotic Kinesin Nano-motors

Authors: Larisa Gheber

Abstract:

Mitosis is an essential process by which duplicated genetic information is transmitted from mother to daughter cells. Incorrect chromosome segregation during mitosis can lead to genetic diseases, chromosome instability and cancer. This process is mediated by a dynamic microtubule-based intracellular structure, the mitotic spindle. One of the major factors that govern the mitotic spindle dynamics are the kinesin-5 biological nano motors that were believed to move unidirectionally on the microtubule filaments, using ATP hydrolysis, thus performing essential functions in mitotic spindle dynamics. Surprisingly, several reports from our and other laboratories have demonstrated that some kinesin-5 motors are bi-directional: they move in minus-end direction on the microtubules as single-molecules and can switch directionality under a number of conditions. These findings broke a twenty-five-years old dogma regarding kinesin directionality (1, 2). The mechanism of this bi-directional motility and its physiological significance remain unclear. To address this unresolved problem, we apply an interdisciplinary approach combining live cell imaging, biophysical single molecule, and structural experiments to examine the activity of these motors and their mutated variants in vivo and in vitro. Our data shows that factors such as protein phosphorylation (3, 4), motor clustering on the microtubules (5, 6) and structural elements (7, 8) regulate the bi-directional motility of kinesin motors. We also show, using Cryo-EM, that bi-directional kinesin motors obtain non-canonical microtubule binding, which is essential to their special motile properties and intracellular functions. We will discuss the implication of these findings to mechanism bi-directional motility and physiological roles in mitosis.

Keywords: mitosis, cancer, kinesin, microtubules, biochemistry, biophysics

Procedia PDF Downloads 81
2191 Using Soil Texture Field Observations as Ordinal Qualitative Variables for Digital Soil Mapping

Authors: Anne C. Richer-De-Forges, Dominique Arrouays, Songchao Chen, Mercedes Roman Dobarco

Abstract:

Most of the digital soil mapping (DSM) products rely on machine learning (ML) prediction models and/or the use or pedotransfer functions (PTF) in which calibration data come from soil analyses performed in labs. However, many other observations (often qualitative, nominal, or ordinal) could be used as proxies of lab measurements or as input data for ML of PTF predictions. DSM and ML are briefly described with some examples taken from the literature. Then, we explore the potential of an ordinal qualitative variable, i.e., the hand-feel soil texture (HFST) estimating the mineral particle distribution (PSD): % of clay (0-2µm), silt (2-50µm) and sand (50-2000µm) in 15 classes. The PSD can also be measured by lab measurements (LAST) to determine the exact proportion of these particle-sizes. However, due to cost constraints, HFST are much more numerous and spatially dense than LAST. Soil texture (ST) is a very important soil parameter to map as it is controlling many of the soil properties and functions. Therefore, comes an essential question: is it possible to use HFST as a proxy of LAST for calibration and/or validation of DSM predictions of ST? To answer this question, the first step is to compare HFST with LAST on a representative set where both information are available. This comparison was made on ca 17,400 samples representative of a French region (34,000 km2). The accuracy of HFST was assessed, and each HFST class was characterized by a probability distribution function (PDF) of its LAST values. This enables to randomly replace HFST observations by LAST values while respecting the PDF previously calculated and results in a very large increase of observations available for the calibration or validation of PTF and ML predictions. Some preliminary results are shown. First, the comparison between HFST classes and LAST analyses showed that accuracies could be considered very good when compared to other studies. The causes of some inconsistencies were explored and most of them were well explained by other soil characteristics. Then we show some examples applying these relationships and the increase of data to several issues related to DSM. The first issue is: do the PDF functions that were established enable to use HSFT class observations to improve the LAST soil texture prediction? For this objective, we replaced all HFST for topsoil by values from the PDF 100 time replicates). Results were promising for the PTF we tested (a PTF predicting soil water holding capacity). For the question related to the ML prediction of LAST soil texture on the region, we did the same kind of replacement, but we implemented a 10-fold cross-validation using points where we had LAST values. We obtained only preliminary results but they were rather promising. Then we show another example illustrating the potential of using HFST as validation data. As in numerous countries, the HFST observations are very numerous; these promising results pave the way to an important improvement of DSM products in all the countries of the world.

Keywords: digital soil mapping, improvement of digital soil mapping predictions, potential of using hand-feel soil texture, soil texture prediction

Procedia PDF Downloads 224
2190 Application of a Universal Distortion Correction Method in Stereo-Based Digital Image Correlation Measurement

Authors: Hu Zhenxing, Gao Jianxin

Abstract:

Stereo-based digital image correlation (also referred to as three-dimensional (3D) digital image correlation (DIC)) is a technique for both 3D shape and surface deformation measurement of a component, which has found increasing applications in academia and industries. The accuracy of the reconstructed coordinate depends on many factors such as configuration of the setup, stereo-matching, distortion, etc. Most of these factors have been investigated in literature. For instance, the configuration of a binocular vision system determines the systematic errors. The stereo-matching errors depend on the speckle quality and the matching algorithm, which can only be controlled in a limited range. And the distortion is non-linear particularly in a complex imaging acquisition system. Thus, the distortion correction should be carefully considered. Moreover, the distortion function is difficult to formulate in a complex imaging acquisition system using conventional models in such cases where microscopes and other complex lenses are involved. The errors of the distortion correction will propagate to the reconstructed 3D coordinates. To address the problem, an accurate mapping method based on 2D B-spline functions is proposed in this study. The mapping functions are used to convert the distorted coordinates into an ideal plane without distortions. This approach is suitable for any image acquisition distortion models. It is used as a prior process to convert the distorted coordinate to an ideal position, which enables the camera to conform to the pin-hole model. A procedure of this approach is presented for stereo-based DIC. Using 3D speckle image generation, numerical simulations were carried out to compare the accuracy of both the conventional method and the proposed approach.

Keywords: distortion, stereo-based digital image correlation, b-spline, 3D, 2D

Procedia PDF Downloads 498