Search results for: event study methodology
51887 Tandem Concentrated Photovoltaic-Thermoelectric Hybrid System: Feasibility Analysis and Performance Enhancement Through Material Assessment Methodology
Authors: Shuwen Hu, Yuancheng Lou, Dongxu Ji
Abstract:
Photovoltaic (PV) power generation, as one of the most commercialized methods to utilize solar power, can only convert a limited range of solar spectrum into electricity, whereas the majority of the solar energy is dissipated as heat. To address this problem, thermoelectric (TE) module is often integrated with the concentrated PV module for waste heat recovery and regeneration. In this research, a feasibility analysis is conducted for the tandem concentrated photovoltaic-thermoelectric (CPV-TE) hybrid system considering various operational parameters as well as TE material properties. Furthermore, the power output density of the CPV-TE hybrid system is maximized by selecting the optimal TE material with application of a systematic assessment methodology. In the feasibility analysis, CPV-TE is found to be more advantageous than sole CPV system except under high optical concentration ratio with low cold side convective coefficient. It is also shown that the effects of the TE material properties, including Seebeck coefficient, thermal conductivity, and electrical resistivity, on the feasibility of CPV-TE are interacted with each other and might have opposite effect on the system performance under different operational conditions. In addition, the optimal TE material selected by the proposed assessment methodology can improve the system power output density by 227 W/m2 under highly concentrated solar irradiance hence broaden the feasible range of CPV-TE considering optical concentration ratio.Keywords: feasibility analysis, material assessment methodology, photovoltaic waste heat recovery, tandem photovoltaic-thermoelectric
Procedia PDF Downloads 7251886 Development of Creatively Integrated Teaching Skills Using Information and Communication Technology for Professional Teacher
Authors: Siwanit Autthawuttikul, Prakob Koraneekid, Sayamon Insa-ard
Abstract:
The purposes of this research were to development creatively integrated teaching skills using Information and Communication Technology (ICT) for professional teacher in schools under the education area of the basic education commission, ministry of education both schools under the office of primary education and those under The office of secondary education in eight western region provinces of Thailand. This is useful in defining a vision for the school strategy and restructuring schools in addition, teachers will have developed skills in teaching creative integrated ICT. The research methodology comprises quantitative and qualitative data collection. The Baseline Survey, focus group for discussions and then the model was developed creatively integrated teaching skills using ICT. The findings showed that 7 elements were important: (1) Academy Transformation (2) Information Technology Infrastructure (3) Personal Development (4) Supervision, Monitoring and Evaluation (5) Motivating and Rewarding (6) Important factor affecting the success of teaching integrated with ICT were knowledge, skills, attitudes and (7) The role of the individual concerned. The comparison creatively integrated teaching skills before and after participating in the overall shows that the average creatively integrated teaching skills using ICT after attending the event is 3.27, and standard deviation was 0.56, higher than before which is 2.60 and the standard deviation was 0.56. There are significant differences significant statistically level of .05. The final average score of the evaluation plan design creatively integrated teaching skills using ICT teachers' average score was 26.94 at the high levels.Keywords: integrated curriculum, information and communications technology, teachers in the western region, schools
Procedia PDF Downloads 44551885 Science School Was Burned: A Case Study of Crisis Management in Thailand
Authors: Proud Arunrangsiwed
Abstract:
This study analyzes the crisis management and image repair strategies during the crisis of Mahidol Wittayanusorn School (MWIT) library burning. The library of this school was burned by a 16-year-old-male student on June 6th, 2010. This student blamed the school that the lesson was difficult, and other students were selfish. Although no one was in the building during the fire, it had caused damage to the building, books and electronic supplies around 130 million bahts (4.4 million USD). This event aroused many discourses arguing about the education system and morality. The strategies which were used during crisis were denial, shift the blame, bolstering, minimization, and uncertainty reduction. The results of using these strategies appeared after the crisis. That was the numbers of new students, who registered for the examination to get into this school in the later years, have remained the same.Keywords: school, crisis management, violence, image repair strategies, uncertainty, burn
Procedia PDF Downloads 47151884 Optimization of Friction Stir Welding Parameters for Joining Aluminium Alloys using Response Surface Methodology and Artificial Neural Network
Authors: A. M. Khourshid, A. M. El-Kassas, I. Sabry
Abstract:
The objective of this work was to investigate the mechanical properties in order to demonstrate the feasibility of friction stir welding for joining Al 6061 aluminium alloys. Welding was performed on pipe with different thickness (2, 3 and 4 mm), five rotational speeds (485, 710, 910, 1120 and 1400 rpm) and a traverse speed of 4mm/min. This work focuses on two methods which are artificial neural networks using software and Response Surface Methodology (RSM) to predict the tensile strength, the percentage of elongation and hardness of friction stir welded 6061 aluminium alloy. An Artificial Neural Network (ANN) model was developed for the analysis of the friction stir welding parameters of 6061 pipe. Tensile strength, the percentage of elongation and hardness of weld joints were predicted by taking the parameters tool rotation speed, material thickness and axial force as a function. A comparison was made between measured and predicted data. Response Surface Methodology (RSM) was also developed and the values obtained for the response tensile strength, the percentage of elongation and hardness are compared with measured values. The effect of FSW process parameters on mechanical properties of 6061 aluminium alloy has been analysed in detail.Keywords: friction stir welding, aluminium alloy, response surface methodology, artificial neural network
Procedia PDF Downloads 29351883 A Distributed Mobile Agent Based on Intrusion Detection System for MANET
Authors: Maad Kamal Al-Anni
Abstract:
This study is about an algorithmic dependence of Artificial Neural Network on Multilayer Perceptron (MPL) pertaining to the classification and clustering presentations for Mobile Adhoc Network vulnerabilities. Moreover, mobile ad hoc network (MANET) is ubiquitous intelligent internetworking devices in which it has the ability to detect their environment using an autonomous system of mobile nodes that are connected via wireless links. Security affairs are the most important subject in MANET due to the easy penetrative scenarios occurred in such an auto configuration network. One of the powerful techniques used for inspecting the network packets is Intrusion Detection System (IDS); in this article, we are going to show the effectiveness of artificial neural networks used as a machine learning along with stochastic approach (information gain) to classify the malicious behaviors in simulated network with respect to different IDS techniques. The monitoring agent is responsible for detection inference engine, the audit data is collected from collecting agent by simulating the node attack and contrasted outputs with normal behaviors of the framework, whenever. In the event that there is any deviation from the ordinary behaviors then the monitoring agent is considered this event as an attack , in this article we are going to demonstrate the signature-based IDS approach in a MANET by implementing the back propagation algorithm over ensemble-based Traffic Table (TT), thus the signature of malicious behaviors or undesirable activities are often significantly prognosticated and efficiently figured out, by increasing the parametric set-up of Back propagation algorithm during the experimental results which empirically shown its effectiveness for the ratio of detection index up to 98.6 percentage. Consequently it is proved in empirical results in this article, the performance matrices are also being included in this article with Xgraph screen show by different through puts like Packet Delivery Ratio (PDR), Through Put(TP), and Average Delay(AD).Keywords: Intrusion Detection System (IDS), Mobile Adhoc Networks (MANET), Back Propagation Algorithm (BPA), Neural Networks (NN)
Procedia PDF Downloads 19451882 Physical Model Testing of Storm-Driven Wave Impact Loads and Scour at a Beach Seawall
Authors: Sylvain Perrin, Thomas Saillour
Abstract:
The Grande-Motte port and seafront development project on the French Mediterranean coastline entailed evaluating wave impact loads (pressures and forces) on the new beach seawall and comparing the resulting scour potential at the base of the existing and new seawall. A physical model was built at ARTELIA’s hydraulics laboratory in Grenoble (France) to provide insight into the evolution of scouring overtime at the front of the wall, quasi-static and impulsive wave force intensity and distribution on the wall, and water and sand overtopping discharges over the wall. The beach was constituted of fine sand and approximately 50 m wide above mean sea level (MSL). Seabed slopes were in the range of 0.5% offshore to 1.5% closer to the beach. A smooth concrete structure will replace the existing concrete seawall with an elevated curved crown wall. Prior the start of breaking (at -7 m MSL contour), storm-driven maximum spectral significant wave heights of 2.8 m and 3.2 m were estimated for the benchmark historical storm event dated of 1997 and the 50-year return period storms respectively, resulting in 1 m high waves at the beach. For the wave load assessment, a tensor scale measured wave forces and moments and five piezo / piezo-resistive pressure sensors were placed on the wall. Light-weight sediment physical model and pressure and force measurements were performed with scale 1:18. The polyvinyl chloride light-weight particles used to model the prototype silty sand had a density of approximately 1 400 kg/m3 and a median diameter (d50) of 0.3 mm. Quantitative assessments of the seabed evolution were made using a measuring rod and also a laser scan survey. Testing demonstrated the occurrence of numerous impulsive wave impacts on the reflector (22%), induced not by direct wave breaking but mostly by wave run-up slamming on the top curved part of the wall. Wave forces of up to 264 kilonewtons and impulsive pressure spikes of up to 127 kilonewtons were measured. Maximum scour of -0.9 m was measured for the new seawall versus -0.6 m for the existing seawall, which is imputable to increased wave reflection (coefficient was 25.7 - 30.4% vs 23.4 - 28.6%). This paper presents a methodology for the setup and operation of a physical model in order to assess the hydrodynamic and morphodynamic processes at a beach seawall during storms events. It discusses the pros and cons of such methodology versus others, notably regarding structures peculiarities and model effects.Keywords: beach, impacts, scour, seawall, waves
Procedia PDF Downloads 15351881 Gamification to Enhance Learning Using Gagne's Learning Model
Authors: M. L. McLain, R. Sreelakshmi, Abhishek, Rajeshwaran, Bhavani Rao, Kamal Bijlani, R. Jayakrishnan
Abstract:
Technology enhanced learning has brought drastic changes in the field of education in the modern world. In this study we explore a novel way to improve how high school students learn by building a serious game that uses a pedagogical model developed by Robert Gagne. By integrating serious game with principles of Gagne’s learning model can provide engaging and meaningful instructions to students. The game developed in this study is a waste sorting game that can easily and succinctly demonstrate the principles of this learning model. All the tasks in the game that the player has to accomplish correspond to Gagne’s “Nine Events of Learning”. A quiz is incorporated in order to get data on the progress made by the player in understanding the concept and as well as to assess them. Additionally, an experimental study was conducted which demonstrates that game based learning using Gagne’s event is more effective than a traditional classroom setup.Keywords: game based learning, sorting and recycling of waste, Gagne’s learning model, e-Learning, technology enhanced learning
Procedia PDF Downloads 63151880 The Crisis of Turkey's Downing the Russian Warplane within the Concept of Country Branding: The Examples of BBC World, and Al Jazeera English
Authors: Derya Gül Ünlü, Oguz Kuş
Abstract:
The branding of a country means that the country has its own position different from other countries in its region and thus it is perceived more specifically. It is made possible by the branding efforts of a country and the uniqueness of all the national structures, by presenting it in a specific way, by creating the desired image and attracting tourists and foreign investors. Establishing a national brand involves, in a sense, the process of managing the perceptions of the citizens of the other country about the target country, by structuring the image of the country permanently and holistically. By this means, countries are not easily affected by their crisis of international relations. Therefore, within the scope of the research that will be carried out from this point, it is aimed to show how the warplane downing crisis between Turkey and Russia is perceived on social media. The Russian warplane was downed by Turkey on November 24, 2015, on the grounds that Turkey violated the airspace on the Syrian border. Whereupon the relations between the two countries have been tensed, and Russia has called on its citizens not to go to Turkey and citizens in Turkey to return to their countries. Moreover, relations between two countries have been weakened, for example, tourism tours organized in Russia to Turkey and visa-free travel were canceled and all military dialogue was cut off. After the event, various news sites on social media published plenty of news related to topic and the readers made various comments about the event and Turkey. In this context, an investigation into the perception of Turkey's national brand before and after the warplane downing crisis has been conducted. through comments fetched from the reports on the BBC World, and from Al Jazeera English news sites on Facebook accounts, which takes place widely in the social media. In order to realize study, user comments were fetched from jet downing-related news which are published on Facebook fan-page of BBC World Service, and Al Jazeera English. Regarding this, all the news published between 24.10.2015-24.12.2015 and containing Turk and Turkey keyword in its title composed data set of our study. Afterwards, comments written to these news were analyzed via text mining technique. Furthermore, by sentiment analysis, it was intended to reveal reader’s emotions before and after the crisis.Keywords: Al Jazeera English, BBC World, country branding, social media, text mining
Procedia PDF Downloads 22351879 An Innovative Approach to Improve Skills of Students in Qatar University Spending in Virtual Class though LMS
Authors: Mohammad Shahid Jamil
Abstract:
In this study we have investigated students’ learning and satisfaction in one of the course offered in the Foundation Program at Qatar University. We implied innovative teaching methodology that emphasizes on enhancing students’ thinking skills, decision making, and problem solving skills. Some interesting results were found which can be used to further improve the teaching methodology. To make sure the full use of technology in Foundation Program at Qatar University has started implementing new ways of teaching Math course by using Blackboard as an innovative interactive tool to support standard teaching such as Discussion board, Virtual class, and Study plan in My Math Lab “MML”. In MML Study Plan is designed in such a way that the student can improve their skills wherever they face difficulties with in their Homework, Quiz or Test. Discussion board and Virtual Class are collaborative learning tools encourages students to engage outside of class time. These tools are useful to share students’ knowledge and learning experiences, promote independent and active learning and they helps students to improve their critical thinking skills through the learning process.Keywords: blackboard, discussion board, critical thinking, active learning, independent learning, problem solving
Procedia PDF Downloads 42851878 Development of Requirements Analysis Tool for Medical Autonomy in Long-Duration Space Exploration Missions
Authors: Lara Dutil-Fafard, Caroline Rhéaume, Patrick Archambault, Daniel Lafond, Neal W. Pollock
Abstract:
Improving resources for medical autonomy of astronauts in prolonged space missions, such as a Mars mission, requires not only technology development, but also decision-making support systems. The Advanced Crew Medical System - Medical Condition Requirements study, funded by the Canadian Space Agency, aimed to create knowledge content and a scenario-based query capability to support medical autonomy of astronauts. The key objective of this study was to create a prototype tool for identifying medical infrastructure requirements in terms of medical knowledge, skills and materials. A multicriteria decision-making method was used to prioritize the highest risk medical events anticipated in a long-term space mission. Starting with those medical conditions, event sequence diagrams (ESDs) were created in the form of decision trees where the entry point is the diagnosis and the end points are the predicted outcomes (full recovery, partial recovery, or death/severe incapacitation). The ESD formalism was adapted to characterize and compare possible outcomes of medical conditions as a function of available medical knowledge, skills, and supplies in a given mission scenario. An extensive literature review was performed and summarized in a medical condition database. A PostgreSQL relational database was created to allow query-based evaluation of health outcome metrics with different medical infrastructure scenarios. Critical decision points, skill and medical supply requirements, and probable health outcomes were compared across chosen scenarios. The three medical conditions with the highest risk rank were acute coronary syndrome, sepsis, and stroke. Our efforts demonstrate the utility of this approach and provide insight into the effort required to develop appropriate content for the range of medical conditions that may arise.Keywords: decision support system, event-sequence diagram, exploration mission, medical autonomy, scenario-based queries, space medicine
Procedia PDF Downloads 12751877 Climate Change Adaptation: Methodologies and Tools to Define Resilience Scenarios for Existing Buildings in Mediterranean Urban Areas
Authors: Francesca Nicolosi, Teresa Cosola
Abstract:
Climate changes in Mediterranean areas, such as the increase of average seasonal temperatures, the urban heat island phenomenon, the intensification of solar radiation and the extreme weather threats, cause disruption events, so that climate adaptation has become a pressing issue. Due to the strategic role that the built heritage holds in terms of environmental impact and energy waste and its potentiality, it is necessary to assess the vulnerability and the adaptive capacity of the existing building to climate change, in order to define different mitigation scenarios. The aim of this research work is to define an optimized and integrated methodology for the assessment of resilience levels and adaptation scenarios for existing buildings in Mediterranean urban areas. Moreover, the study of resilience indicators allows us to define building environmental and energy performance in order to identify the design and technological solutions for the improvement of the building and its urban area potentialities. The methodology identifies step-by-step different phases, starting from the detailed study of characteristic elements of urban system: climatic, natural, human, typological and functional components are analyzed in their critical factors and their potential. Through the individuation of the main perturbing factors and the vulnerability degree of the system to the risks linked to climate change, it is possible to define mitigation and adaptation scenarios. They can be different, according to the typological, functional and constructive features of the analyzed system, divided into categories of intervention, and characterized by different analysis levels (from the single building to the urban area). The use of software simulations allows obtaining information on the overall behavior of the building and the urban system, to generate predictive models in the medium and long-term environmental and energy retrofit and to make a comparative study of the mitigation scenarios identified. The studied methodology is validated on a case study.Keywords: climate impact mitigation, energy efficiency, existing building heritage, resilience
Procedia PDF Downloads 24051876 Effect of Process Variables of Wire Electrical Discharge Machining on Surface Roughness for AA-6063 by Response Surface Methodology
Authors: Deepak
Abstract:
WEDM is an amazingly potential electro-wire process for machining of hard metal compounds and metal grid composites without making contact. Wire electrical machining is a developing noncustomary machining process for machining hard to machine materials that are electrically conductive. It is an exceptionally exact, precise, and one of the most famous machining forms in nontraditional machining. WEDM has turned into the fundamental piece of many assembling process ventures, which require precision, variety, and accuracy. In the present examination, AA-6063 is utilized as a workpiece, and execution investigation is done to discover the critical control factors. Impact of different parameters like a pulse on time, pulse off time, servo voltage, peak current, water pressure, wire tension, wire feed upon surface hardness has been researched while machining on AA-6063. RSM has been utilized to advance the yield variable. A variety of execution measures with input factors was demonstrated by utilizing the response surface methodology.Keywords: AA-6063, response surface methodology, WEDM, surface roughness
Procedia PDF Downloads 11651875 Risk Analysis of Leaks from a Subsea Oil Facility Based on Fuzzy Logic Techniques
Authors: Belén Vinaixa Kinnear, Arturo Hidalgo López, Bernardo Elembo Wilasi, Pablo Fernández Pérez, Cecilia Hernández Fuentealba
Abstract:
The expanded use of risk assessment in legislative and corporate decision-making has increased the role of expert judgement in giving data for security-related decision-making. Expert judgements are required in most steps of risk assessment: danger recognizable proof, hazard estimation, risk evaluation, and examination of choices. This paper presents a fault tree analysis (FTA), which implies a probabilistic failure analysis applied to leakage of oil in a subsea production system. In standard FTA, the failure probabilities of items of a framework are treated as exact values while evaluating the failure probability of the top event. There is continuously insufficiency of data for calculating the failure estimation of components within the drilling industry. Therefore, fuzzy hypothesis can be used as a solution to solve the issue. The aim of this paper is to examine the leaks from the Zafiro West subsea oil facility by using fuzzy fault tree analysis (FFTA). As a result, the research has given theoretical and practical contributions to maritime safety and environmental protection. It has been also an effective strategy used traditionally in identifying hazards in nuclear installations and power industries.Keywords: expert judgment, probability assessment, fault tree analysis, risk analysis, oil pipelines, subsea production system, drilling, quantitative risk analysis, leakage failure, top event, off-shore industry
Procedia PDF Downloads 19051874 Damage Detection in a Cantilever Beam under Different Excitation and Temperature Conditions
Authors: A. Kyprianou, A. Tjirkallis
Abstract:
Condition monitoring of structures in service is very important as it provides information about the risk of damage development. One of the essential constituents of structural condition monitoring is the damage detection methodology. In the context of condition monitoring of in service structures a damage detection methodology analyses data obtained from the structure while it is in operation. Usually, this means that the data could be affected by operational and environmental conditions in a way that could mask the effects of a possible damage on the data. This, depending on the damage detection methodology, could lead to either false alarms or miss existing damages. In this article a damage detection methodology that is based on the Spatio-temporal continuous wavelet transform (SPT-CWT) analysis of a sequence of experimental time responses of a cantilever beam is proposed. The cantilever is subjected to white and pink noise excitation to simulate different operating conditions. In addition, in order to simulate changing environmental conditions, the cantilever is subjected to heating by a heat gun. The response of the cantilever beam is measured by a high-speed camera. Edges are extracted from the series of images of the beam response captured by the camera. Subsequent processing of the edges gives a series of time responses on 439 points on the beam. This sequence is then analyzed using the SPT-CWT to identify damage. The algorithm proposed was able to clearly identify damage under any condition when the structure was excited by white noise force. In addition, in the case of white noise excitation, the analysis could also reveal the position of the heat gun when it was used to heat the structure. The analysis could identify the different operating conditions i.e. between responses due to white noise excitation and responses due to pink noise excitation. During the pink noise excitation whereas damage and changing temperature were identified it was not possible to clearly identify the effect of damage from that of temperature. The methodology proposed in this article for damage detection enables the separation the damage effect from that due to temperature and excitation on data obtained from measurements of a cantilever beam. This methodology does not require information about the apriori state of the structure.Keywords: spatiotemporal continuous wavelet transform, damage detection, data normalization, varying temperature
Procedia PDF Downloads 27951873 Optimization of Process Parameters Affecting Biogas Production from Organic Fraction of Municipal Solid Waste via Anaerobic Digestion
Authors: B. Sajeena Beevi, P. P. Jose, G. Madhu
Abstract:
The aim of this study was to obtain the optimal conditions for biogas production from anaerobic digestion of organic fraction of municipal solid waste (OFMSW) using response surface methodology (RSM). The parameters studied were initial pH, substrate concentration and total organic carbon (TOC). The experimental results showed that the linear model terms of initial pH and substrate concentration and the quadratic model terms of the substrate concentration and TOC had significant individual effect (p < 0.05) on biogas yield. However, there was no interactive effect between these variables (p > 0.05). The highest level of biogas produced was 53.4 L/Kg VS at optimum pH, substrate concentration and total organic carbon of 6.5, 99gTS/L, and 20.32 g/L respectively.Keywords: anaerobic digestion, biogas, optimization, response surface methodology
Procedia PDF Downloads 43351872 The Impact of Government Expenditure on Economic Growth: A Study of Asian Countries
Authors: K. P. K. S. Lahirushan, W. G. V. Gunasekara
Abstract:
Main purpose of this study is to identifying the impact of government expenditure on economic growth in Asian Countries. Consequently, Fist, objective is to analyze whether government expenditure causes economic growth in Asian countries vice versa and then scrutinizing long-run equilibrium relationship exists between them. The study completely based on secondary data. The methodology being quantitative that includes econometrical techniques of cointegration, panel fixed effects model and granger causality in the context of panel data of Asian countries; Singapore, Malaysia, Thailand, South Korea, Japan, China, Sri Lanka, India and Bhutan with 44 observations in each country, totaling to 396 observations from 1970 to 2013. The model used is the random effects panel OLS model. As with the above methodology, the study found the fascinating outcome. At first, empirical findings exhibit a momentous positive impact of government expenditure on Gross Domestic Production in Asian region. Secondly, government expenditure and economic growth indicate a long-run relationship in Asian countries. In conclusion, there is a unidirectional causality from economic growth to government expenditure and government expenditure to economic growth in Asian countries. Hence the study is validated that it is in line with the Keynesian theory and Wagner’s law as well. Consequently, it can be concluded that role of government would play a vital role in economic growth of Asian Countries .However; if government expenditure did not figure out with the economy’s needs it might be considerably inspiration the economy in a negative way so that society bears the costs.Keywords: Asian countries, government expenditure, Keynesian theory, Wagner’s theory, random effects panel ols model
Procedia PDF Downloads 35251871 Harmonizing Spatial Plans: A Methodology to Integrate Sustainable Mobility and Energy Plans to Promote Resilient City Planning
Authors: B. Sanchez, D. Zambrana-Vasquez, J. Fresner, C. Krenn, F. Morea, L. Mercatelli
Abstract:
Local administrations are facing established targets on sustainable development from different disciplines at the heart of different city departments. Nevertheless, some of these targets, such as CO2 reduction, relate to two or more disciplines, as it is the case of sustainable mobility and energy plans (SUMP & SECAP/SEAP). This opens up the possibility to efficiently cooperate among different city departments and to create and develop harmonized spatial plans by using available resources and together achieving more ambitious goals in cities. The steps of the harmonization processes developed result in the identification of areas to achieve common strategic objectives. Harmonization, in other words, helps different departments in local authorities to work together and optimize the use or resources by sharing the same vision, involving key stakeholders, and promoting common data assessment to better optimize the resources. A methodology to promote resilient city planning via the harmonization of sustainable mobility and energy plans is presented in this paper. In order to validate the proposed methodology, a representative city engaged in an innovation process in efficient spatial planning is used as a case study. The harmonization process of sustainable mobility and energy plans covers identifying matching targets between different fields, developing different spatial plans with dual benefit and common indicators guaranteeing the continuous improvement of the harmonized plans. The proposed methodology supports local administrations in consistent spatial planning, considering both energy efficiency and sustainable mobility. Thus, municipalities can use their human and economic resources efficiently. This guarantees an efficient upgrade of land use plans integrating energy and mobility aspects in order to achieve sustainability targets, as well as to improve the wellbeing of its citizens.Keywords: integrated multi-sector planning, spatial plans harmonization, sustainable energy and climate action plan, sustainable urban mobility plan
Procedia PDF Downloads 17851870 Application Reliability Method for the Analysis of the Stability Limit States of Large Concrete Dams
Authors: Mustapha Kamel Mihoubi, Essadik Kerkar, Abdelhamid Hebbouche
Abstract:
According to the randomness of most of the factors affecting the stability of a gravity dam, probability theory is generally used to TESTING the risk of failure and there is a confusing logical transition from the state of stability failed state, so the stability failure process is considered as a probable event. The control of risk of product failures is of capital importance for the control from a cross analysis of the gravity of the consequences and effects of the probability of occurrence of identified major accidents and can incur a significant risk to the concrete dam structures. Probabilistic risk analysis models are used to provide a better understanding the reliability and structural failure of the works, including when calculating stability of large structures to a major risk in the event of an accident or breakdown. This work is interested in the study of the probability of failure of concrete dams through the application of the reliability analysis methods including the methods used in engineering. It is in our case of the use of level II methods via the study limit state. Hence, the probability of product failures is estimated by analytical methods of the type FORM (First Order Reliability Method), SORM (Second Order Reliability Method). By way of comparison, a second level III method was used which generates a full analysis of the problem and involving an integration of the probability density function of, random variables are extended to the field of security by using of the method of Mont-Carlo simulations. Taking into account the change in stress following load combinations: normal, exceptional and extreme the acting on the dam, calculation results obtained have provided acceptable failure probability values which largely corroborate the theory, in fact, the probability of failure tends to increase with increasing load intensities thus causing a significant decrease in strength, especially in the presence of combinations of unique and extreme loads. Shear forces then induce a shift threatens the reliability of the structure by intolerable values of the probability of product failures. Especially, in case THE increase of uplift in a hypothetical default of the drainage system.Keywords: dam, failure, limit state, monte-carlo, reliability, probability, sliding, Taylor
Procedia PDF Downloads 31851869 A Three Elements Vector Valued Structure’s Ultimate Strength-Strong Motion-Intensity Measure
Authors: A. Nicknam, N. Eftekhari, A. Mazarei, M. Ganjvar
Abstract:
This article presents an alternative collapse capacity intensity measure in the three elements form which is influenced by the spectral ordinates at periods longer than that of the first mode period at near and far source sites. A parameter, denoted by β, is defined by which the spectral ordinate effects, up to the effective period (2T_1), on the intensity measure are taken into account. The methodology permits to meet the hazard-levelled target extreme event in the probabilistic and deterministic forms. A MATLAB code is developed involving OpenSees to calculate the collapse capacities of the 8 archetype RC structures having 2 to 20 stories for regression process. The incremental dynamic analysis (IDA) method is used to calculate the structure’s collapse values accounting for the element stiffness and strength deterioration. The general near field set presented by FEMA is used in a series of performing nonlinear analyses. 8 linear relationships are developed for the 8structutres leading to the correlation coefficient up to 0.93. A collapse capacity near field prediction equation is developed taking into account the results of regression processes obtained from the 8 structures. The proposed prediction equation is validated against a set of actual near field records leading to a good agreement. Implementation of the proposed equation to the four archetype RC structures demonstrated different collapse capacities at near field site compared to those of FEMA. The reasons of differences are believed to be due to accounting for the spectral shape effects.Keywords: collapse capacity, fragility analysis, spectral shape effects, IDA method
Procedia PDF Downloads 23951868 Urban Growth Analysis Using Multi-Temporal Satellite Images, Non-stationary Decomposition Methods and Stochastic Modeling
Authors: Ali Ben Abbes, ImedRiadh Farah, Vincent Barra
Abstract:
Remotely sensed data are a significant source for monitoring and updating databases for land use/cover. Nowadays, changes detection of urban area has been a subject of intensive researches. Timely and accurate data on spatio-temporal changes of urban areas are therefore required. The data extracted from multi-temporal satellite images are usually non-stationary. In fact, the changes evolve in time and space. This paper is an attempt to propose a methodology for changes detection in urban area by combining a non-stationary decomposition method and stochastic modeling. We consider as input of our methodology a sequence of satellite images I1, I2, … In at different periods (t = 1, 2, ..., n). Firstly, a preprocessing of multi-temporal satellite images is applied. (e.g. radiometric, atmospheric and geometric). The systematic study of global urban expansion in our methodology can be approached in two ways: The first considers the urban area as one same object as opposed to non-urban areas (e.g. vegetation, bare soil and water). The objective is to extract the urban mask. The second one aims to obtain a more knowledge of urban area, distinguishing different types of tissue within the urban area. In order to validate our approach, we used a database of Tres Cantos-Madrid in Spain, which is derived from Landsat for a period (from January 2004 to July 2013) by collecting two frames per year at a spatial resolution of 25 meters. The obtained results show the effectiveness of our method.Keywords: multi-temporal satellite image, urban growth, non-stationary, stochastic model
Procedia PDF Downloads 42851867 Investigating the Effect of Orthographic Transparency on Phonological Awareness in Bilingual Children with Dyslexia
Authors: Sruthi Raveendran
Abstract:
Developmental dyslexia, characterized by reading difficulties despite normal intelligence, presents a significant challenge for bilingual children navigating languages with varying degrees of orthographic transparency. This study bridges a critical gap in dyslexia interventions for bilingual populations in India by examining how consistency and predictability of letter-sound relationships in a writing system (orthographic transparency) influence the ability to understand and manipulate the building blocks of sound in language (phonological processing). The study employed a computerized visual rhyme-judgment task with concurrent EEG (electroencephalogram) recording. The task compared reaction times, accuracy of performance, and event-related potential (ERP) components (N170, N400, and LPC) for rhyming and non-rhyming stimuli in two orthographies: English (opaque orthography) and Kannada (transparent orthography). As hypothesized, the results revealed advantages in phonological processing tasks for transparent orthography (Kannada). Children with dyslexia were faster and more accurate when judging rhymes in Kannada compared to English. This suggests that a language with consistent letter-sound relationships (transparent orthography) facilitates processing, especially for tasks that involve manipulating sounds within words (rhyming). Furthermore, brain activity measured by event-related potentials (ERP) showed less effort required for processing words in Kannada, as reflected by smaller N170, N400, and LPC amplitudes. These findings highlight the crucial role of orthographic transparency in optimizing reading performance for bilingual children with dyslexia. These findings emphasize the need for language-specific intervention strategies that consider the unique linguistic characteristics of each language. While acknowledging the complexity of factors influencing dyslexia, this research contributes valuable insights into the impact of orthographic transparency on phonological awareness in bilingual children. This knowledge paves the way for developing tailored interventions that promote linguistic inclusivity and optimize literacy outcomes for children with dyslexia.Keywords: developmental dyslexia, phonological awareness, rhyme judgment, orthographic transparency, Kannada, English, N170, N400, LPC
Procedia PDF Downloads 751866 A Model Approach of Good Practice Based on the Project Management Body of Knowledge® Guide in the Project Owner
Authors: Claudia Marcela Munoz Gonzalez, Diego Fernando Hernandez Losada, Hugo Alberto Herrera Fonseca
Abstract:
The project owner's role in the public-private investment consists of controlling and verifying the correct execution of the project's objectives and resources. Likewise, it is a discipline little explored in the academic field, whereby this work wishes to contribute with a model of good practices based on the project management methodology proposed by the Project Management Body of Knowledge® Guide. In the same way, highlight what are the controls that an integral project owner should take into account in its exercise and application, through the stages in which its contract runs. This proposal aims to structure its practice and integrate its functions according to a project management methodology. In addition, these practices will be applied in a case study of projects in the agricultural sector, particularly in the construction of irrigation district in Cundinamarca, Colombia.Keywords: controls, construction of irrigation district, PMBOK®, project owner
Procedia PDF Downloads 45851865 Capital Market Reaction to Governance and Disclosure Violations: Evidence from the Saudi Arabian Capital Market
Authors: Nasser Alsadoun
Abstract:
Today's companies in Saudi Arabian capital market must comply with strict criteria and adhere to rigid corporate governance rules and continuous disclosure requirements. Unlike other regulators in the region, decision makers of the Capital Market Authority (hereafter CMA) of Saudi Arabia believes that the announcements of economic sanctions and penalties for non-compliance firms will foster more effective regulatory compliance and hence improve the quality of financial reporting. An implied argument put forward by the opponents, however, states that such penalties are unnecessary and stated to be onerous for non-compliance firms. Over that last years, the CMA has publicly announced several economic fines levied on some listed companies for their failing to comply with corporate governance and continuous disclosure regulation clauses, with the amount of fine levied ranges between 50,000 SR to 100,000 SR for each failing. Economic theory suggests that rational investors make decisions based on a cost-benefit principal. The regulatory intervention made by CMA on the announcement of economic sanctions has been costly to the society (economy) hoping that it improves the transparency of financial statements. It is argued, therefore, that threat of regulators and economic sanctions will provide incentives for firms’ managers to report more relevant and reliable accounting information, and the benefit of such announcements is likely to be reflected in the context of the quality of the financial reports. Yet, the economic consequences of the revealed fines announcement for non-compliance firms in Saudi Arabian market have not been examined. Thus, this study attempts to empirically examine whether market participants are pricing the supposed benefits of rigid governance and disclosure rules in the Saudi market. The study employs an event study methodology to assess the impact of CMA economic sanctions announcements on the market price of non-compliance firms. The study also estimates and examines bid–ask spread behavior of violated firms around the CMA announcements. The findings indicate that the CMA fines announcements for failing to comply with governance and disclosure rules do not appear to play any significant role in securities pricing. In addition, tests of bid-ask behavior does not indicate any significant increases in information asymmetry surrounding these announcements. While the CMA has developed many goals to increase the awareness of listed companies with the best governance and disclosure practices, it seems they have to develop more goals to improve market efficiency and increase investors and public awareness.Keywords: governance and disclosure violations, financial reporting quality, regulatory intervention, market efficiency
Procedia PDF Downloads 30551864 Optimization of Diluted Organic Acid Pretreatment on Rice Straw Using Response Surface Methodology
Authors: Rotchanaphan Hengaroonprasan, Malinee Sriariyanun, Prapakorn Tantayotai, Supacharee Roddecha, Kraipat Cheenkachorn
Abstract:
Lignocellolusic material is a substance that is resistant to be degraded by microorganisms or hydrolysis enzymes. To be used as materials for biofuel production, it needs pretreatment process to improve efficiency of hydrolysis. In this work, chemical pretreatments on rice straw using three diluted organic acids, including acetic acid, citric acid, oxalic acid, were optimized. Using Response Surface Methodology (RSM), the effect of three pretreatment parameters, acid concentration, treatment time, and reaction temperature, on pretreatment efficiency were statistically evaluated. The results indicated that dilute oxalic acid pretreatment led to the highest enhancement of enzymatic saccharification by commercial cellulase and yielded sugar up to 10.67 mg/ml when using 5.04% oxalic acid at 137.11 oC for 30.01 min. Compared to other acid pretreatment by acetic acid, citric acid, and hydrochloric acid, the maximum sugar yields are 7.07, 6.30, and 8.53 mg/ml, respectively. Here, it was demonstrated that organic acids can be used for pretreatment of lignocellulosic materials to enhance of hydrolysis process, which could be integrated to other applications for various biorefinery processes.Keywords: lignocellolusic biomass, pretreatment, organic acid response surface methodology, biorefinery
Procedia PDF Downloads 65451863 Applying Serious Game Design Frameworks to Existing Games for Integration of Custom Learning Objectives
Authors: Jonathan D. Moore, Mark G. Reith, David S. Long
Abstract:
Serious games (SGs) have been shown to be an effective teaching tool in many contexts. Because of the success of SGs, several design frameworks have been created to expedite the process of making original serious games to teach specific learning objectives (LOs). Even with these frameworks, the time required to create a custom SG from conception to implementation can range from months to years. Furthermore, it is even more difficult to design a game framework that allows an instructor to create customized game variants supporting multiple LOs within the same field. This paper proposes a refactoring methodology to apply the theoretical principles from well-established design frameworks to a pre-existing serious game. The expected result is a generalized game that can be quickly customized to teach LOs not originally targeted by the game. This methodology begins by describing the general components in a game, then uses a combination of two SG design frameworks to extract the teaching elements present in the game. The identified teaching elements are then used as the theoretical basis to determine the range of LOs that can be taught by the game. This paper evaluates the proposed methodology by presenting a case study of refactoring the serious game Battlespace Next (BSN) to teach joint military capabilities. The range of LOs that can be taught by the generalized BSN are identified, and examples of creating custom LOs are given. Survey results from users of the generalized game are also provided. Lastly, the expected impact of this work is discussed and a road map for future work and evaluation is presented.Keywords: serious games, learning objectives, game design, learning theory, game framework
Procedia PDF Downloads 11551862 Understanding Hydrodynamic in Lake Victoria Basin in a Catchment Scale: A Literature Review
Authors: Seema Paul, John Mango Magero, Prosun Bhattacharya, Zahra Kalantari, Steve W. Lyon
Abstract:
The purpose of this review paper is to develop an understanding of lake hydrodynamics and the potential climate impact on the Lake Victoria (LV) catchment scale. This paper briefly discusses the main problems of lake hydrodynamics and its’ solutions that are related to quality assessment and climate effect. An empirical methodology in modeling and mapping have considered for understanding lake hydrodynamic and visualizing the long-term observational daily, monthly, and yearly mean dataset results by using geographical information system (GIS) and Comsol techniques. Data were obtained for the whole lake and five different meteorological stations, and several geoprocessing tools with spatial analysis are considered to produce results. The linear regression analyses were developed to build climate scenarios and a linear trend on lake rainfall data for a long period. A potential evapotranspiration rate has been described by the MODIS and the Thornthwaite method. The rainfall effect on lake water level observed by Partial Differential Equations (PDE), and water quality has manifested by a few nutrients parameters. The study revealed monthly and yearly rainfall varies with monthly and yearly maximum and minimum temperatures, and the rainfall is high during cool years and the temperature is high associated with below and average rainfall patterns. Rising temperatures are likely to accelerate evapotranspiration rates and more evapotranspiration is likely to lead to more rainfall, drought is more correlated with temperature and cloud is more correlated with rainfall. There is a trend in lake rainfall and long-time rainfall on the lake water surface has affected the lake level. The onshore and offshore have been concentrated by initial literature nutrients data. The study recommended that further studies should consider fully lake bathymetry development with flow analysis and its’ water balance, hydro-meteorological processes, solute transport, wind hydrodynamics, pollution and eutrophication these are crucial for lake water quality, climate impact assessment, and water sustainability.Keywords: climograph, climate scenarios, evapotranspiration, linear trend flow, rainfall event on LV, concentration
Procedia PDF Downloads 9951861 Modeling and Optimization of Algae Oil Extraction Using Response Surface Methodology
Authors: I. F. Ejim, F. L. Kamen
Abstract:
Aims: In this experiment, algae oil extraction with a combination of n-hexane and ethanol was investigated. The effects of extraction solvent concentration, extraction time and temperature on the yield and quality of oil were studied using Response Surface Methodology (RSM). Experimental Design: Optimization of algae oil extraction using Box-Behnken design was used to generate 17 experimental runs in a three-factor-three-level design where oil yield, specific gravity, acid value and saponification value were evaluated as the response. Result: In this result, a minimum oil yield of 17% and maximum of 44% was realized. The optimum values for yield, specific gravity, acid value and saponification value from the overlay plot were 40.79%, 0.8788, 0.5056 mg KOH/g and 180.78 mg KOH/g respectively with desirability of 0.801. The maximum point prediction was yield 40.79% at solvent concentration 66.68 n-hexane, temperature of 40.0°C and extraction time of 4 hrs. Analysis of Variance (ANOVA) results showed that the linear and quadratic coefficient were all significant at p<0.05. The experiment was validated and results obtained were with the predicted values. Conclusion: Algae oil extraction was successfully optimized using RSM and its quality indicated it is suitable for many industrial uses.Keywords: algae oil, response surface methodology, optimization, Box-Bohnken, extraction
Procedia PDF Downloads 33751860 Qualitative and Quantitative Case Study Research Method on Social Science: Accounting Perspective
Authors: Bubaker F. Shareia
Abstract:
The main aim of this paper is to set the parameters within which the study is to be conducted, specifically justifying the use of qualitative research, informed by theory. This paper argues that the social world is subjective in nature and may be accessed through the interpretive approach provided by the people involved in the context of the study. The paper defines and distinguishes between qualitative and quantitative research methodologies, explores Burrell and Morgan's framework for social research, and presents the study's adopted methodology and methods, with the rationale for these choices.Keywords: accounting, methodologies, qualitative, quantitative research
Procedia PDF Downloads 23351859 A Study for Area-level Mosquito Abundance Prediction by Using Supervised Machine Learning Point-level Predictor
Authors: Theoktisti Makridou, Konstantinos Tsaprailis, George Arvanitakis, Charalampos Kontoes
Abstract:
In the literature, the data-driven approaches for mosquito abundance prediction relaying on supervised machine learning models that get trained with historical in-situ measurements. The counterpart of this approach is once the model gets trained on pointlevel (specific x,y coordinates) measurements, the predictions of the model refer again to point-level. These point-level predictions reduce the applicability of those solutions once a lot of early warning and mitigation actions applications need predictions for an area level, such as a municipality, village, etc... In this study, we apply a data-driven predictive model, which relies on public-open satellite Earth Observation and geospatial data and gets trained with historical point-level in-Situ measurements of mosquito abundance. Then we propose a methodology to extract information from a point-level predictive model to a broader area-level prediction. Our methodology relies on the randomly spatial sampling of the area of interest (similar to the Poisson hardcore process), obtaining the EO and geomorphological information for each sample, doing the point-wise prediction for each sample, and aggregating the predictions to represent the average mosquito abundance of the area. We quantify the performance of the transformation from the pointlevel to the area-level predictions, and we analyze it in order to understand which parameters have a positive or negative impact on it. The goal of this study is to propose a methodology that predicts the mosquito abundance of a given area by relying on point-level prediction and to provide qualitative insights regarding the expected performance of the area-level prediction. We applied our methodology to historical data (of Culex pipiens) of two areas of interest (Veneto region of Italy and Central Macedonia of Greece). In both cases, the results were consistent. The mean mosquito abundance of a given area can be estimated with similar accuracy to the point-level predictor, sometimes even better. The density of the samples that we use to represent one area has a positive effect on the performance in contrast to the actual number of sampling points which is not informative at all regarding the performance without the size of the area. Additionally, we saw that the distance between the sampling points and the real in-situ measurements that were used for training did not strongly affect the performance.Keywords: mosquito abundance, supervised machine learning, culex pipiens, spatial sampling, west nile virus, earth observation data
Procedia PDF Downloads 14751858 Times Series Analysis of Depositing in Industrial Design in Brazil between 1996 and 2013
Authors: Jonas Pedro Fabris, Alberth Almeida Amorim Souza, Maria Emilia Camargo, Suzana Leitão Russo
Abstract:
With the law Nº. 9279, of May 14, 1996, the Brazilian government regulates rights and obligations relating to industrial property considering the economic development of the country as granting patents, trademark registration, registration of industrial designs and other forms of protection copyright. In this study, we show the application of the methodology of Box and Jenkins in the series of deposits of industrial design at the National Institute of Industrial Property for the period from May 1996 to April 2013. First, a graphical analysis of the data was done by observing the behavior of the data and the autocorrelation function. The best model found, based on the analysis of charts and statistical tests suggested by Box and Jenkins methodology, it was possible to determine the model number for the deposit of industrial design, SARIMA (2,1,0)(2,0,0), with an equal to 9.88% MAPE.Keywords: ARIMA models, autocorrelation, Box and Jenkins Models, industrial design, MAPE, time series
Procedia PDF Downloads 544