Search results for: the characteristics method
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 24414

Search results for: the characteristics method

20604 A Basic Concept for Installing Cooling and Heating System Using Seawater Thermal Energy from the West Coast of Korea

Authors: Jun Byung Joon, Seo Seok Hyun, Lee Seo Young

Abstract:

As carbon dioxide emissions increase due to rapid industrialization and reckless development, abnormal climates such as floods and droughts are occurring. In order to respond to such climate change, the use of existing fossil fuels is reduced, and the proportion of eco-friendly renewable energy is gradually increasing. Korea is an energy resource-poor country that depends on imports for 93% of its total energy. As the global energy supply chain instability experienced due to the Russia-Ukraine crisis increases, countries around the world are resetting energy policies to minimize energy dependence and strengthen security. Seawater thermal energy is a renewable energy that replaces the existing air heat energy. It uses the characteristic of having a higher specific heat than air to cool and heat main spaces of buildings to increase heat transfer efficiency and minimize power consumption to generate electricity using fossil fuels, and Carbon dioxide emissions can be minimized. In addition, the effect on the marine environment is very small by using only the temperature characteristics of seawater in a limited way. K-water carried out a demonstration project of supplying cooling and heating energy to spaces such as the central control room and presentation room in the management building by acquiring the heat source of seawater circulated through the power plant's waterway by using the characteristics of the tidal power plant. Compared to the East Sea and the South Sea, the main system was designed in consideration of the large tidal difference, small temperature difference, and low-temperature characteristics, and its performance was verified through operation during the demonstration period. In addition, facility improvements were made for major deficiencies to strengthen monitoring functions, provide user convenience, and improve facility soundness. To spread these achievements, the basic concept was to expand the seawater heating and cooling system with a scale of 200 USRT at the Tidal Culture Center. With the operational experience of the demonstration system, it will be possible to establish an optimal seawater heat cooling and heating system suitable for the characteristics of the west coast ocean. Through this, it is possible to reduce operating costs by KRW 33,31 million per year compared to air heat, and through industry-university-research joint research, it is possible to localize major equipment and materials and develop key element technologies to revitalize the seawater heat business and to advance into overseas markets. The government's efforts are needed to expand the seawater heating and cooling system. Seawater thermal energy utilizes only the thermal energy of infinite seawater. Seawater thermal energy has less impact on the environment than river water thermal energy, except for environmental pollution factors such as bottom dredging, excavation, and sand or stone extraction. Therefore, it is necessary to increase the sense of speed in project promotion by innovatively simplifying unnecessary licensing/permission procedures. In addition, support should be provided to secure business feasibility by dramatically exempting the usage fee of public waters to actively encourage development in the private sector.

Keywords: seawater thermal energy, marine energy, tidal power plant, energy consumption

Procedia PDF Downloads 99
20603 Labview-Based System for Fiber Links Events Detection

Authors: Bo Liu, Qingshan Kong, Weiqing Huang

Abstract:

With the rapid development of modern communication, diagnosing the fiber-optic quality and faults in real-time is widely focused. In this paper, a Labview-based system is proposed for fiber-optic faults detection. The wavelet threshold denoising method combined with Empirical Mode Decomposition (EMD) is applied to denoise the optical time domain reflectometer (OTDR) signal. Then the method based on Gabor representation is used to detect events. Experimental measurements show that signal to noise ratio (SNR) of the OTDR signal is improved by 1.34dB on average, compared with using the wavelet threshold denosing method. The proposed system has a high score in event detection capability and accuracy. The maximum detectable fiber length of the proposed Labview-based system can be 65km.

Keywords: empirical mode decomposition, events detection, Gabor transform, optical time domain reflectometer, wavelet threshold denoising

Procedia PDF Downloads 120
20602 Cracking Performance of Bituminous Concrete Mixes Containing High Percentage of RAP Material

Authors: Bicky Agarwal, Ambika Behl, Rajiv Kumar, Ashish Dhamaniya

Abstract:

India ranks second for having the largest road network in the world after the United States (U.S.). According to the National Asphalt Pavement Association (NAPA), the U.S. produced about 94.6 million tons of Reclaimed Asphalt Pavement (RAP) in 2021. Despite the benefits of RAP usage, it is not widely adopted in many countries, including India. Rising asphalt binder costs and environmental concerns have spurred interest in using RAP material in asphalt mixtures. However, increasing RAP content may have adverse effects on certain characteristics of asphalt mixtures, such as cracking resistance. Cracking is a common pavement issue that affects the lifespan and durability of hot-mix asphalt pavements. Assessing cracking resistance is crucial in pavement design. Various laboratory tests and performance indicators are utilized to evaluate cracking resistance. This study aims to use the Texas Overlay Tester (TOT) to assess the impact of reclaimed asphalt pavement (RAP) on the cracking resistance of Bituminous Concrete (BC-II) mixes. Following the Marshall Mix Design method, asphalt mixes with RAP contents of 0% (Control), 30%, 40%, 50%, and 60% were prepared and tested at their Optimum Binder Content (OBC). The ITS results showed that the control mix had an ITS value of 1.2 MPa, with slight decreases observed in mixes containing up to 60% RAP, although these changes were not statistically significant (p=0.538>0.05). The TSR tests indicated that all mixes exceeded the minimum requirement of 80%. The Texas Overlay Test (TOT) was used to evaluate cracking performance and revealed that higher RAP contents had a negative impact on fatigue resistance. The 50% RAP mix exhibited the highest CFE, indicating that it has the best resistance to crack propagation despite having a lower number of cycles to failure. All mixes were categorized as falling into the Soft-crack-resistant quadrant, indicating their ability to resist crack propagation while being more susceptible to crack initiation.

Keywords: RAP, BC-II, HMA, TOT

Procedia PDF Downloads 19
20601 Rounded-off Measurements and Their Implication on Control Charts

Authors: Ran Etgar

Abstract:

The process of rounding off measurements in continuous variables is commonly encountered. Although it usually has minor effects, sometimes it can lead to poor outcomes in statistical process control using X ̅-chart. The traditional control limits can cause incorrect conclusions if applied carelessly. This study looks into the limitations of classical control limits, particularly the impact of asymmetry. An approach to determining the distribution function of the measured parameter (Y ̅) is presented, resulting in a more precise method to establish the upper and lower control limits. The proposed method, while slightly more complex than Shewhart's original idea, is still user-friendly and accurate and only requires the use of two straightforward tables.

Keywords: inaccurate measurement, SPC, statistical process control, rounded-off, control chart

Procedia PDF Downloads 31
20600 A Generative Adversarial Framework for Bounding Confounded Causal Effects

Authors: Yaowei Hu, Yongkai Wu, Lu Zhang, Xintao Wu

Abstract:

Causal inference from observational data is receiving wide applications in many fields. However, unidentifiable situations, where causal effects cannot be uniquely computed from observational data, pose critical barriers to applying causal inference to complicated real applications. In this paper, we develop a bounding method for estimating the average causal effect (ACE) under unidentifiable situations due to hidden confounders. We propose to parameterize the unknown exogenous random variables and structural equations of a causal model using neural networks and implicit generative models. Then, with an adversarial learning framework, we search the parameter space to explicitly traverse causal models that agree with the given observational distribution and find those that minimize or maximize the ACE to obtain its lower and upper bounds. The proposed method does not make any assumption about the data generating process and the type of the variables. Experiments using both synthetic and real-world datasets show the effectiveness of the method.

Keywords: average causal effect, hidden confounding, bound estimation, generative adversarial learning

Procedia PDF Downloads 184
20599 An Intelligent Thermal-Aware Task Scheduler in Multiprocessor System on a Chip

Authors: Sina Saadati

Abstract:

Multiprocessors Systems-On-Chips (MPSOCs) are used widely on modern computers to execute sophisticated software and applications. These systems include different processors for distinct aims. Most of the proposed task schedulers attempt to improve energy consumption. In some schedulers, the processor's temperature is considered to increase the system's reliability and performance. In this research, we have proposed a new method for thermal-aware task scheduling which is based on an artificial neural network (ANN). This method enables us to consider a variety of factors in the scheduling process. Some factors like ambient temperature, season (which is important for some embedded systems), speed of the processor, computing type of tasks and have a complex relationship with the final temperature of the system. This Issue can be solved using a machine learning algorithm. Another point is that our solution makes the system intelligent So that It can be adaptive. We have also shown that the computational complexity of the proposed method is cheap. As a consequence, It is also suitable for battery-powered systems.

Keywords: task scheduling, MOSOC, artificial neural network, machine learning, architecture of computers, artificial intelligence

Procedia PDF Downloads 99
20598 Energy Consumption Statistic of Gas-Solid Fluidized Beds through Computational Fluid Dynamics-Discrete Element Method Simulations

Authors: Lei Bi, Yunpeng Jiao, Chunjiang Liu, Jianhua Chen, Wei Ge

Abstract:

Two energy paths are proposed from thermodynamic viewpoints. Energy consumption means total power input to the specific system, and it can be decomposed into energy retention and energy dissipation. Energy retention is the variation of accumulated mechanical energy in the system, and energy dissipation is the energy converted to heat by irreversible processes. Based on the Computational Fluid Dynamics-Discrete Element Method (CFD-DEM) framework, different energy terms are quantified from the specific flow elements of fluid cells and particles as well as their interactions with the wall. Direct energy consumption statistics are carried out for both cold and hot flow in gas-solid fluidization systems. To clarify the statistic method, it is necessary to identify which system is studied: the particle-fluid system or the particle sub-system. For the cold flow, the total energy consumption of the particle sub-system can predict the onset of bubbling and turbulent fluidization, while the trends of local energy consumption can reflect the dynamic evolution of mesoscale structures. For the hot flow, different heat transfer mechanisms are analyzed, and the original solver is modified to reproduce the experimental results. The influence of the heat transfer mechanisms and heat source on energy consumption is also investigated. The proposed statistic method has proven to be energy-conservative and easy to conduct, and it is hopeful to be applied to other multiphase flow systems.

Keywords: energy consumption statistic, gas-solid fluidization, CFD-DEM, regime transition, heat transfer mechanism

Procedia PDF Downloads 63
20597 Oil Recovery Study by Low Temperature Carbon Dioxide Injection in High-Pressure High-Temperature Micromodels

Authors: Zakaria Hamdi, Mariyamni Awang

Abstract:

For the past decades, CO2 flooding has been used as a successful method for enhanced oil recovery (EOR). However, high mobility ratio and fingering effect are considered as important drawbacka of this process. Low temperature injection of CO2 into high temperature reservoirs may improve the oil recovery, but simulating multiphase flow in the non-isothermal medium is difficult, and commercial simulators are very unstable in these conditions. Furthermore, to best of authors’ knowledge, no experimental work was done to verify the results of the simulations and to understand the pore-scale process. In this paper, we present results of investigations on injection of low temperature CO2 into a high-pressure high-temperature micromodel with injection temperature range from 34 to 75 °F. Effect of temperature and saturation changes of different fluids are measured in each case. The results prove the proposed method. The injection of CO2 at low temperatures increased the oil recovery in high temperature reservoirs significantly. Also, CO2 rich phases available in the high temperature system can affect the oil recovery through the better sweep of the oil which is initially caused by penetration of LCO2 inside the system. Furthermore, no unfavorable effect was detected using this method. Low temperature CO2 is proposed to be used as early as secondary recovery.

Keywords: enhanced oil recovery, CO₂ flooding, micromodel studies, miscible flooding

Procedia PDF Downloads 348
20596 Implementation of Maqasid Sharia in Islamic Financial Institution in Indonesia

Authors: Deden Misbahudin Muayyad, Lavlimatria Esya

Abstract:

Up to the month of June 2015, Indonesia has 12 Islamic Commercial Banks, 22 Islamic Business Unit, 327 offices in 33 provinces. The initial purpose of the establishment of Islamic financial institutions is to achieve and the welfare of the people in this world and in the hereafter. To realize these goals, the Islamic financial institutions in every kind of aspect of product development and in terms of operations should be based on maqashid sharia, namely keeping the faith, keep the soul, keep the sense, maintain the property, keeping the offspring. To see whether Islamic banking grounded in maqasid sharia, the Islamic banking performance measurements based on the principles of maqasid sharia. Banking performance measurement is not only focusing on profit and other financial measures, but put other values of banks reflects the size of the benefit of non-profit in accordance with the bank's objectives. The measurement using the measurement of financial performance called maqasid sharia index. Maqasid syariah index is a model of Islamic banking performance measurement in accordance with the objectives and characteristics of Islamic banking. Maqasid syariah index was developed based on three main factors, namely the education of individuals, the creation of justice, the achievement of well-being, where the three factors were in accordance with the common goal of maqasid sharia is achieving prosperity and avoid evil. Maqasid syariah index shows that maqasid sharia approach can be a strategic alternative approach to describe how good the performance of the banking system and it can be implemented in the comprehensive policy strategy. This study uses a model of performance measurement framework based on maqasid syariah, in addition to financial performance measures that already exist. Methods to develop the idea of a performance measurement framework of Islamic banking by maqasid sharia is the Sekaran method. Operationally, the methods have now able to describe the elements that will be measured by this study. This is done by observing the behavior of the dimensions illustrated through a concept that has been set. These dimensions translate into derivative elements that can be observed and more scalable, so it can establish measurement indices. This research is descriptive quantitative. Techniques are being made to collect data in this paper is by using purposive sampling method, with 12 Islamic Commercial Banks that qualify as research samples. The financial data taken at 12 banks was sourced from the annual financial statements the period 2008 to 2012 with consideration of the database and ease of access to data. The ratio measured in this study only 7 ratio used in determining the performance of Islamic banking, namely: four ratio refers to the sharia objectives related to education. three ratio while again referring to sharia objectives related to the achievement of welfare. While other ratios associated with justice can not be used in this study because of the limited data used. Total overall calculation of performance indicators and performance ratios on each goal for each bank describes the maqasid syariah index.

Keywords: Islamic banking, Maslahah, maqashid syariah, maqashid syariah index

Procedia PDF Downloads 265
20595 VISMA: A Method for System Analysis in Early Lifecycle Phases

Authors: Walter Sebron, Hans Tschürtz, Peter Krebs

Abstract:

The choice of applicable analysis methods in safety or systems engineering depends on the depth of knowledge about a system, and on the respective lifecycle phase. However, the analysis method chain still shows gaps as it should support system analysis during the lifecycle of a system from a rough concept in pre-project phase until end-of-life. This paper’s goal is to discuss an analysis method, the VISSE Shell Model Analysis (VISMA) method, which aims at closing the gap in the early system lifecycle phases, like the conceptual or pre-project phase, or the project start phase. It was originally developed to aid in the definition of the system boundary of electronic system parts, like e.g. a control unit for a pump motor. Furthermore, it can be also applied to non-electronic system parts. The VISMA method is a graphical sketch-like method that stratifies a system and its parts in inner and outer shells, like the layers of an onion. It analyses a system in a two-step approach, from the innermost to the outermost components followed by the reverse direction. To ensure a complete view of a system and its environment, the VISMA should be performed by (multifunctional) development teams. To introduce the method, a set of rules and guidelines has been defined in order to enable a proper shell build-up. In the first step, the innermost system, named system under consideration (SUC), is selected, which is the focus of the subsequent analysis. Then, its directly adjacent components, responsible for providing input to and receiving output from the SUC, are identified. These components are the content of the first shell around the SUC. Next, the input and output components to the components in the first shell are identified and form the second shell around the first one. Continuing this way, shell by shell is added with its respective parts until the border of the complete system (external border) is reached. Last, two external shells are added to complete the system view, the environment and the use case shell. This system view is also stored for future use. In the second step, the shells are examined in the reverse direction (outside to inside) in order to remove superfluous components or subsystems. Input chains to the SUC, as well as output chains from the SUC are described graphically via arrows, to highlight functional chains through the system. As a result, this method offers a clear and graphical description and overview of a system, its main parts and environment; however, the focus still remains on a specific SUC. It helps to identify the interfaces and interfacing components of the SUC, as well as important external interfaces of the overall system. It supports the identification of the first internal and external hazard causes and causal chains. Additionally, the method promotes a holistic picture and cross-functional understanding of a system, its contributing parts, internal relationships and possible dangers within a multidisciplinary development team.

Keywords: analysis methods, functional safety, hazard identification, system and safety engineering, system boundary definition, system safety

Procedia PDF Downloads 223
20594 Iron Recovery from Red Mud As Zero-Valent Iron Metal Powder Using Direct Electrochemical Reduction Method

Authors: Franky Michael Hamonangan Siagian, Affan Maulana, Himawan Tri Bayu Murti Petrus, Panut Mulyono, Widi Astuti

Abstract:

In this study, the feasibility of the direct electrowinning method was used to produce zero-valent iron from red mud. The bauxite residue sample came from the Tayan mine, Indonesia, which contains high hematite (Fe₂O₃). Before electrolysis, the samples were characterized by various analytical techniques (ICP-AES, SEM, XRD) to determine their chemical composition and mineralogy. The direct electrowinning method of red mud suspended in NaOH was introduced at low temperatures ranging from 30 - 110 °C. Variations of current density, red mud: NaOH ratio and temperature were carried out to determine the optimum operation of the direct electrowinning process. Cathode deposits and residues in electrochemical cells were analyzed using XRD, XRF, and SEM to determine the chemical composition and current recovery. The low-temperature electrolysis current efficiency on Redmud can reach 20% recovery at a current density of 920,945 A/m². The moderate performance of the process was investigated with red mud, which was attributed to the troublesome adsorption of red mud particles on the cathode, making the reduction far less efficient than that with hematite.

Keywords: alumina, red mud, electrochemical reduction, iron production

Procedia PDF Downloads 75
20593 Sequential Covering Algorithm for Nondifferentiable Global Optimization Problem and Applications

Authors: Mohamed Rahal, Djaouida Guetta

Abstract:

In this paper, the one-dimensional unconstrained global optimization problem of continuous functions satifying a Hölder condition is considered. We extend the algorithm of sequential covering SCA for Lipschitz functions to a large class of Hölder functions. The convergence of the method is studied and the algorithm can be applied to systems of nonlinear equations. Finally, some numerical examples are presented and illustrate the efficiency of the present approach.

Keywords: global optimization, Hölder functions, sequential covering method, systems of nonlinear equations

Procedia PDF Downloads 367
20592 Describing Cognitive Decline in Alzheimer's Disease via a Picture Description Writing Task

Authors: Marielle Leijten, Catherine Meulemans, Sven De Maeyer, Luuk Van Waes

Abstract:

For the diagnosis of Alzheimer's disease (AD), a large variety of neuropsychological tests are available. In some of these tests, linguistic processing - both oral and written - is an important factor. Language disturbances might serve as a strong indicator for an underlying neurodegenerative disorder like AD. However, the current diagnostic instruments for language assessment mainly focus on product measures, such as text length or number of errors, ignoring the importance of the process that leads to written or spoken language production. In this study, it is our aim to describe and test differences between cognitive and impaired elderly on the basis of a selection of writing process variables (inter- and intrapersonal characteristics). These process variables are mainly related to pause times, because the number, length, and location of pauses have proven to be an important indicator of the cognitive complexity of a process. Method: Participants that were enrolled in our research were chosen on the basis of a number of basic criteria necessary to collect reliable writing process data. Furthermore, we opted to match the thirteen cognitively impaired patients (8 MCI and 5 AD) with thirteen cognitively healthy elderly. At the start of the experiment, participants were each given a number of tests, such as the Mini-Mental State Examination test (MMSE), the Geriatric Depression Scale (GDS), the forward and backward digit span and the Edinburgh Handedness Inventory (EHI). Also, a questionnaire was used to collect socio-demographic information (age, gender, eduction) of the subjects as well as more details on their level of computer literacy. The tests and questionnaire were followed by two typing tasks and two picture description tasks. For the typing tasks participants had to copy (type) characters, words and sentences from a screen, whereas the picture description tasks each consisted of an image they had to describe in a few sentences. Both the typing and the picture description tasks were logged with Inputlog, a keystroke logging tool that allows us to log and time stamp keystroke activity to reconstruct and describe text production processes. The main rationale behind keystroke logging is that writing fluency and flow reveal traces of the underlying cognitive processes. This explains the analytical focus on pause (length, number, distribution, location, etc.) and revision (number, type, operation, embeddedness, location, etc.) characteristics. As in speech, pause times are seen as indexical of cognitive effort. Results. Preliminary analysis already showed some promising results concerning pause times before, within and after words. For all variables, mixed effects models were used that included participants as a random effect and MMSE scores, GDS scores and word categories (such as determiners and nouns) as a fixed effect. For pause times before and after words cognitively impaired patients paused longer than healthy elderly. These variables did not show an interaction effect between the group participants (cognitively impaired or healthy elderly) belonged to and word categories. However, pause times within words did show an interaction effect, which indicates pause times within certain word categories differ significantly between patients and healthy elderly.

Keywords: Alzheimer's disease, keystroke logging, matching, writing process

Procedia PDF Downloads 362
20591 Comparison between Hardy-Cross Method and Water Software to Solve a Pipe Networking Design Problem for a Small Town

Authors: Ahmed Emad Ahmed, Zeyad Ahmed Hussein, Mohamed Salama Afifi, Ahmed Mohammed Eid

Abstract:

Water has a great importance in life. In order to deliver water from resources to the users, many procedures should be taken by the water engineers. One of the main procedures to deliver water to the community is by designing pressurizer pipe networks for water. The main aim of this work is to calculate the water demand of a small town and then design a simple water network to distribute water resources among the town with the smallest losses. Literature has been mentioned to cover the main point related to water distribution. Moreover, the methodology has introduced two approaches to solve the research problem, one by the iterative method of Hardy-cross and the other by water software Pipe Flow. The results have introduced two main designs to satisfy the same research requirements. Finally, the researchers have concluded that the use of water software provides more abilities and options for water engineers.

Keywords: looping pipe networks, hardy cross networks accuracy, relative error of hardy cross method

Procedia PDF Downloads 159
20590 Analysis and Re-Design Ergonomic Mineral Water Gallon Trolley

Authors: Dessy Laksyana Utami

Abstract:

Manual material handling activities often make it difficult for humans to work like this. Muscle injury due to incorrect posture.Workers need to facilitate their activities. One tool to assist their activities in the transportation of ordinary materials is a trolley. This tool is very useful because it can be used.It can bring many items without having to spend more energy to operate it. Very Comfortable used a trolley in the community. But the old design still have a complaint by worker, because lack of grip and capacity. After posture analysis with the REBA method, the value of risk need to be increased is obtained tool. Re design use Indonesian anthropometric data with the 50th percentile.

Keywords: Material Handling, REBA method, postural assessment, Trolley.

Procedia PDF Downloads 133
20589 A Development of Personalized Edutainment Contents through Storytelling

Authors: Min Kyeong Cha, Ju Yeon Mun, Seong Baeg Kim

Abstract:

Recently, ‘play of learning’ became important and is emphasized as a useful learning tool. Therefore, interest in edutainment contents is growing. Storytelling is considered first as a method that improves the transmission of information and learner's interest when planning edutainment contents. In this study, we designed edutainment contents in the form of an adventure game that applies the storytelling method. This content provides questions and items constituted dynamically and reorganized learning contents through analysis of test results. It allows learners to solve various questions through effective iterative learning. As a result, the learners can reach mastery learning.

Keywords: storytelling, edutainment, mastery learning, computer operating principle

Procedia PDF Downloads 311
20588 Teacher Knowledge: Unbridling Teacher Agency in the Context of Professional Development for Transformative Teaching and Learning

Authors: Bernice Badal

Abstract:

This article addresses a persistent challenge related to teacher agency in knowledge acquisition in professional development (PD) workshops in contexts of educational change, given that scholarship identifies a need for more teacher involvement and amplification of teacher's voices. Theoretical concepts are drawn from Bandura’s Social cognitive theory, incorporating the triadic causation model of agency to examine the reciprocal nature of the context, teacher characteristics, and systemic influences that shape how knowledge is transmitted and acquired in PD workshops. This qualitative study, using a mix of classroom observations and interviews, explored the political, contextual, and personal characteristics of teacher agency in PD through an analysis of data extracted from a PhD study. The narratives of six teachers from three township schools are examined to show how PD efforts in South Africa have failed to take account of the holistic development of teacher agency in knowledge dissemination and how this shapes teacher self-efficacy beliefs about being able to masterfully apply the tenets of the reform. Agency, teacher voice, and contextual considerations were used as markers of the quality of the training provided to understand how knowledge is acquired and meaning is made. The findings suggest that systemic influences of institutionally imposed PD offer partial understandings of the reform, which is offered in traditional formats that do not consider teacher empowerment in knowledge production and the development of teacher agency. Common in all the participants’ responses is the need for more information and training on the prescribed approach for teaching English as a second language; however, this paper holds the view that more information may not solve teachers’ dilemmas. Accordingly, it recommends a restructuring of the programme with facilitators being more cognisant of teacher agency for the development of transformative teachers. The findings of the study contribute to the field of teacher knowledge, teacher training, and professional development in the context of educational reforms.

Keywords: teacher professional development, teacher voice, teacher agency, educational reforms, teacher knowledge

Procedia PDF Downloads 65
20587 Cuckoo Search (CS) Optimization Algorithm for Solving Constrained Optimization

Authors: Sait Ali Uymaz, Gülay Tezel

Abstract:

This paper presents the comparison results on the performance of the Cuckoo Search (CS) algorithm for constrained optimization problems. For constraint handling, CS algorithm uses penalty method. CS algorithm is tested on thirteen well-known test problems and the results obtained are compared to Particle Swarm Optimization (PSO) algorithm. Mean, best, median and worst values were employed for the analyses of performance.

Keywords: cuckoo search, particle swarm optimization, constrained optimization problems, penalty method

Procedia PDF Downloads 551
20586 Epileptic Seizure Prediction by Exploiting Signal Transitions Phenomena

Authors: Mohammad Zavid Parvez, Manoranjan Paul

Abstract:

A seizure prediction method is proposed by extracting global features using phase correlation between adjacent epochs for detecting relative changes and local features using fluctuation/deviation within an epoch for determining fine changes of different EEG signals. A classifier and a regularization technique are applied for the reduction of false alarms and improvement of the overall prediction accuracy. The experiments show that the proposed method outperforms the state-of-the-art methods and provides high prediction accuracy (i.e., 97.70%) with low false alarm using EEG signals in different brain locations from a benchmark data set.

Keywords: Epilepsy, seizure, phase correlation, fluctuation, deviation.

Procedia PDF Downloads 464
20585 Comparison of Power Generation Status of Photovoltaic Systems under Different Weather Conditions

Authors: Zhaojun Wang, Zongdi Sun, Qinqin Cui, Xingwan Ren

Abstract:

Based on multivariate statistical analysis theory, this paper uses the principal component analysis method, Mahalanobis distance analysis method and fitting method to establish the photovoltaic health model to evaluate the health of photovoltaic panels. First of all, according to weather conditions, the photovoltaic panel variable data are classified into five categories: sunny, cloudy, rainy, foggy, overcast. The health of photovoltaic panels in these five types of weather is studied. Secondly, a scatterplot of the relationship between the amount of electricity produced by each kind of weather and other variables was plotted. It was found that the amount of electricity generated by photovoltaic panels has a significant nonlinear relationship with time. The fitting method was used to fit the relationship between the amount of weather generated and the time, and the nonlinear equation was obtained. Then, using the principal component analysis method to analyze the independent variables under five kinds of weather conditions, according to the Kaiser-Meyer-Olkin test, it was found that three types of weather such as overcast, foggy, and sunny meet the conditions for factor analysis, while cloudy and rainy weather do not satisfy the conditions for factor analysis. Therefore, through the principal component analysis method, the main components of overcast weather are temperature, AQI, and pm2.5. The main component of foggy weather is temperature, and the main components of sunny weather are temperature, AQI, and pm2.5. Cloudy and rainy weather require analysis of all of their variables, namely temperature, AQI, pm2.5, solar radiation intensity and time. Finally, taking the variable values in sunny weather as observed values, taking the main components of cloudy, foggy, overcast and rainy weather as sample data, the Mahalanobis distances between observed value and these sample values are obtained. A comparative analysis was carried out to compare the degree of deviation of the Mahalanobis distance to determine the health of the photovoltaic panels under different weather conditions. It was found that the weather conditions in which the Mahalanobis distance fluctuations ranged from small to large were: foggy, cloudy, overcast and rainy.

Keywords: fitting, principal component analysis, Mahalanobis distance, SPSS, MATLAB

Procedia PDF Downloads 141
20584 Improving the Design of Blood Pressure and Blood Saturation Monitors

Authors: L. Parisi

Abstract:

A blood pressure monitor or sphygmomanometer can be either manual or automatic, employing respectively either the auscultatory method or the oscillometric method. The manual version of the sphygmomanometer involves an inflatable cuff with a stethoscope adopted to detect the sounds generated by the arterial walls to measure blood pressure in an artery. An automatic sphygmomanometer can be effectively used to monitor blood pressure through a pressure sensor, which detects vibrations provoked by oscillations of the arterial walls. The pressure sensor implemented in this device improves the accuracy of the measurements taken.

Keywords: blood pressure, blood saturation, sensors, actuators, design improvement

Procedia PDF Downloads 451
20583 Erodibility Analysis of Cikapundung Hulu: A Study Case of Mekarwangi Catchment Area

Authors: Shantosa Yudha Siswanto, Rachmat Harryanto

Abstract:

The aim of the research was to investigate the effect of land use and slope steepness on soil erodibility index. The research was conducted from September to December 2013 in Mekarwangi catchment area, sub watershed of Cikapundung Hulu, Indonesia. The study was carried out using descriptive method. Physiographic free survey method was used as survey method, it was a survey based on land physiographic appearance. Soil sampling was carried out into transect on the similarity of slope without calculating the range between points of observation. Soil samples were carried onto three classes of land use such as: forest, plantation and dry cultivation area. Each land use consists of three slope classes such as: 8-15%, 16-25%, and 26-40% class. Five samples of soil were taken from each of them, resulting 45 points of observation. The result of the research showed that type of land use and slope classes gave different effect on soil erodibility. The highest C-organic and permeability was found on forest with slope 16-25%. Slope of 8-15% with forest land use give the lowest effect on soil erodibility.

Keywords: land use, slope, erodibility, erosion

Procedia PDF Downloads 246
20582 A Genetic Algorithm Based Ensemble Method with Pairwise Consensus Score on Malware Cacophonous Labels

Authors: Shih-Yu Wang, Shun-Wen Hsiao

Abstract:

In the field of cybersecurity, there exists many vendors giving malware samples classified results, namely naming after the label that contains some important information which is also called AV label. Lots of researchers relay on AV labels for research. Unfortunately, AV labels are too cluttered. They do not have a fixed format and fixed naming rules because the naming results were based on each classifiers' viewpoints. A way to fix the problem is taking a majority vote. However, voting can sometimes create problems of bias. Thus, we create a novel ensemble approach which does not rely on the cacophonous naming result but depend on group identification to aggregate everyone's opinion. To achieve this purpose, we develop an scoring system called Pairwise Consensus Score (PCS) to calculate result similarity. The entire method architecture combine Genetic Algorithm and PCS to find maximum consensus in the group. Experimental results revealed that our method outperformed the majority voting by 10% in term of the score.

Keywords: genetic algorithm, ensemble learning, malware family, malware labeling, AV labels

Procedia PDF Downloads 83
20581 Development and Validation of a Green Analytical Method for the Analysis of Daptomycin Injectable by Fourier-Transform Infrared Spectroscopy (FTIR)

Authors: Eliane G. Tótoli, Hérida Regina N. Salgado

Abstract:

Daptomycin is an important antimicrobial agent used in clinical practice nowadays, since it is very active against some Gram-positive bacteria that are particularly challenges for the medicine, such as methicillin-resistant Staphylococcus aureus (MRSA) and vancomycin-resistant Enterococci (VRE). The importance of environmental preservation has receiving special attention since last years. Considering the evident need to protect the natural environment and the introduction of strict quality requirements regarding analytical procedures used in pharmaceutical analysis, the industries must seek environmentally friendly alternatives in relation to the analytical methods and other processes that they follow in their routine. In view of these factors, green analytical chemistry is prevalent and encouraged nowadays. In this context, infrared spectroscopy stands out. This is a method that does not use organic solvents and, although it is formally accepted for the identification of individual compounds, also allows the quantitation of substances. Considering that there are few green analytical methods described in literature for the analysis of daptomycin, the aim of this work was the development and validation of a green analytical method for the quantification of this drug in lyophilized powder for injectable solution, by Fourier-transform infrared spectroscopy (FT-IR). Method: Translucent potassium bromide pellets containing predetermined amounts of the drug were prepared and subjected to spectrophotometric analysis in the mid-infrared region. After obtaining the infrared spectrum and with the assistance of the IR Solution software, quantitative analysis was carried out in the spectral region between 1575 and 1700 cm-1, related to a carbonyl band of the daptomycin molecule, and this band had its height analyzed in terms of absorbance. The method was validated according to ICH guidelines regarding linearity, precision (repeatability and intermediate precision), accuracy and robustness. Results and discussion: The method showed to be linear (r = 0.9999), precise (RSD% < 2.0), accurate and robust, over a concentration range from 0.2 to 0.6 mg/pellet. In addition, this technique does not use organic solvents, which is one great advantage over the most common analytical methods. This fact contributes to minimize the generation of organic solvent waste by the industry and thereby reduces the impact of its activities on the environment. Conclusion: The validated method proved to be adequate to quantify daptomycin in lyophilized powder for injectable solution and can be used for its routine analysis in quality control. In addition, the proposed method is environmentally friendly, which is in line with the global trend.

Keywords: daptomycin, Fourier-transform infrared spectroscopy, green analytical chemistry, quality control, spectrometry in IR region

Procedia PDF Downloads 378
20580 Probabilistic Graphical Model for the Web

Authors: M. Nekri, A. Khelladi

Abstract:

The world wide web network is a network with a complex topology, the main properties of which are the distribution of degrees in power law, A low clustering coefficient and a weak average distance. Modeling the web as a graph allows locating the information in little time and consequently offering a help in the construction of the research engine. Here, we present a model based on the already existing probabilistic graphs with all the aforesaid characteristics. This work will consist in studying the web in order to know its structuring thus it will enable us to modelize it more easily and propose a possible algorithm for its exploration.

Keywords: clustering coefficient, preferential attachment, small world, web community

Procedia PDF Downloads 269
20579 Nonlinear Dynamic Analysis of Base-Isolated Structures Using a Mixed Integration Method: Stability Aspects and Computational Efficiency

Authors: Nicolò Vaiana, Filip C. Filippou, Giorgio Serino

Abstract:

In order to reduce numerical computations in the nonlinear dynamic analysis of seismically base-isolated structures, a Mixed Explicit-Implicit time integration Method (MEIM) has been proposed. Adopting the explicit conditionally stable central difference method to compute the nonlinear response of the base isolation system, and the implicit unconditionally stable Newmark’s constant average acceleration method to determine the superstructure linear response, the proposed MEIM, which is conditionally stable due to the use of the central difference method, allows to avoid the iterative procedure generally required by conventional monolithic solution approaches within each time step of the analysis. The main aim of this paper is to investigate the stability and computational efficiency of the MEIM when employed to perform the nonlinear time history analysis of base-isolated structures with sliding bearings. Indeed, in this case, the critical time step could become smaller than the one used to define accurately the earthquake excitation due to the very high initial stiffness values of such devices. The numerical results obtained from nonlinear dynamic analyses of a base-isolated structure with a friction pendulum bearing system, performed by using the proposed MEIM, are compared to those obtained adopting a conventional monolithic solution approach, i.e. the implicit unconditionally stable Newmark’s constant acceleration method employed in conjunction with the iterative pseudo-force procedure. According to the numerical results, in the presented numerical application, the MEIM does not have stability problems being the critical time step larger than the ground acceleration one despite of the high initial stiffness of the friction pendulum bearings. In addition, compared to the conventional monolithic solution approach, the proposed algorithm preserves its computational efficiency even when it is adopted to perform the nonlinear dynamic analysis using a smaller time step.

Keywords: base isolation, computational efficiency, mixed explicit-implicit method, partitioned solution approach, stability

Procedia PDF Downloads 273
20578 An Alternative Method for Computing Clothoids

Authors: Gerardo Casal, Miguel E. Vázquez-Méndez

Abstract:

The clothoid (also known as Cornu spiral or Euler spiral) is a curve that is characterized because its curvature is proportional to its length. This property makes that it would be widely used as transition curve for designing the layout of roads and railway tracks. In this work, from the geometrical property characterizing the clothoid, its parametric equations are obtained and two algorithms to compute it are compared. The first (classical), is widely used in Surveying Schools and it is based on the use of explicit formulas obtained from Taylor expansions of sine and cosine functions. The second one (alternative) is a very simple algorithm, based on the numerical solution of the initial value problems giving the clothoid parameterization. Both methods are compared in some typical surveying problems. The alternative method does not use complex formulas and so it is conceptually very simple and easy to apply. It gives good results, even if the classical method goes wrong (if the quotient between length and radius of curvature is high), needs no subsequent translations nor rotations and, consequently, it seems an efficient tool for designing the layout of roads and railway tracks.

Keywords: transition curves, railroad and highway engineering, Runge-Kutta methods

Procedia PDF Downloads 277
20577 The Application of Patterned Injuries in Reconstruction of Motorcycle Accidents

Authors: Chun-Liang Wu, Kai-Ping Shaw, Cheng-Ping Yu, Wu-Chien Chien, Hsiao-Ting Chen, Shao-Huang Wu

Abstract:

Objective: This study analyzed three criminal judicial cases. We applied the patterned injuries of the rider to demonstrate the facts of each accident, reconstruct the scenes, and pursue the truth. Methods: Case analysis, a method that collects evidence and reasons the results in judicial procedures, then the importance of the pattern of injury as evidence will be compared and evaluated. The patterned injuries analysis method is to compare the collision situation between an object and human body injuries to determine whether the characteristics can reproduce the unique pattern of injury. Result: Case 1: Two motorcycles, A and B, head-on collided; rider A dead, and rider B was accused. During the prosecutor’s investigation, the defendant learned that rider A had an 80 mm open wound on his neck. During the court trial, the defendant requested copies of the case file and found out that rider A had a large contusion on his chest wall, and the cause of death was traumatic hemothorax and abdominal wall contusion. The defendant compared all the evidence at the scene and determined that the injury was obviously not caused by the collision of the body or the motorcycle of rider B but that rider was out of control and injured himself when he crossed the double yellow line. In this case, the defendant was innocent in the High Court judgment in April 2022. Case 2: Motorcycles C and D head-on crashed, and rider C died of massive abdominal bleeding. The prosecutor decided that rider C was driving under the influence (DUI), but rider D was negligent and sued rider D. The defendant requested the copies’ file and found the special phenomenon that the front wheel of motorcycle C was turned left. The defendant’s injuries were a left facial bone fracture, a left femur fracture, and other injuries on the left side. The injuries were of human-vehicle separation and human-vehicle collision, which proved that rider C suddenly turned left when the two motorcycles approached, knocked down motorcycle D, and the defendant flew forward. Case 3: Motorcycle E and F’s rear end collided, the front rider E was sentenced to 3 months, and the rear rider F sued rider E for more than 7 million N.T. The defendant found in the copies’ file that the injury of rider F was the left tibial platform fracture, etc., and then proved that rider F made the collision with his left knee, causing motorcycle E to fall out of control. This evidence was accepted by the court and is still on trial. Conclusion: The application of patterned injuries in the reconstruction of a motorcycle accident could discover the truth and provide the basis for judicial justice. The cases and methods could be the reference for the policy of preventing traffic accident casualties.

Keywords: judicial evidence, patterned injuries analysis, accident reconstruction, fatal motorcycle injuries

Procedia PDF Downloads 77
20576 Step Height Calibration Using Hamming Window: Band-Pass Filter

Authors: Dahi Ghareab Abdelsalam Ibrahim

Abstract:

Calibration of step heights with high accuracy is needed for many applications in the industry. In general, step height consists of three bands: pass band, transition band (roll-off), and stop band. Abdelsalam used a convolution of the transfer functions of both Chebyshev type 2 and elliptic filters with WFF of the Fresnel transform in the frequency domain for producing a steeper roll-off with the removal of ripples in the pass band- and stop-bands. In this paper, we used a new method based on the Hamming window: band-pass filter for calibration of step heights in terms of perfect adjustment of pass-band, roll-off, and stop-band. The method is applied to calibrate a nominal step height of 40 cm. The step height is measured first by asynchronous dual-wavelength phase-shift interferometry. The measured step height is then calibrated by the simulation of the Hamming window: band-pass filter. The spectrum of the simulated band-pass filter is simulated at N = 881 and f0 = 0.24. We can conclude that the proposed method can calibrate any step height by adjusting only two factors which are N and f0.

Keywords: optical metrology, step heights, hamming window, band-pass filter

Procedia PDF Downloads 79
20575 Flow Field Analysis of Different Intake Bump (Compression Surface) Configurations on a Supersonic Aircraft

Authors: Mudassir Ghafoor, Irsalan Arif, Shuaib Salamat

Abstract:

This paper presents modeling and analysis of different intake bump (compression surface) configurations and comparison with an existing supersonic aircraft having bump intake configuration. Many successful aircraft models have shown that Diverter less Supersonic Inlet (DSI) as compared to conventional intake can reduce weight, complexity and also maintenance cost. The research is divided into two parts. In the first part, four different intake bumps are modeled for comparative analysis keeping in view the consistency of outer perimeter dimensions of fighter aircraft and various characteristics such as flow behavior, boundary layer diversion and pressure recovery are analyzed. In the second part, modeled bumps are integrated with intake duct for performance analysis and comparison with existing supersonic aircraft data is carried out. The bumps are named as uniform large (Config 1), uniform small (Config 2), uniform sharp (Config 3), non-uniform (Config 4) based on their geometric features. Analysis is carried out at different Mach Numbers to analyze flow behavior in subsonic and supersonic regime. Flow behavior, boundary layer diversion and Pressure recovery are examined for each bump characteristics, and comparative study is carried out. The analysis reveals that at subsonic speed, Config 1 and Config 2 give similar pressure recoveries as diverterless supersonic intake, but difference in pressure recoveries becomes significant at supersonic speed. It was concluded from research that Config 1 gives better results as compared to Config 3. Also, higher amplitude (Config 1) is preferred over lower (Config 2 and 4). It was observed that maximum height of bump is preferred to be placed near cowl lip of intake duct.

Keywords: bump intake, boundary layer, computational fluid dynamics, diverter-less supersonic inlet

Procedia PDF Downloads 240