Search results for: time domain reflectometry (TDR)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 19215

Search results for: time domain reflectometry (TDR)

18375 Merging of Results in Distributed Information Retrieval Systems

Authors: Larbi Guezouli, Imane Azzouz

Abstract:

This work is located in the domain of distributed information retrieval ‘DIR’. A simplified view of the DIR requires a multi-search in a set of collections, which forces the system to analyze results found in these collections, and merge results back before sending them to the user in a single list. Our work is to find a fusion method based on the relevance score of each result received from collections and the relevance of the local search engine of each collection.

Keywords: information retrieval, distributed IR systems, merging results, datamining

Procedia PDF Downloads 331
18374 Time-Frequency Feature Extraction Method Based on Micro-Doppler Signature of Ground Moving Targets

Authors: Ke Ren, Huiruo Shi, Linsen Li, Baoshuai Wang, Yu Zhou

Abstract:

Since some discriminative features are required for ground moving targets classification, we propose a new feature extraction method based on micro-Doppler signature. Firstly, the time-frequency analysis of measured data indicates that the time-frequency spectrograms of the three kinds of ground moving targets, i.e., single walking person, two people walking and a moving wheeled vehicle, are discriminative. Then, a three-dimensional time-frequency feature vector is extracted from the time-frequency spectrograms to depict these differences. At last, a Support Vector Machine (SVM) classifier is trained with the proposed three-dimensional feature vector. The classification accuracy to categorize ground moving targets into the three kinds of the measured data is found to be over 96%, which demonstrates the good discriminative ability of the proposed micro-Doppler feature.

Keywords: micro-doppler, time-frequency analysis, feature extraction, radar target classification

Procedia PDF Downloads 402
18373 Advanced Stability Criterion for Time-Delayed Systems of Neutral Type and Its Application

Authors: M. J. Park, S. H. Lee, C. H. Lee, O. M. Kwon

Abstract:

This paper investigates stability problem for linear systems of neutral type with time-varying delay. By constructing various Lyapunov-Krasovskii functional, and utilizing some mathematical techniques, the sufficient stability conditions for the systems are established in terms of linear matrix inequalities (LMIs), which can be easily solved by various effective optimization algorithms. Finally, some illustrative examples are given to show the effectiveness of the proposed criterion.

Keywords: neutral systems, time-delay, stability, Lyapnov method, LMI

Procedia PDF Downloads 344
18372 Time Variance and Spillover Effects between International Crude Oil Price and Ten Emerging Equity Markets

Authors: Murad A. Bein

Abstract:

This paper empirically examines the time-varying relationship and spillover effects between the international crude oil price and ten emerging equity markets, namely three oil-exporting countries (Brazil, Mexico, and Russia) and seven Central and Eastern European (CEE) countries (Bulgaria, Croatia, Czech Republic, Hungary, Poland, Romania, and Slovakia). The results revealed that there are spillover effects from oil markets into almost all emerging equity markets save Slovakia. Besides, the oil supply glut had a homogenous effect on the emerging markets, both net oil-exporting, and oil-importing countries (CEE). Further, the time variance drastically increased during financial turmoil. Indeed, the time variance remained high from 2009 to 2012 in response to aggregate demand shocks (global financial crisis and Eurozone debt crisis) and quantitative easing measures. Interestingly, the time variance was slightly higher for the oil-exporting countries than for some of the CEE countries. Decision-makers in emerging economies should therefore seek policy coordination when dealing with financial turmoil.

Keywords: crude oil, spillover effects, emerging equity, time-varying, aggregate demand shock

Procedia PDF Downloads 120
18371 A Probabilistic Study on Time to Cover Cracking Due to Corrosion

Authors: Chun-Qing Li, Hassan Baji, Wei Yang

Abstract:

Corrosion of steel in reinforced concrete structures is a major problem worldwide. The volume expansion of corrosion products causes concrete cover cracking, which could lead to delamination of concrete cover. The time to cover cracking plays a key role to the assessment of serviceability of reinforced concrete structures subjected to corrosion. Many analytical, numerical, and empirical models have been developed to predict the time to cracking initiation due to corrosion. In this study, a numerical model based on finite element modeling of corrosion-induced cracking process is used. In order to predict the service life based on time to cover initiation, the numerical approach is coupled with a probabilistic procedure. In this procedure, all the influential factors affecting time to cover cracking are modeled as random variables. The results show that the time to cover cracking is highly variables. It is also shown that rust product expansion ratio and the size of more porous concrete zone around the rebar are the most influential factors in predicting service life of corrosion-affected structures.

Keywords: corrosion, crack width, probabilistic, service life

Procedia PDF Downloads 205
18370 Impact of Reclamation on the Water Exchange in Bohai Bay

Authors: Luyao Liu, Dekui Yuan, Xu Li

Abstract:

As one of the most important bays of China, the water exchange capacity of Bohai Bay can influence the economic development and urbanization of surrounding cities. However, the rapid reclamation has influenced the weak water exchange capacity of this semi-enclosed bay in recent years. This paper sets two hydrodynamic models of Bohai Bay with two shorelines before and after reclamation. The mean value and distribution of Turn-over Time, the distribution of residual current, and the feature of the tracer path are compared. After comparison, it is found that Bohai Bay keeps these characteristics; the spending time of water exchange in the northern is longer than southern, and inshore is longer than offshore. However, the mean water exchange time becomes longer after reclamation. In addition, the material spreading is blocked because of the inwardly extending shorelines, and the direction changed from along the shoreline to towards the center after reclamation.

Keywords: Bohai Bay, water exchange, reclamation, turn-over time

Procedia PDF Downloads 136
18369 Design and Development of Sustained Release Floating Tablet of Stavudine

Authors: Surajj Sarode, G. Vidya Sagar, G. P. Vadnere

Abstract:

The purpose of the present study was to prolong the gastric residence time of Stavudine by developing gastric floating drug delivery system (GFDDS). Moreover, to study influence of different polymers on its release rate using gas-forming agents, like sodium bicarbonate, citric acid. Floating tablets were prepared by wet granulation method using PVP K-30 as a binder and the other polymers include Pullulan Gum, HPMC K100M, six different formulations with the varying concentrations of polymers were prepared and the tablets were evaluated in terms of their pre-compression parameters like bulk density, tapped density, Haunsner ratio, angle of repose, compressibility index, post compression physical characteristics, in vitro release, buoyancy, floating lag time (FLT), total floating time (TFT) and swelling index. All the formulations showed good floating lag time i.e. less than 3 mins. The batch containing combination of Pullulan Gum and HPMC 100M (i.e. F-6) showed total floating lag time more than 12 h., the highest swelling index among all the prepared batches. The drug release was found to follow zero order kinetics.

Keywords: Suavudine, floating, total floating time (TFT), gastric residence

Procedia PDF Downloads 391
18368 Modeling and Analysis of Laser Sintering Process Scanning Time for Optimal Planning and Control

Authors: Agarana Michael C., Akinlabi Esther T., Pule Kholopane

Abstract:

In order to sustain the advantages of an advanced manufacturing technique, such as laser sintering, minimization of total processing cost of the parts being produced is very important. An efficient time management would usually very important in optimal cost attainment which would ultimately result in an efficient advanced manufacturing process planning and control. During Laser Scanning Process Scanning (SLS) procedures it is possible to adjust various manufacturing parameters which are used to influence the improvement of various mechanical and other properties of the products. In this study, Modelling and mathematical analysis, including sensitivity analysis, of the laser sintering process time were carried out. The results of the analyses were represented with graphs, from where conclusions were drawn. It was specifically observed that achievement of optimal total scanning time is key for economic efficiency which is required for sustainability of the process.

Keywords: modeling and analysis, optimal planning and control, laser sintering process, scanning time

Procedia PDF Downloads 95
18367 Vertically Coupled III-V/Silicon Single Mode Laser with a Hybrid Grating Structure

Authors: Zekun Lin, Xun Li

Abstract:

Silicon photonics has gained much interest and extensive research for a promising aspect for fabricating compact, high-speed and low-cost photonic devices compatible with complementary metal-oxide-semiconductor (CMOS) process. Despite the remarkable progress made on the development of silicon photonics, high-performance, cost-effective, and reliable silicon laser sources are still missing. In this work, we present a 1550 nm III-V/silicon laser design with stable single-mode lasing property and robust and high-efficiency vertical coupling. The InP cavity consists of two uniform Bragg grating sections at sides for mode selection and feedback, as well as a central second-order grating for surface emission. A grating coupler is etched on the SOI waveguide by which the light coupling between the parallel III-V and SOI is reached vertically rather than by evanescent wave coupling. Laser characteristic is simulated and optimized by the traveling-wave model (TWM) and a Green’s function analysis as well as a 2D finite difference time domain (FDTD) method for the coupling process. The simulation results show that single-mode lasing with SMSR better than 48dB is achievable, and the threshold current is less than 15mA with a slope efficiency of around 0.13W/A. The coupling efficiency is larger than 42% and possesses a high tolerance with less than 10% reduction for 10 um horizontal or 15 um vertical dislocation. The design can be realized by standard flip-chip bonding techniques without co-fabrication of III-V and silicon or precise alignment.

Keywords: III-V/silicon integration, silicon photonics, single mode laser, vertical coupling

Procedia PDF Downloads 151
18366 On Privacy-Preserving Search in the Encrypted Domain

Authors: Chun-Shien Lu

Abstract:

Privacy-preserving query has recently received considerable attention in the signal processing and multimedia community. It is also a critical step in wireless sensor network for retrieval of sensitive data. The purposes of privacy-preserving query in both the areas of signal processing and sensor network are the same, but the similarity and difference of the adopted technologies are not fully explored. In this paper, we first review the recently developed methods of privacy-preserving query, and then describe in a comprehensive manner what we can learn from the mutual of both areas.

Keywords: encryption, privacy-preserving, search, security

Procedia PDF Downloads 252
18365 Nonstationarity Modeling of Economic and Financial Time Series

Authors: C. Slim

Abstract:

Traditional techniques for analyzing time series are based on the notion of stationarity of phenomena under study, but in reality most economic and financial series do not verify this hypothesis, which implies the implementation of specific tools for the detection of such behavior. In this paper, we study nonstationary non-seasonal time series tests in a non-exhaustive manner. We formalize the problem of nonstationary processes with numerical simulations and take stock of their statistical characteristics. The theoretical aspects of some of the most common unit root tests will be discussed. We detail the specification of the tests, showing the advantages and disadvantages of each. The empirical study focuses on the application of these tests to the exchange rate (USD/TND) and the Consumer Price Index (CPI) in Tunisia, in order to compare the Power of these tests with the characteristics of the series.

Keywords: stationarity, unit root tests, economic time series, ADF tests

Procedia PDF Downloads 416
18364 Hybrid Adaptive Modeling to Enhance Robustness of Real-Time Optimization

Authors: Hussain Syed Asad, Richard Kwok Kit Yuen, Gongsheng Huang

Abstract:

Real-time optimization has been considered an effective approach for improving energy efficient operation of heating, ventilation, and air-conditioning (HVAC) systems. In model-based real-time optimization, model mismatches cannot be avoided. When model mismatches are significant, the performance of the real-time optimization will be impaired and hence the expected energy saving will be reduced. In this paper, the model mismatches for chiller plant on real-time optimization are considered. In the real-time optimization of the chiller plant, simplified semi-physical or grey box model of chiller is always used, which should be identified using available operation data. To overcome the model mismatches associated with the chiller model, hybrid Genetic Algorithms (HGAs) method is used for online real-time training of the chiller model. HGAs combines Genetic Algorithms (GAs) method (for global search) and traditional optimization method (i.e. faster and more efficient for local search) to avoid conventional hit and trial process of GAs. The identification of model parameters is synthesized as an optimization problem; and the objective function is the Least Square Error between the output from the model and the actual output from the chiller plant. A case study is used to illustrate the implementation of the proposed method. It has been shown that the proposed approach is able to provide reliability in decision making, enhance the robustness of the real-time optimization strategy and improve on energy performance.

Keywords: energy performance, hybrid adaptive modeling, hybrid genetic algorithms, real-time optimization, heating, ventilation, and air-conditioning

Procedia PDF Downloads 411
18363 Multi-Label Approach to Facilitate Test Automation Based on Historical Data

Authors: Warda Khan, Remo Lachmann, Adarsh S. Garakahally

Abstract:

The increasing complexity of software and its applicability in a wide range of industries, e.g., automotive, call for enhanced quality assurance techniques. Test automation is one option to tackle the prevailing challenges by supporting test engineers with fast, parallel, and repetitive test executions. A high degree of test automation allows for a shift from mundane (manual) testing tasks to a more analytical assessment of the software under test. However, a high initial investment of test resources is required to establish test automation, which is, in most cases, a limitation to the time constraints provided for quality assurance of complex software systems. Hence, a computer-aided creation of automated test cases is crucial to increase the benefit of test automation. This paper proposes the application of machine learning for the generation of automated test cases. It is based on supervised learning to analyze test specifications and existing test implementations. The analysis facilitates the identification of patterns between test steps and their implementation with test automation components. For the test case generation, this approach exploits historical data of test automation projects. The identified patterns are the foundation to predict the implementation of unknown test case specifications. Based on this support, a test engineer solely has to review and parameterize the test automation components instead of writing them manually, resulting in a significant time reduction for establishing test automation. Compared to other generation approaches, this ML-based solution can handle different writing styles, authors, application domains, and even languages. Furthermore, test automation tools require expert knowledge by means of programming skills, whereas this approach only requires historical data to generate test cases. The proposed solution is evaluated using various multi-label evaluation criteria (EC) and two small-sized real-world systems. The most prominent EC is ‘Subset Accuracy’. The promising results show an accuracy of at least 86% for test cases, where a 1:1 relationship (Multi-Class) between test step specification and test automation component exists. For complex multi-label problems, i.e., one test step can be implemented by several components, the prediction accuracy is still at 60%. It is better than the current state-of-the-art results. It is expected the prediction quality to increase for larger systems with respective historical data. Consequently, this technique facilitates the time reduction for establishing test automation and is thereby independent of the application domain and project. As a work in progress, the next steps are to investigate incremental and active learning as additions to increase the usability of this approach, e.g., in case labelled historical data is scarce.

Keywords: machine learning, multi-class, multi-label, supervised learning, test automation

Procedia PDF Downloads 129
18362 Real-Time Implementation of Self-Tuning Fuzzy-PID Controller for First Order Plus Dead Time System Base on Microcontroller STM32

Authors: Maitree Thamma, Witchupong Wiboonjaroen, Thanat Suknuan, Karan Homchat

Abstract:

First order plus dead time (FOPDT) is a high dynamic system. Therefore, the controller must be intelligent. This paper presents the development and implementation of self-tuning Fuzzy-PID controller for controlling the FOPDT system. The water level process used represented FOPDT system and the mathematical model of the system was approximated by using System Identification toolbox in Matlab. The control programming and Fuzzy-PID algorithm used Matlab/Simulink and run on Microcontroller STM32.

Keywords: real-time control, self-tuning fuzzy-PID, FOPDT system, the water lever process

Procedia PDF Downloads 287
18361 Improving Part-Time Instructors’ Academic Outcomes with Gamification

Authors: Jared R. Chapman

Abstract:

This study introduces a type of motivational information system called an educational engagement information system (EEIS). An EEIS draws on principles of behavioral economics, motivation theory, and learning cognition theory to design information systems that help students want to improve their performance. This study compares academic outcomes for course sections taught by part- and full-time instructors both with and without an EEIS. Without an EEIS, students in the part-time instructor's course sections demonstrated significantly higher failure rates (a 143.8% increase) and dropout rates (a 110.4% increase) with significantly fewer students scoring a B- or higher (39.8% decrease) when compared to students in the course sections taught by a full-time instructor. It is concerning that students in the part-time instructor’s course without an EEIS had significantly lower academic outcomes, suggesting less understanding of the course content. This could impact retention and continuation in a major. With an EEIS, when comparing part- and full-time instructors, there was no significant difference in failure and dropout rates or in the number of students scoring a B- or higher in the course. In fact, with an EEIS, the failure and dropout rates were statistically identical for part- and full-time instructor courses. When using an EEIS (compared with not using an EEIS), the part-time instructor showed a 62.1% decrease in failures, a 61.4% decrease in dropouts, and a 41.7% increase in the number of students scoring a B- or higher in the course. We are unaware of other interventions that yield such large improvements in academic performance. This suggests that using an EEIS such as Delphinium may compensate for part-time instructors’ limitations of expertise, time, or rewards that can have a negative impact on students’ academic outcomes. The EEIS had only a minimal impact on failure rates (7.7% decrease) and dropout rates (18.8% decrease) for the full-time instructor. This suggests there is a ceiling effect for the improvements that an EEIS can make in student performance. This may be because experienced instructors are already doing the kinds of things that an EEIS does, such as motivating students, tracking grades, and providing feedback about progress. Additionally, full-time instructors have more time to dedicate to students outside of class than part-time instructors and more rewards for doing so. Using adjunct and other types of part-time instructors will likely remain a prevalent practice in higher education management courses. Given that using part-time instructors can have a negative impact on student graduation and persistence in a field of study, it is important to identify ways we can augment part-time instructors’ performance. We demonstrated that when part-time instructors use an EEIS, it can result in significantly lower students’ failure and dropout rates and an increase in the rate of students earning a B- or above; and bring their students’ performance to parity with the performance of students taught by a full-time instructor.

Keywords: gamification, engagement, motivation, academic outcomes

Procedia PDF Downloads 65
18360 Performance Comparison of Space-Time Block and Trellis Codes under Rayleigh Channels

Authors: Jing Qingfeng, Wu Jiajia

Abstract:

Due to the crowded orbits and shortage of frequency resources, utilizing of MIMO technology to improve spectrum efficiency and increase the capacity has become a necessary trend of broadband satellite communication. We analyze the main influenced factors and compare the BER performance of space-time block code (STBC) scheme and space-time trellis code (STTC) scheme. This paper emphatically studies the bit error rate (BER) performance of STTC and STBC under Rayleigh channel. The main emphasis is placed on the effects of the factors, such as terminal environment and elevation angles, on the BER performance of STBC and STTC schemes. Simulation results indicate that performance of STTC under Rayleigh channel is obviously improved with the increasing of transmitting and receiving antennas numbers, but the encoder state has little impact on the performance. Under Rayleigh channel, performance of Alamouti code is better than that of STTC.

Keywords: MIMO, space time block code (STBC), space time trellis code (STTC), Rayleigh channel

Procedia PDF Downloads 344
18359 Real-Time Data Stream Partitioning over a Sliding Window in Real-Time Spatial Big Data

Authors: Sana Hamdi, Emna Bouazizi, Sami Faiz

Abstract:

In recent years, real-time spatial applications, like location-aware services and traffic monitoring, have become more and more important. Such applications result dynamic environments where data as well as queries are continuously moving. As a result, there is a tremendous amount of real-time spatial data generated every day. The growth of the data volume seems to outspeed the advance of our computing infrastructure. For instance, in real-time spatial Big Data, users expect to receive the results of each query within a short time period without holding in account the load of the system. But with a huge amount of real-time spatial data generated, the system performance degrades rapidly especially in overload situations. To solve this problem, we propose the use of data partitioning as an optimization technique. Traditional horizontal and vertical partitioning can increase the performance of the system and simplify data management. But they remain insufficient for real-time spatial Big data; they can’t deal with real-time and stream queries efficiently. Thus, in this paper, we propose a novel data partitioning approach for real-time spatial Big data named VPA-RTSBD (Vertical Partitioning Approach for Real-Time Spatial Big data). This contribution is an implementation of the Matching algorithm for traditional vertical partitioning. We find, firstly, the optimal attribute sequence by the use of Matching algorithm. Then, we propose a new cost model used for database partitioning, for keeping the data amount of each partition more balanced limit and for providing a parallel execution guarantees for the most frequent queries. VPA-RTSBD aims to obtain a real-time partitioning scheme and deals with stream data. It improves the performance of query execution by maximizing the degree of parallel execution. This affects QoS (Quality Of Service) improvement in real-time spatial Big Data especially with a huge volume of stream data. The performance of our contribution is evaluated via simulation experiments. The results show that the proposed algorithm is both efficient and scalable, and that it outperforms comparable algorithms.

Keywords: real-time spatial big data, quality of service, vertical partitioning, horizontal partitioning, matching algorithm, hamming distance, stream query

Procedia PDF Downloads 154
18358 High School Gain Analytics From National Assessment Program – Literacy and Numeracy and Australian Tertiary Admission Rankin Linkage

Authors: Andrew Laming, John Hattie, Mark Wilson

Abstract:

Nine Queensland Independent high schools provided deidentified student-matched ATAR and NAPLAN data for all 1217 ATAR graduates since 2020 who also sat NAPLAN at the school. Graduating cohorts from the nine schools contained a mean 100 ATAR graduates with previous NAPLAN data from their school. Excluded were vocational students (mean=27) and any ATAR graduates without NAPLAN data (mean=20). Based on Index of Community Socio-Educational Access (ICSEA) prediction, all schools had larger that predicted proportions of their students graduating with ATARs. There were an additional 173 students not releasing their ATARs to their school (14%), requiring this data to be inferred by schools. Gain was established by first converting each student’s strongest NAPLAN domain to a statewide percentile, then subtracting this result from final ATAR. The resulting ‘percentile shift’ was corrected for plausible ATAR participation at each NAPLAN level. Strongest NAPLAN domain had the highest correlation with ATAR (R2=0.58). RESULTS School mean NAPLAN scores fitted ICSEA closely (R2=0.97). Schools achieved a mean cohort gain of two ATAR rankings, but only 66% of students gained. This ranged from 46% of top-NAPLAN decile students gaining, rising to 75% achieving gains outside the top decile. The 54% of top-decile students whose ATAR fell short of prediction lost a mean 4.0 percentiles (or 6.2 percentiles prior to correction for regression to the mean). 71% of students in smaller schools gained, compared to 63% in larger schools. NAPLAN variability in each of the 13 ICSEA1100 cohorts was 17%, with both intra-school and inter-school variation of these values extremely low (0.3% to 1.8%). Mean ATAR change between years in each school was just 1.1 ATAR ranks. This suggests consecutive school cohorts and ICSEA-similar schools share very similar distributions and outcomes over time. Quantile analysis of the NAPLAN/ATAR revealed heteroscedasticity, but splines offered little additional benefit over simple linear regression. The NAPLAN/ATAR R2 was 0.33. DISCUSSION Standardised data like NAPLAN and ATAR offer educators a simple no-cost progression metric to analyse performance in conjunction with their internal test results. Change is expressed in percentiles, or ATAR shift per student, which is layperson intuitive. Findings may also reduce ATAR/vocational stream mismatch, reveal proportions of cohorts meeting or falling short of expectation and demonstrate by how much. Finally, ‘crashed’ ATARs well below expectation are revealed, which schools can reasonably work to minimise. The percentile shift method is neither value-add nor a growth percentile. In the absence of exit NAPLAN testing, this metric is unable to discriminate academic gain from legitimate ATAR-maximizing strategies. But by controlling for ICSEA, ATAR proportion variation and student mobility, it uncovers progression to ATAR metrics which are not currently publicly available. However achieved, ATAR maximisation is a sought-after private good. So long as standardised nationwide data is available, this analysis offers useful analytics for educators and reasonable predictivity when counselling subsequent cohorts about their ATAR prospects.  

Keywords: NAPLAN, ATAR, analytics, measurement, gain, performance, data, percentile, value-added, high school, numeracy, reading comprehension, variability, regression to the mean

Procedia PDF Downloads 62
18357 Applying a Noise Reduction Method to Reveal Chaos in the River Flow Time Series

Authors: Mohammad H. Fattahi

Abstract:

Chaotic analysis has been performed on the river flow time series before and after applying the wavelet based de-noising techniques in order to investigate the noise content effects on chaotic nature of flow series. In this study, 38 years of monthly runoff data of three gauging stations were used. Gauging stations were located in Ghar-e-Aghaj river basin, Fars province, Iran. The noise level of time series was estimated with the aid of Gaussian kernel algorithm. This step was found to be crucial in preventing removal of the vital data such as memory, correlation and trend from the time series in addition to the noise during de-noising process.

Keywords: chaotic behavior, wavelet, noise reduction, river flow

Procedia PDF Downloads 466
18356 Performance Evaluation and Comparison between the Empirical Mode Decomposition, Wavelet Analysis, and Singular Spectrum Analysis Applied to the Time Series Analysis in Atmospheric Science

Authors: Olivier Delage, Hassan Bencherif, Alain Bourdier

Abstract:

Signal decomposition approaches represent an important step in time series analysis, providing useful knowledge and insight into the data and underlying dynamics characteristics while also facilitating tasks such as noise removal and feature extraction. As most of observational time series are nonlinear and nonstationary, resulting of several physical processes interaction at different time scales, experimental time series have fluctuations at all time scales and requires the development of specific signal decomposition techniques. Most commonly used techniques are data driven, enabling to obtain well-behaved signal components without making any prior-assumptions on input data. Among the most popular time series decomposition techniques, most cited in the literature, are the empirical mode decomposition and its variants, the empirical wavelet transform and singular spectrum analysis. With increasing popularity and utility of these methods in wide ranging applications, it is imperative to gain a good understanding and insight into the operation of these algorithms. In this work, we describe all of the techniques mentioned above as well as their ability to denoise signals, to capture trends, to identify components corresponding to the physical processes involved in the evolution of the observed system and deduce the dimensionality of the underlying dynamics. Results obtained with all of these methods on experimental total ozone columns and rainfall time series will be discussed and compared

Keywords: denoising, empirical mode decomposition, singular spectrum analysis, time series, underlying dynamics, wavelet analysis

Procedia PDF Downloads 105
18355 Optimization of Machining Parameters of Wire Electric Discharge Machining (WEDM) of Inconel 625 Super Alloy

Authors: Amitesh Goswami, Vishal Gulati, Annu Yadav

Abstract:

In this paper, WEDM has been used to investigate the machining characteristics of Inconel-625 alloy. The machining characteristics namely material removal rate (MRR) and surface roughness (SR) have been investigated along with surface microstructure analysis using SEM and EDS of the machined surface. Taguchi’s L27 Orthogonal array design has been used by considering six varying input parameters viz. Pulse-on time (Ton), Pulse-off time (Toff), Spark Gap Set Voltage (SV), Peak Current (IP), Wire Feed (WF) and Wire Tension (WT) for the responses of interest. It has been found out that Pulse-on time (Ton) and Spark Gap Set Voltage (SV) are the most significant parameters affecting material removal rate (MRR) and surface roughness (SR) are. Microstructure analysis of workpiece was also done using Scanning Electron Microscope (SEM). It was observed that, variations in pulse-on time and pulse-off time causes varying discharge energy and as a result of which deep craters / micro cracks and large/ small number of debris were formed. These results were helpful in studying the effects of pulse-on time and pulse-off time on MRR and SR. Energy Dispersive Spectrometry (EDS) was also done to check the compositional analysis of the material and it was observed that Copper and Zinc which were initially not present in the Inconel 625, later migrated on the material surface from the brass wire electrode during machining

Keywords: MRR, SEM, SR, taguchi, Wire Electric Discharge Machining

Procedia PDF Downloads 348
18354 A Review Study on the Importance and Correlation of Crisis Literacy and Media Communications for Vulnerable Marginalized People During Crisis

Authors: Maryam Jabeen

Abstract:

In recent times, there has been a notable surge in attention towards diverse literacy concepts such as media literacy, information literacy, and digital literacy. These concepts have garnered escalating interest, spurring the emergence of novel approaches, particularly in the aftermath of the Covid-19 crisis. However, amidst discussions of crises, the domain of crisis literacy remains largely uncharted within academic exploration. Crisis literacy, also referred to as disaster literacy, denotes an individual's aptitude to not only comprehend but also effectively apply information, enabling well-informed decision-making and adherence to instructions about disaster mitigation, preparedness, response, and recovery. This theoretical and descriptive study seeks to transcend foundational literacy concepts, underscoring the urgency for an in-depth exploration of crisis literacy and its interplay with the realm of media communication. Given the profound impact of the pandemic experience and the looming uncertainty of potential future crises, there arises a pressing need to elevate crisis literacy, or disaster literacy, towards heightened autonomy and active involvement within the spheres of critical disaster preparedness, recovery initiatives, and media communication domains. This research paper is part of my ongoing Ph.D. research study, which explores on a broader level the Encoding and decoding of media communications in relation to crisis literacy. The primary objective of this research paper is to expound upon a descriptive, theoretical research endeavor delving into this domain. The emphasis lies in highlighting the paramount significance of media communications in literacy of crisis, coupled with an accentuated focus on its role in providing information to marginalized populations amidst crises. In conclusion, this research bridges the gap in crisis literacy correlation to media communications exploration, advocating for a comprehensive understanding of its dynamics and its symbiotic relationship with media communications. It intends to foster a heightened sense of crisis literacy, particularly within marginalized communities, catalyzing proactive participation in disaster preparedness, recovery processes, and adept media interactions.

Keywords: covid-19, crisis literacy, crisis, marginalized, media and communications, pandemic, vulnerable people

Procedia PDF Downloads 56
18353 Investigation of the Fading Time Effects on Microstructure and Mechanical Properties in Vermicular Cast Iron

Authors: Mehmet Ekici

Abstract:

In this study, the fading time affecting the mechanical properties and microstructures of vermicular cast iron were studied. Pig iron and steel scrap weighing about 12 kg were charged into the high-frequency induction furnace crucible and completely melted for production of vermicular cast iron. The slag was skimmed using a common flux. After fading time was set at 1. 3 and 5 minutes. In this way, three vermicular cast iron was produced that same composition but different phase structures. The microstructure of specimens was investigated, and uni-axial tensile test and the Charpy impact test were performed, and their micro-hardness measurements were done in order to characterize the mechanical behaviours of vermicular cast iron.

Keywords: vermicular cast iron, fading time, hardness, tensile test and impact test

Procedia PDF Downloads 343
18352 On the Accuracy of Basic Modal Displacement Method Considering Various Earthquakes

Authors: Seyed Sadegh Naseralavi, Sadegh Balaghi, Ehsan Khojastehfar

Abstract:

Time history seismic analysis is supposed to be the most accurate method to predict the seismic demand of structures. On the other hand, the required computational time of this method toward achieving the result is its main deficiency. While being applied in optimization process, in which the structure must be analyzed thousands of time, reducing the required computational time of seismic analysis of structures makes the optimization algorithms more practical. Apparently, the invented approximate methods produce some amount of errors in comparison with exact time history analysis but the recently proposed method namely, Complete Quadratic Combination (CQC) and Sum Root of the Sum of Squares (SRSS) drastically reduces the computational time by combination of peak responses in each mode. In the present research, the Basic Modal Displacement (BMD) method is introduced and applied towards estimation of seismic demand of main structure. Seismic demand of sampled structure is estimated by calculation of modal displacement of basic structure (in which the modal displacement has been calculated). Shear steel sampled structures are selected as case studies. The error applying the introduced method is calculated by comparison of the estimated seismic demands with exact time history dynamic analysis. The efficiency of the proposed method is demonstrated by application of three types of earthquakes (in view of time of peak ground acceleration).

Keywords: time history dynamic analysis, basic modal displacement, earthquake-induced demands, shear steel structures

Procedia PDF Downloads 351
18351 Solution Approaches for Some Scheduling Problems with Learning Effect and Job Dependent Delivery Times

Authors: M. Duran Toksari, Berrin Ucarkus

Abstract:

In this paper, we propose two algorithms to optimally solve makespan and total completion time scheduling problems with learning effect and job dependent delivery times in a single machine environment. The delivery time is the extra time to eliminate adverse effect between the main processing and delivery to the customer. In this paper, we introduce the job dependent delivery times for some single machine scheduling problems with position dependent learning effect, which are makespan are total completion. The results with respect to two algorithms proposed for solving of the each problem are compared with LINGO solutions for 50-jobs, 100-jobs and 150-jobs problems. The proposed algorithms can find the same results in shorter time.

Keywords: delivery Times, learning effect, makespan, scheduling, total completion time

Procedia PDF Downloads 467
18350 IT-Aided Business Process Enabling Real-Time Analysis of Candidates for Clinical Trials

Authors: Matthieu-P. Schapranow

Abstract:

Recruitment of participants for clinical trials requires the screening of a big number of potential candidates, i.e. the testing for trial-specific inclusion and exclusion criteria, which is a time-consuming and complex task. Today, a significant amount of time is spent on identification of adequate trial participants as their selection may affect the overall study results. We introduce a unique patient eligibility metric, which allows systematic ranking and classification of candidates based on trial-specific filter criteria. Our web application enables real-time analysis of patient data and assessment of candidates using freely definable inclusion and exclusion criteria. As a result, the overall time required for identifying eligible candidates is tremendously reduced whilst additional degrees of freedom for evaluating the relevance of individual candidates are introduced by our contribution.

Keywords: in-memory technology, clinical trials, screening, eligibility metric, data analysis, clustering

Procedia PDF Downloads 490
18349 In vivo Antiplatelet Activity Test of Wet Extract of Mimusops elengi L.'s Leaves on DDY Strain Mice as an Effort to Treat Atherosclerosis

Authors: Dewi Tristantini, Jason Jonathan

Abstract:

Coronary Artery Disease (CAD) is one of the deathliest diseases which is caused by atherosclerosis. Atherosclerosis is a disease that plaque builds up inside the arteries. Plaque is made up of fat, cholesterol, calcium, platelet, and other substances found in blood. The current treatment of atherosclerosis is to provide antiplatelet therapy treatment, but such treatments often cause gastrointestinal irritation, muscle pain and hormonal imbalance. Mimusops elengi L.’s leaves can be utilized as a natural and cheap antiplatelet’s source because it contains flavonoids such as quertecin. Antiplatelet aggregation effect of Mimusops elengi L.’s leaves’ wet extract was measured by bleeding time on DDY strain mice with the test substances were given orally during the period of 8 days. The bleeding time was measured on first day and 9th day. Empirically, the dose which is used for humans is 8.5 g of leaves in 600 ml of water. This dose is equivalent to 2.1 g of leaves in 350 ml of water for mice. The extract was divided into 3 doses for mice: 0.05 ml/day; 0.1 ml/day; 0.2 ml/day. After getting the percentage of the increase in bleeding time, data were analyzed by analysis of variance test (Anova), followed by individual comparison within the groups by LSD test. The test substances above respectively increased bleeding time 21%, 62%, and 128%. As the conclusion, the 0.02 ml/day dose of Mimusops elengi L.’s leaves’ wet extract could increase bleeding time better than clopidogrel as positive controls with 110% increase in bleeding time.

Keywords: antiplatelets, atheroschlerosis, bleeding time, Mimusops elengi

Procedia PDF Downloads 259
18348 Eliciting and Confirming Data, Information, Knowledge and Wisdom in a Specialist Health Care Setting - The Wicked Method

Authors: Sinead Impey, Damon Berry, Selma Furtado, Miriam Galvin, Loretto Grogan, Orla Hardiman, Lucy Hederman, Mark Heverin, Vincent Wade, Linda Douris, Declan O'Sullivan, Gaye Stephens

Abstract:

Healthcare is a knowledge-rich environment. This knowledge, while valuable, is not always accessible outside the borders of individual clinics. This research aims to address part of this problem (at a study site) by constructing a maximal data set (knowledge artefact) for motor neurone disease (MND). This data set is proposed as an initial knowledge base for a concurrent project to develop an MND patient data platform. It represents the domain knowledge at the study site for the duration of the research (12 months). A knowledge elicitation method was also developed from the lessons learned during this process - the WICKED method. WICKED is an anagram of the words: eliciting and confirming data, information, knowledge, wisdom. But it is also a reference to the concept of wicked problems, which are complex and challenging, as is eliciting expert knowledge. The method was evaluated at a second site, and benefits and limitations were noted. Benefits include that the method provided a systematic way to manage data, information, knowledge and wisdom (DIKW) from various sources, including healthcare specialists and existing data sets. Limitations surrounded the time required and how the data set produced only represents DIKW known during the research period. Future work is underway to address these limitations.

Keywords: healthcare, knowledge acquisition, maximal data sets, action design science

Procedia PDF Downloads 337
18347 Investigating the Factors Affecting on One Time Passwords Technology Acceptance: A Case Study in Banking Environment

Authors: Sajad Shokohuyar, Mahsa Zomorrodi Anbaji, Saghar Pouyan Shad

Abstract:

According to fast technology growth, modern banking tries to decrease going to banks’ branches and increase customers’ consent. One of the problems which banks face is securing customer’s password. The banks’ solution is one time password creation system. In this research by adapting from acceptance of technology model theory, assesses factors that are effective on banking in Iran especially in using one time password machine by one of the private banks of Iran customers. The statistical population is all of this bank’s customers who use electronic banking service and one time password technology and the questionnaires were distributed among members of statistical population in 5 selected groups of north, south, center, east and west of Tehran. Findings show that confidential preservation, education, ease of utilization and advertising and informing has positive relations and distinct hardware and age has negative relations.

Keywords: security, electronic banking, one time password, information technology

Procedia PDF Downloads 446
18346 Asymmetric of the Segregation-Enhanced Brazil Nut Effect

Authors: Panupat Chaiworn, Soraya lama

Abstract:

We study the motion of particles in cylinders which are subjected to a sinusoidal vertical vibration. We measure the rising time of a large intruder from the bottom of the container to free surface of the bed particles and find that the rising time as a function of intruder density increases to a maximum and then decreases monotonically. The result is qualitatively accord to the previous findings in experiments using relative humidity of the bed particles and found speed convection of the bed particles containers it moving slowly, and the rising time of the intruder where a minimal instead of maximal rising time in the small density region was found. Our experimental results suggest that the topology of the container plays an important role in the Brazil nut effect.

Keywords: granular particles, Brazil nut effect, cylinder container, vertical vibration, convection

Procedia PDF Downloads 526