Search results for: time series
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 19593

Search results for: time series

18573 Time's Arrow and Entropy: Violations to the Second Law of Thermodynamics Disrupt Time Perception

Authors: Jason Clarke, Michaela Porubanova, Angela Mazzoli, Gulsah Kut

Abstract:

What accounts for our perception that time inexorably passes in one direction, from the past to the future, the so-called arrow of time, given that the laws of physics permit motion in one temporal direction to also happen in the reverse temporal direction? Modern physics says that the reason for time’s unidirectional physical arrow is the relationship between time and entropy, the degree of disorder in the universe, which is evolving from low entropy (high order; thermal disequilibrium) toward high entropy (high disorder; thermal equilibrium), the second law of thermodynamics. Accordingly, our perception of the direction of time, from past to future, is believed to emanate as a result of the natural evolution of entropy from low to high, with low entropy defining our notion of ‘before’ and high entropy defining our notion of ‘after’. Here we explored this proposed relationship between entropy and the perception of time’s arrow. We predicted that if the brain has some mechanism for detecting entropy, whose output feeds into processes involved in constructing our perception of the direction of time, presentation of violations to the expectation that low entropy defines ‘before’ and high entropy defines ‘after’ would alert this mechanism, leading to measurable behavioral effects, namely a disruption in duration perception. To test this hypothesis, participants were shown briefly-presented (1000 ms or 500 ms) computer-generated visual dynamic events: novel 3D shapes that were seen either to evolve from whole figures into parts (low to high entropy condition) or were seen in the reverse direction: parts that coalesced into whole figures (high to low entropy condition). On each trial, participants were instructed to reproduce the duration of their visual experience of the stimulus by pressing and releasing the space bar. To ensure that attention was being deployed to the stimuli, a secondary task was to report the direction of the visual event (forward or reverse motion). Participants completed 60 trials. As predicted, we found that duration reproduction was significantly longer for the high to low entropy condition compared to the low to high entropy condition (p=.03). This preliminary data suggests the presence of a neural mechanism that detects entropy, which is used by other processes to construct our perception of the direction of time or time’s arrow.

Keywords: time perception, entropy, temporal illusions, duration perception

Procedia PDF Downloads 163
18572 Gaze Patterns of Skilled and Unskilled Sight Readers Focusing on the Cognitive Processes Involved in Reading Key and Time Signatures

Authors: J. F. Viljoen, Catherine Foxcroft

Abstract:

Expert sight readers rely on their ability to recognize patterns in scores, their inner hearing and prediction skills in order to perform complex sight reading exercises. They also have the ability to observe deviations from expected patterns in musical scores. This increases the “Eye-hand span” (reading ahead of the point of playing) in order to process the elements in the score. The study aims to investigate the gaze patterns of expert and non-expert sight readers focusing on key and time signatures. 20 musicians were tasked with playing 12 sight reading examples composed for one hand and five examples composed for two hands to be performed on a piano keyboard. These examples were composed in different keys and time signatures and included accidentals and changes of time signature to test this theory. Results showed that the experts fixate more and for longer on key and time signatures as well as deviations in examples for two hands than the non-expert group. The inverse was true for the examples for one hand, where expert sight readers showed fewer and shorter fixations on key and time signatures as well as deviations. This seems to suggest that experts focus more on the key and time signatures as well as deviations in complex scores to facilitate sight reading. The examples written for one appeared to be too easy for the expert sight readers, compromising gaze patterns.

Keywords: cognition, eye tracking, musical notation, sight reading

Procedia PDF Downloads 136
18571 Volarization of Sugarcane Bagasse: The Effect of Alkali Concentration, Soaking Time and Temperature on Fibre Yield

Authors: Tamrat Tesfaye, Tilahun Seyoum, K. Shabaridharan

Abstract:

The objective of this paper was to determine the effect of NaOH concentration, soaking time, soaking temperature and their interaction on percentage yield of fibre extract using Response Surface Methodology (RSM). A Box-Behnken design was employed to optimize the extraction process of cellulosic fibre from sugar cane by-product bagasse using low alkaline extraction technique. The quadratic model with the optimal technological conditions resulted in a maximum fibre yield of 56.80% at 0.55N NaOH concentration, 4 h steeping time and 60ᵒC soaking temperature. Among the independent variables concentration was found to be the most significant (P < 0.005) variable and the interaction effect of concentration and soaking time leads to securing the optimized processes.

Keywords: sugarcane bagasse, low alkaline, Box-Behnken, fibre

Procedia PDF Downloads 242
18570 Filling the Gaps with Representation: Netflix’s Anne with an E as a Way to Reveal What the Text Hid

Authors: Arkadiusz Adam Gardaś

Abstract:

In his theory of gaps, Wolfgang Iser states that literary texts often lack direct messages. Instead of using straightforward descriptions, authors leave the gaps or blanks, i.e., the spaces within the text that come into existence only when readers fill them with their understanding and experiences. This paper’s aim is to present Iser’s literary theory in an intersectional way by comparing it to the idea of intersemiotic translation. To be more precise, the author uses the example of Netflix’s adaption of Lucy Maud Montgomery’s Anne of Green Gables as a form of rendering a book into a film in such a way that certain textual gaps are filled with film images. Intersemiotic translation is a rendition in which signs of one kind of media are translated into the signs of the other media. Film adaptions are the most common, but not the only, type of intersemiotic translation. In this case, the role of the translator is taken by a screenwriter. A screenwriter’s role can reach beyond the direct meaning presented by the author, and instead, it can delve into the source material (here – a novel) in a deeper way. When it happens, a screenwriter is able to spot the gaps in the text and fill them with images that can later be presented to the viewers. Anne with an E, the Netflix adaption of Montgomery’s novel, may be used as a highly meaningful example of such a rendition. It is due to the fact that the 2017 series was broadcasted more than a hundred years after the first edition of the novel was published. This means that what the author might not have been able to show in her text can now be presented in a more open way. The screenwriter decided to use this opportunity to represent certain groups in the film, i.e., racial and sexual minorities, and women. Nonetheless, the series does not alter the novel; in fact, it adds to it by filling the blanks with more direct images. In the paper, fragments of the first season of Anne with an E are analysed in comparison to its source, the novel by Montgomery. The main purpose of that is to show how intersemiotic translation connected with the Iser’s literary theory can enrich the understanding of works of art, culture, media, and literature.

Keywords: intersemiotic translation, film, literary gaps, representation

Procedia PDF Downloads 310
18569 Climate Changes Impact on Artificial Wetlands

Authors: Carla Idely Palencia-Aguilar

Abstract:

Artificial wetlands play an important role at Guasca Municipality in Colombia, not only because they are used for the agroindustry, but also because more than 45 species were found, some of which are endemic and migratory birds. Remote sensing was used to determine the changes in the area occupied by water of artificial wetlands by means of Aster and Modis images for different time periods. Evapotranspiration was also determined by three methods: Surface Energy Balance System-Su (SEBS) algorithm, Surface Energy Balance- Bastiaanssen (SEBAL) algorithm, and Potential Evapotranspiration- FAO. Empirical equations were also developed to determine the relationship between Normalized Difference Vegetation Index (NDVI) versus net radiation, ambient temperature and rain with an obtained R2 of 0.83. Groundwater level fluctuations on a daily basis were studied as well. Data from a piezometer placed next to the wetland were fitted with rain changes (with two weather stations located at the proximities of the wetlands) by means of multiple regression and time series analysis, the R2 from the calculated and measured values resulted was higher than 0.98. Information from nearby weather stations provided information for ordinary kriging as well as the results for the Digital Elevation Model (DEM) developed by using PCI software. Standard models (exponential, spherical, circular, gaussian, linear) to describe spatial variation were tested. Ordinary Cokriging between height and rain variables were also tested, to determine if the accuracy of the interpolation would increase. The results showed no significant differences giving the fact that the mean result of the spherical function for the rain samples after ordinary kriging was 58.06 and a standard deviation of 18.06. The cokriging using for the variable rain, a spherical function; for height variable, the power function and for the cross variable (rain and height), the spherical function had a mean of 57.58 and a standard deviation of 18.36. Threatens of eutrophication were also studied, given the unconsciousness of neighbours and government deficiency. Water quality was determined over the years; different parameters were studied to determine the chemical characteristics of water. In addition, 600 pesticides were studied by gas and liquid chromatography. Results showed that coliforms, nitrogen, phosphorous and prochloraz were the most significant contaminants.

Keywords: DEM, evapotranspiration, geostatistics, NDVI

Procedia PDF Downloads 115
18568 Votes - Commercialization in Nigeria: A Crime Against Sustainable Democracy

Authors: Oluwasaanmi Lawrence Adesuyi, Igbekoyi Kayode Emmanuel

Abstract:

This study examined vote - commercialization during elections among the voters in Nigeria, a series of elections in Ekiti State, Southwestern Nigeria. Democracy in Nigeria that came to replace the unwanted ruling and dictating mission of the military government has been facing a societal terror “crime of votes commercialization” that stands in jeopardy against its sustainability in Nigeria. Social exchange and action-bound theories were employed as the theoretical framework. Forty-Eight in-depth interviews, key informant interviews, and case studies were conducted with purposively selected respondents in the three senatorial districts that captured the sixteen local governments of the state. The results show that really commercialization of votes has become the order of the day in all series of electioneering among Ekiti people. Also, it was recorded that true democracy is no longer allowed to triumph as a result of vote buying that allows the highest bidder to be the winner. The result also shows that this attitude is not limited to only one political party or one candidate but involved all the political parties that participated in Election. It has become a frequent idea among the electorates during every festive period of election in Ekiti State. The tyrannical attitude has been given a nickname to suit the conditional situation of votes commercialization - (Diboki o se obe), which means vote and have a pot of soup, this implies that you will get money to take care of yourself and the family when you vote and collect money on the vote you cast, notwithstanding the money is being collected from all candidates that participated in the election, but the highest bidder has the day. The main challenge this has on democracy is that the contestants challenge the result of the election results based on the act of vote commercialization. Also, those that bought people’s votes with a huge amount of money relent on their democratic promises. The study showed that the crime of vote commercialization that threatens democracy must be addressed for sustainability.

Keywords: crime, democracy, jeopardy, military, sustainability, votes-commercialization

Procedia PDF Downloads 90
18567 Numerical Experiments for the Purpose of Studying Space-Time Evolution of Various Forms of Pulse Signals in the Collisional Cold Plasma

Authors: N. Kh. Gomidze, I. N. Jabnidze, K. A. Makharadze

Abstract:

The influence of inhomogeneities of plasma and statistical characteristics on the propagation of signal is very actual in wireless communication systems. While propagating in the media, the deformation and evaluation of the signal in time and space take place and on the receiver we get a deformed signal. The present article is dedicated to studying the space-time evolution of rectangular, sinusoidal, exponential and bi-exponential impulses via numerical experiment in the collisional, cold plasma. The presented method is not based on the Fourier-presentation of the signal. Analytically, we have received the general image depicting the space-time evolution of the radio impulse amplitude that gives an opportunity to analyze the concrete results in the case of primary impulse.

Keywords: collisional, cold plasma, rectangular pulse signal, impulse envelope

Procedia PDF Downloads 379
18566 On Flexible Preferences for Standard Taxis, Electric Taxis, and Peer-to-Peer Ridesharing

Authors: Ricardo Daziano

Abstract:

In the analysis and planning of the mobility ecosystem, preferences for ride-hailing over incumbent street-hailing services need better understanding. In this paper, a seminonparametric discrete choice model that allows for flexible preference heterogeneity is fitted with data from a discrete choice experiment among adult commuters in Montreal, Canada (N=760). Participants chose among Uber, Teo (a local electric ride-hailing service that was in operation when data was collected in 2018), and a standard taxi when presented with information about cost, time (on-trip, waiting, walking), powertrain of the car (gasoline/hybrid) for Uber and taxi, and whether the available electric Teo was a Tesla (which was one of the actual features of the Teo fleet). The fitted flexible model offers several behavioral insights. Waiting time for ride-hailing services is associated with a statistically significant but low marginal disutility. For other time components, including on-ride, and street-hailing waiting and walking the estimates of the value of time show an interesting pattern: whereas in a conditional logit on-ride time reductions are valued higher, in the flexible LML specification means of the value of time follow the expected pattern of waiting and walking creating a higher disutility. At the same time, the LML estimates show the presence of important, multimodal unobserved preference heterogeneity.

Keywords: discrete choice, electric taxis, ridehailing, semiparametrics

Procedia PDF Downloads 158
18565 Analyses and Optimization of Physical and Mechanical Properties of Direct Recycled Aluminium Alloy (AA6061) Wastes by ANOVA Approach

Authors: Mohammed H. Rady, Mohd Sukri Mustapa, S Shamsudin, M. A. Lajis, A. Wagiman

Abstract:

The present study is aimed at investigating microhardness and density of aluminium alloy chips when subjected to various settings of preheating temperature and preheating time. Three values of preheating temperature were taken as 450 °C, 500 °C, and 550 °C. On the other hand, three values of preheating time were chosen (1, 2, 3) hours. The influences of the process parameters (preheating temperature and time) were analyzed using Design of Experiments (DOE) approach whereby full factorial design with center point analysis was adopted. The total runs were 11 and they comprise of two factors of full factorial design with 3 center points. The responses were microhardness and density. The results showed that the density and microhardness increased with decreasing the preheating temperature. The results also found that the preheating temperature is more important to be controlled rather than the preheating time in microhardness analysis while both the preheating temperature and preheating time are important in density analysis. It can be concluded that setting temperature at 450 °C for 1 hour resulted in the optimum responses.

Keywords: AA6061, density, DOE, hot extrusion, microhardness

Procedia PDF Downloads 345
18564 Assessment of Five Photoplethysmographic Methods for Estimating Heart Rate Variability

Authors: Akshay B. Pawar, Rohit Y. Parasnis

Abstract:

Heart Rate Variability (HRV) is a widely used indicator of the regulation between the autonomic nervous system (ANS) and the cardiovascular system. Besides being non-invasive, it also has the potential to predict mortality in cases involving critical injuries. The gold standard method for determining HRV is based on the analysis of RR interval time series extracted from ECG signals. However, because it is much more convenient to obtain photoplethysmogramic (PPG) signals as compared to ECG signals (which require the attachment of several electrodes to the body), many researchers have used pulse cycle intervals instead of RR intervals to estimate HRV. They have also compared this method with the gold standard technique. Though most of their observations indicate a strong correlation between the two methods, recent studies show that in healthy subjects, except for a few parameters, the pulse-based method cannot be a surrogate for the standard RR interval- based method. Moreover, the former tends to overestimate short-term variability in heart rate. This calls for improvements in or alternatives to the pulse-cycle interval method. In this study, besides the systolic peak-peak interval method (PP method) that has been studied several times, four recent PPG-based techniques, namely the first derivative peak-peak interval method (P1D method), the second derivative peak-peak interval method (P2D method), the valley-valley interval method (VV method) and the tangent-intersection interval method (TI method) were compared with the gold standard technique. ECG and PPG signals were obtained from 10 young and healthy adults (consisting of both males and females) seated in the armchair position. In order to de-noise these signals and eliminate baseline drift, they were passed through certain digital filters. After filtering, the following HRV parameters were computed from PPG using each of the five methods and also from ECG using the gold standard method: time domain parameters (SDNN, pNN50 and RMSSD), frequency domain parameters (Very low-frequency power (VLF), Low-frequency power (LF), High-frequency power (HF) and Total power or “TP”). Besides, Poincaré plots were also plotted and their SD1/SD2 ratios determined. The resulting sets of parameters were compared with those yielded by the standard method using measures of statistical correlation (correlation coefficient) as well as statistical agreement (Bland-Altman plots). From the viewpoint of correlation, our results show that the best PPG-based methods for the determination of most parameters and Poincaré plots are the P2D method (shows more than 93% correlation with the standard method) and the PP method (mean correlation: 88%) whereas the TI, VV and P1D methods perform poorly (<70% correlation in most cases). However, our evaluation of statistical agreement using Bland-Altman plots shows that none of the five techniques agrees satisfactorily well with the gold standard method as far as time-domain parameters are concerned. In conclusion, excellent statistical correlation implies that certain PPG-based methods provide a good amount of information on the pattern of heart rate variation, whereas poor statistical agreement implies that PPG cannot completely replace ECG in the determination of HRV.

Keywords: photoplethysmography, heart rate variability, correlation coefficient, Bland-Altman plot

Procedia PDF Downloads 320
18563 Characterization on Molecular Weight of Polyamic Acids Using GPC Coupled with Multiple Detectors

Authors: Mei Hong, Wei Liu, Xuemin Dai, Yanxiong Pan, Xiangling Ji

Abstract:

Polyamic acid (PAA) is the precursor of polyimide (PI) prepared by a two-step method, its molecular weight and molecular weight distribution not only play an important role during the preparation and processing, but also influence the final performance of PI. However, precise characterization on molecular weight of PAA is still a challenge because of the existence of very complicated interactions in the solution system, including the electrostatic interaction, hydrogen bond interaction, dipole-dipole interaction, etc. Thus, it is necessary to establisha suitable strategy which can completely suppress these complex effects and get reasonable data on molecular weight. Herein, the gel permeation chromatography (GPC) coupled with differential refractive index (RI) and multi-angle laser light scattering (MALLS) detectors were applied to measure the molecular weight of (6FDA-DMB) PAA using different mobile phases, LiBr/DMF, LiBr/H3PO4/THF/DMF, LiBr/HAc/THF/DMF, and LiBr/HAc/DMF, respectively. It was found that combination of LiBr with HAc can shield the above-mentioned complex interactions and is more conducive to the separation of PAA than only addition of LiBr in DMF. LiBr/HAc/DMF was employed for the first time as a mild mobile phase to effectively separate PAA and determine its molecular weight. After a series of conditional experiments, 0.02M LiBr/0.2M HAc/DMF was fixed as an optimized mobile phase to measure the relative and absolute molecular weights of (6FDA-DMB) PAA prepared, and the obtained Mw from GPC-MALLS and GPC-RI were 35,300 g/mol and 125,000 g/mol, respectively. Particularly, such a mobile phase is also applicable to other PAA samples with different structures, and the final results on molecular weight are also reproducible.

Keywords: Polyamic acids, Polyelectrolyte effects, Gel permeation chromatography, Mobile phase, Molecular weight

Procedia PDF Downloads 51
18562 Driver Take-Over Time When Resuming Control from Highly Automated Driving in Truck Platooning Scenarios

Authors: Bo Zhang, Ellen S. Wilschut, Dehlia M. C. Willemsen, Marieke H. Martens

Abstract:

With the rapid development of intelligent transportation systems, automated platooning of trucks is drawing increasing interest for its beneficial effects on safety, energy consumption and traffic flow efficiency. Nevertheless, one major challenge lies in the safe transition of control from the automated system back to the human drivers, especially when they have been inattentive after a long period of highly automated driving. In this study, we investigated driver take-over time after a system initiated request to leave the platooning system Virtual Tow Bar in a non-critical scenario. 22 professional truck drivers participated in the truck driving simulator experiment, and each was instructed to drive under three experimental conditions before the presentation of the take-over request (TOR): driver ready (drivers were instructed to monitor the road constantly), driver not-ready (drivers were provided with a tablet) and eye-shut. The results showed significantly longer take-over time in both driver not-ready and eye-shut conditions compared with the driver ready condition. Further analysis revealed hand movement time as the main factor causing long response time in the driver not-ready condition, while in the eye-shut condition, gaze reaction time also influenced the total take-over time largely. In addition to comparing the means, large individual differences can be found especially in two driver, not attentive conditions. The importance of a personalized driver readiness predictor for a safe transition is concluded.

Keywords: driving simulation, highly automated driving, take-over time, transition of control, truck platooning

Procedia PDF Downloads 247
18561 Arthroscopic Assisted Fibertape Technique For Recurrent MPFL Reconstruction - Case Series Done In The UK Population

Authors: Naufal Ahmed, Michael Lwin

Abstract:

Background: MPFL reconstructions are ideally performed with au-tografts like gracilis semitendinosus tendon, which may be associated with donor site morbidity and complications. In this case series, we have tried to use fiber tape, which avoids the above complications and also keeps the graft virgin. This kind of synthetic graft has been used successfully in rotator cuffs and ACJ reconstructions with good results. Materials and methods: It was a retrospective data analysis of 45 patients who underwent this procedure from 2014-2020 under a single consultant in a DGH . These patiens have been followed up at 6 weeks, 6 months, 1 year, and 1 ½ years with clinical assessment and KOOS scores. We compared the results with the NJR and also with the Belgium report and was found to be satisfactory and comparable with them. Surgical technique : We used Arthrex fiber tape for the reconstruction of MPFL . Initially, two parallel holes drilled over sup aspect of the patella with help of an image intensifier, and then fiber wire passed through them from the medial to the lateral side and back to the medial side. The fiber wire was attached to the schottle point on the femoral side, giving a good extra articular internal brac-ing to the MPFL. All patients were scoped before the procedure, and the final tightening over the femoral side was done directly under vision to see the position of the patella. Results: We had 45 MPFL reconstructions along with 4 additional procedures 1 ACLR, 2 ACL REPAIR, 1 TTT advancement ( revision MPFL ). There were 14 males and 31 females, and their average age was 25 (13-55 ). We did not have any donor site morbidity, no infection, no fractures, no recurrent dislocations, no reoperations yet. Conclusion: Fiber tape is a feasible and appropriate option for MPFL reconstruction. We haven’t seen any re -operation in our 5 year follow up. This technique avoids the use of autograft, which can be used in the future if needed for revision surgeries. We don’t lose anything by following this simple novel technique.

Keywords: arthroscopy, fibertape, MPFL reconstruction, recurrent patella dislocation

Procedia PDF Downloads 135
18560 Developing Research Involving Different Species: Opportunities and Empirical Foundations

Authors: A. V. Varfolomeeva, N. S. Tkachenko, A. G. Tishchenko

Abstract:

The problem of violation of internal validity in studies of psychological structures is considered. The role of epistemological attitudes of researchers in the planning of research within the methodology of the system-evolutionary approach is assessed. Alternative programs of psychological research involving representatives of different biological species are presented. On the example of the results of two research series the variants of solving the problem are discussed.

Keywords: epistemological attitudes, experimental design, validity, psychological structure, learning

Procedia PDF Downloads 109
18559 Project Time and Quality Management during Construction

Authors: Nahed Al-Hajeri

Abstract:

Time and cost is an integral part of every construction plan and can affect each party’s contractual obligations. The performance of both time and cost are usually important to the client and contractor during the project. Almost all construction projects are experiencing time overrun. These time overruns always contributed as expensive to both client and contractor. Construction of any project inside the gathering centers involves complex management skills related to work force, materials, plant, machineries, new technologies etc. It also involves many agencies interdependent on each other like the vendors, structural and functional designers including various types of specialized engineers and it includes support of contractors and specialized contractors. This paper mainly highlights the types of construction delays due to which project suffer time and cost overrun. This paper also speaks about the delay causes and factors that contribute to the construction sequence delay for the oil and gas projects. Construction delay is supposed to be one of the repeated problems in the construction projects and it has an opposing effect on project success in terms of time, cost and quality. Some effective methods are identified to minimize delays in construction projects such as: 1. Site management and supervision, 2. Effective strategic planning, 3. Clear information and communication channel. Our research paper studies the types of delay with some real examples with statistic results and suggests solutions to overcome this problem.

Keywords: non-compensable delay, delays caused by force majeure, compensable delay, delays caused by the owner or the owner’s representative, non-excusable delay, delay caused by the contractor or the contractor’s representative, concurrent delay, delays resulting from two separate causes at the same time

Procedia PDF Downloads 238
18558 Coupled Space and Time Homogenization of Viscoelastic-Viscoplastic Composites

Authors: Sarra Haouala, Issam Doghri

Abstract:

In this work, a multiscale computational strategy is proposed for the analysis of structures, which are described at a refined level both in space and in time. The proposal is applied to two-phase viscoelastic-viscoplastic (VE-VP) reinforced thermoplastics subjected to large numbers of cycles. The main aim is to predict the effective long time response while reducing the computational cost considerably. The proposed computational framework is a combination of the mean-field space homogenization based on the generalized incrementally affine formulation for VE-VP composites, and the asymptotic time homogenization approach for coupled isotropic VE-VP homogeneous solids under large numbers of cycles. The time homogenization method is based on the definition of micro and macro-chronological time scales, and on asymptotic expansions of the unknown variables. First, the original anisotropic VE-VP initial-boundary value problem of the composite material is decomposed into coupled micro-chronological (fast time scale) and macro-chronological (slow time-scale) problems. The former is purely VE, and solved once for each macro time step, whereas the latter problem is nonlinear and solved iteratively using fully implicit time integration. Second, mean-field space homogenization is used for both micro and macro-chronological problems to determine the micro and macro-chronological effective behavior of the composite material. The response of the matrix material is VE-VP with J2 flow theory assuming small strains. The formulation exploits the return-mapping algorithm for the J2 model, with its two steps: viscoelastic predictor and plastic corrections. The proposal is implemented for an extended Mori-Tanaka scheme, and verified against finite element simulations of representative volume elements, for a number of polymer composite materials subjected to large numbers of cycles.

Keywords: asymptotic expansions, cyclic loadings, inclusion-reinforced thermoplastics, mean-field homogenization, time homogenization

Procedia PDF Downloads 364
18557 Dynamic Change of Floods Disaster Monitoring for River Central Bar by Remote Sensing Time-Series Images

Authors: Zuoji Huang, Jinyan Sun, Chunlin Wang, Haiming Qian, Nan Xu

Abstract:

The spatial extent and area of central river bars can always vary due to the impact of water level, sediment supply and human activities. In 2016, a catastrophic flood disaster caused by sustained and heavy rainfall happened in the middle and lower Yangtze River. The flood led to the most serious economic and social loss since 1954, and strongly affected the central river bar. It is essential to continuously monitor the dynamics change of central bars because it can avoid frequent field measurements in central bars before and after the flood disaster and is helpful for flood warning. This paper focused on the dynamic change of central bars of Phoenix bar and Changsha bar in the Yangtze River in 2016. In this study, GF-1 (GaoFen-1) WFV(wide field view) data was employed owing to its high temporal frequency and high spatial resolution. A simple NDWI (Normalized Difference Water Index) method was utilized for river central bar mapping. Human-checking was then performed to ensure the mapping quality. The relationship between the area of central bars and the measured water level was estimated using four mathematical models. Furthermore, a risk assessment index was proposed to map the spatial pattern of inundation risk of central bars. The results indicate a good ability of the GF-1 WFV imagery with a 16-m spatial resolution to characterize the seasonal variation of central river bars and to capture the impact of a flood disaster on the area of central bars. This paper observed a significant negative but nonlinear relationship between the water level and the area of central bars, and found that the cubic function fits best among four models (R² = 0.9839, P < 0.000001, RMSE = 0.4395). The maximum of the inundated area of central bars appeared during the rainy season on July 8, 2016, and the minimum occurred during the dry season on December 28, 2016, which are consistent with the water level measured by the hydrological station. The results derived from GF-1 data could provide a useful reference for decision-making of real-time disaster early warning and post-disaster reconstruction.

Keywords: central bars, dynamic change, water level, the Yangtze river

Procedia PDF Downloads 237
18556 Agile Project Management: A Real Application in a Multi-Project Research and Development Center

Authors: Aysegul Sarac

Abstract:

The aim of this study is to analyze the impacts of integrating agile development principles and practices, in particular to reduce project lead time in a multi-project environment. We analyze Arçelik Washing Machine R&D Center in which multiple projects are conducted by shared resources. In the first part of the study, we illustrate the current waterfall model system by using a value stream map. We define all activities starting from the first idea of the project to the customer and measure process time and lead time of projects. In the second part of the study we estimate potential improvements and select a set of these improvements to integrate agile principles. We aim to develop a future state map and analyze the impacts of integrating lean principles on project lead time. The main contribution of this study is that we analyze and integrate agile product development principles in a real multi-project system.

Keywords: agile project management, multi project system, project lead time, product development

Procedia PDF Downloads 297
18555 Traffic Prediction with Raw Data Utilization and Context Building

Authors: Zhou Yang, Heli Sun, Jianbin Huang, Jizhong Zhao, Shaojie Qiao

Abstract:

Traffic prediction is essential in a multitude of ways in modern urban life. The researchers of earlier work in this domain carry out the investigation chiefly with two major focuses: (1) the accurate forecast of future values in multiple time series and (2) knowledge extraction from spatial-temporal correlations. However, two key considerations for traffic prediction are often missed: the completeness of raw data and the full context of the prediction timestamp. Concentrating on the two drawbacks of earlier work, we devise an approach that can address these issues in a two-phase framework. First, we utilize the raw trajectories to a greater extent through building a VLA table and data compression. We obtain the intra-trajectory features with graph-based encoding and the intertrajectory ones with a grid-based model and the technique of back projection that restore their surrounding high-resolution spatial-temporal environment. To the best of our knowledge, we are the first to study direct feature extraction from raw trajectories for traffic prediction and attempt the use of raw data with the least degree of reduction. In the prediction phase, we provide a broader context for the prediction timestamp by taking into account the information that are around it in the training dataset. Extensive experiments on several well-known datasets have verified the effectiveness of our solution that combines the strength of raw trajectory data and prediction context. In terms of performance, our approach surpasses several state-of-the-art methods for traffic prediction.

Keywords: traffic prediction, raw data utilization, context building, data reduction

Procedia PDF Downloads 122
18554 Ant-Tracking Attribute: A Model for Understanding Production Response

Authors: Prince Suka Neekia Momta, Rita Iheoma Achonyeulo

Abstract:

Ant Tracking seismic attribute applied over 4-seconds seismic volume revealed structural features triggered by clay diapirism, growth fault development, rapid deltaic sedimentation and intense drilling. The attribute was extracted on vertical seismic sections and time slices. Mega tectonic structures such as growth faults and clay diapirs are visible on vertical sections with obscured minor lineaments or fractures. Fractures are distinctively visible on time slices yielding recognizable patterns corroborating established geologic models. This model seismic attribute enabled the understanding of fluid flow characteristics and production responses. Three structural patterns recognized in the field include: major growth faults, minor faults or lineaments and network of fractures. Three growth faults mapped on seismic section form major deformation bands delimiting the area into three blocks or depocenters. The growth faults trend E-W, dip down-to-south in the basin direction, and cut across the study area. The faults initiating from about 2000ms extended up to 500ms, and tend to progress parallel and opposite to the growth direction of an upsurging diapiric structure. The diapiric structures form the major deformational bands originating from great depths (below 2000ms) and rising to about 1200ms where series of sedimentary layers onlapped and pinchout stratigraphically against the diapir. Several other secondary faults or lineaments that form parallel streaks to one another also accompanied the growth faults. The fracture networks have no particular trend but form a network surrounding the well area. Faults identified in the study area have potentials for structural hydrocarbon traps whereas the presence of fractures created a fractured-reservoir condition that enhanced rapid fluid flow especially water. High aquifer flow potential aided by possible fracture permeability resulted in rapid decline in oil rate. Through the application of Ant Tracking attribute, it is possible to obtain detailed interpretation of structures that can have direct influence on oil and gas production.

Keywords: seismic, attributes, production, structural

Procedia PDF Downloads 57
18553 Cross-Cultural Competence Development through 'Learning by Reflection': A Case Study of Chinese International Students Learning through Taking Part-Time Jobs in the UK

Authors: Xin Zhao

Abstract:

The project aims to expand the notion of narrative learning and address the importance of learning by reflection in our learning and teaching context at a British university. Drawing on the key concepts such as development ZPD, transition and reflection-in and –on-action, this project analyses the learning experiences of a small sample of Chinese postgraduate students in a British University, who use part-time job experience to develop cross-cultural communication skills. The project adopts a mixed methods approach. Questionnaires and focus group interviews are used to examine the way in which students adapt (or not adapt) to the culture of learning in a British university and develop a renewed sense of self in transitions from one culture to the other. The project also looks at how the students appropriate opportunities for learning not just from classrooms but outside classrooms from everyday encounters. The project aims to address the implication of learning by reflection as development in transition. Time in and for learning, or duration, is taken for granted in theorising narrative learning. The project shall explore this very issue of time in relation to learning by reflection in considering time in/of/for learning as duration.

Keywords: cross-cultural competence, learning by refection, international student transition, part-time work experience

Procedia PDF Downloads 181
18552 Evaluation of Hand Grip Strength and EMG Signal on Visual Reaction

Authors: Sung-Wook Shin, Sung-Taek Chung

Abstract:

Hand grip strength has been utilized as an indicator to evaluate the motor ability of hands, responsible for performing multiple body functions. It is, however, difficult to evaluate other factors (other than hand muscular strength) utilizing the hand grip strength only. In this study, we analyzed the motor ability of hands using EMG and the hand grip strength, simultaneously in order to evaluate concentration, muscular strength reaction time, instantaneous muscular strength change, and agility in response to visual reaction. In results, the average time (and their standard deviations) of muscular strength reaction EMG signal and hand grip strength was found to be 209.6 ± 56.2 ms and 354.3 ± 54.6 ms, respectively. In addition, the onset time which represents acceleration time to reach 90% of maximum hand grip strength, was 382.9 ± 129.9 ms.

Keywords: hand grip strength, EMG, visual reaction, endurance

Procedia PDF Downloads 455
18551 Number of Parametrization of Discrete-Time Systems without Unit-Delay Element: Single-Input Single-Output Case

Authors: Kazuyoshi Mori

Abstract:

In this paper, we consider the parametrization of the discrete-time systems without the unit-delay element within the framework of the factorization approach. In the parametrization, we investigate the number of required parameters. We consider single-input single-output systems in this paper. By the investigation, we find, on the discrete-time systems without the unit-delay element, three cases that are (1) there exist plants which require only one parameter and (2) two parameters, and (3) the number of parameters is at most three.

Keywords: factorization approach, discrete-time system, parameterization of stabilizing controllers, system without unit-delay

Procedia PDF Downloads 235
18550 Time Travel Testing: A Mechanism for Improving Renewal Experience

Authors: Aritra Majumdar

Abstract:

While organizations strive to expand their new customer base, retaining existing relationships is a key aspect of improving overall profitability and also showcasing how successful an organization is in holding on to its customers. It is an experimentally proven fact that the lion’s share of profit always comes from existing customers. Hence seamless management of renewal journeys across different channels goes a long way in improving trust in the brand. From a quality assurance standpoint, time travel testing provides an approach to both business and technology teams to enhance the customer experience when they look to extend their partnership with the organization for a defined phase of time. This whitepaper will focus on key pillars of time travel testing: time travel planning, time travel data preparation, and enterprise automation. Along with that, it will call out some of the best practices and common accelerator implementation ideas which are generic across verticals like healthcare, insurance, etc. In this abstract document, a high-level snapshot of these pillars will be provided. Time Travel Planning: The first step of setting up a time travel testing roadmap is appropriate planning. Planning will include identifying the impacted systems that need to be time traveled backward or forward depending on the business requirement, aligning time travel with other releases, frequency of time travel testing, preparedness for handling renewal issues in production after time travel testing is done and most importantly planning for test automation testing during time travel testing. Time Travel Data Preparation: One of the most complex areas in time travel testing is test data coverage. Aligning test data to cover required customer segments and narrowing it down to multiple offer sequencing based on defined parameters are keys for successful time travel testing. Another aspect is the availability of sufficient data for similar combinations to support activities like defect retesting, regression testing, post-production testing (if required), etc. This section will talk about the necessary steps for suitable data coverage and sufficient data availability from a time travel testing perspective. Enterprise Automation: Time travel testing is never restricted to a single application. The workflow needs to be validated in the downstream applications to ensure consistency across the board. Along with that, the correctness of offers across different digital channels needs to be checked in order to ensure a smooth customer experience. This section will talk about the focus areas of enterprise automation and how automation testing can be leveraged to improve the overall quality without compromising on the project schedule. Along with the above-mentioned items, the white paper will elaborate on the best practices that need to be followed during time travel testing and some ideas pertaining to accelerator implementation. To sum it up, this paper will be written based on the real-time experience author had on time travel testing. While actual customer names and program-related details will not be disclosed, the paper will highlight the key learnings which will help other teams to implement time travel testing successfully.

Keywords: time travel planning, time travel data preparation, enterprise automation, best practices, accelerator implementation ideas

Procedia PDF Downloads 153
18549 FLIME - Fast Low Light Image Enhancement for Real-Time Video

Authors: Vinay P., Srinivas K. S.

Abstract:

Low Light Image Enhancement is of utmost impor- tance in computer vision based tasks. Applications include vision systems for autonomous driving, night vision devices for defence systems, low light object detection tasks. Many of the existing deep learning methods are resource intensive during the inference step and take considerable time for processing. The algorithm should take considerably less than 41 milliseconds in order to process a real-time video feed with 24 frames per second and should be even less for a video with 30 or 60 frames per second. The paper presents a fast and efficient solution which has two main advantages, it has the potential to be used for a real-time video feed, and it can be used in low compute environments because of the lightweight nature. The proposed solution is a pipeline of three steps, the first one is the use of a simple function to map input RGB values to output RGB values, the second is to balance the colors and the final step is to adjust the contrast of the image. Hence a custom dataset is carefully prepared using images taken in low and bright lighting conditions. The preparation of the dataset, the proposed model, the processing time are discussed in detail and the quality of the enhanced images using different methods is shown.

Keywords: low light image enhancement, real-time video, computer vision, machine learning

Procedia PDF Downloads 194
18548 Development of Quasi Real-Time Comprehensive System for Earthquake Disaster

Authors: Zhi Liu, Hui Jiang, Jin Li, Kunhao Chen, Langfang Zhang

Abstract:

Fast acquisition of the seismic information and accurate assessment of the earthquake disaster is the key problem for emergency rescue after a destructive earthquake. In order to meet the requirements of the earthquake emergency response and rescue for the cities and counties, a quasi real-time comprehensive evaluation system for earthquake disaster is developed. Based on monitoring data of Micro-Electro-Mechanical Systems (MEMS) strong motion network, structure database of a county area and the real-time disaster information by the mobile terminal after an earthquake, fragility analysis method and dynamic correction algorithm are synthetically obtained in the developed system. Real-time evaluation of the seismic disaster in the county region is finally realized to provide scientific basis for seismic emergency command, rescue and assistant decision.

Keywords: quasi real-time, earthquake disaster data collection, MEMS accelerometer, dynamic correction, comprehensive evaluation

Procedia PDF Downloads 208
18547 Toward Green Islamic Finance: A Case Study from an Emirati Islamic Bank

Authors: Nada Hamed, Mariam Aldhaheri, Sonia Abdennadher

Abstract:

Islamic Finance is not a new term that emerging in the global market, but it is still under scope by many countries. Its characteristics and regulation are not widely clear and implemented. In 2015, The United Nation announced a plan about potential benefits of using Islamic Finance as a sustainable development approach. Enhancing its application in financial markets could protect from unexpected crisis that might be created from the traditional tools of finance. This paper focuses on this area to test if Islamic finance could be used for maintaining sustainable development and if the term of 'Green Islamic Finance' could be implemented to minimize the deficiencies and 'pollution’ generated from traditional techniques and tools of finance. This paper intends to measure the impact on financial performance and sustainability when financial institutions use Islamic finance or better practice it. The objective of this explanatory research is to measure the performance of Islamic Finance with using a case study of an Islamic bank. The paper would analyze and compare the behavior of financial institutions that used traditional financing tools and converted to Islamic banking system. The methodology used is based on a case study of an Islamic bank in Dubai with comparing its performance before implementing Islamic Finance and after. The selected case study represents the first national bank in Emirates Arab Unis who adopt the Islamic finance approach. Based on a time series analysis, a quantitative analysis would be also used through looking at various set of ratios that are routinely used to measure bank performance.

Keywords: Islamic finance, financial stability, green finance, Islamic finance practices, financial ratios

Procedia PDF Downloads 218
18546 Revisiting Classic Triad of Japanese Spotted Fever: A Case Series of Forty-Three Patients

Authors: Y. Kunitani, Y. Nakashima, S. Yamauchi, Y. Ishigami, K. Naito, K. Numata, M. Mizobe, Y. Homma, J. Takahashi, T. Inoue, T. Shiga, H. Funakoshi

Abstract:

Background: Japanese Spotted Fever (JSF) is one of the Rickettsial infections, caused by Rickettsia japonica, which is transmitted by ticks. JSF is seen in limited area, such as Japan and South Korea. Its clinical triad is rash, eschar and fever. It often shows leukocytopenia, thrombopenia, elevated transaminase and high C-reactive protein (CRP). Sometimes it can be life-threatening due to disseminated intravascular coagulation or multiple organ failure. Study Aim: The aim of this study is to describe the features of JSF, as this unique infection is rapidly growing in Japan. Methods: This is a case series of JSF from 2009 to 2016, in Mie Prefectural Hospital in Japan. We collected JSF cases, which were diagnosed by polymerase chain reaction (PCR) of the skin or blood serum, or the elevation of the antibody titer of paired blood samples. Results: There were 43 JSF patients (19 male, 24 female) with a median age of 71 years [IQR:65-80]. The median body temperature was 38.1°C[IQR: 37.5-39.0]. 95% had a rash, 67% had eschar and 50% had fever. The median WBC count was 6,700 [IQR: 5,750-8,200] and leukocytopenia was observed in only 7%. The median platelet count was 14x104 [IQR10x104-17x104], thrombopenia was observed in 65%. The median aspartate transaminase (AST) was 53 IU/L [IQR: 41-93]; the median alanine aminotransferase (ALT) was 34 IU/L [IQR: 24-54]; the median CRP was 10.4 mg/dL [IQR:7.2-13.9]; the median lactate dehydrogenase (LDH) was 352IU/L [IQR:282-451]. CRP and LDH were elevated in almost all of the patients. Median length of stay in hospital was 8 days [IQR: 6-11]. All patients were treated with tetracycline and quinolone on the day of the presentation. There was no fatality from JSF. Conclusion: The patients with JSF classically presents with eschar, rash and fever. However, in this study, the half of the patients were afebrile. Although JSF is not a common infectious disease worldwide, if the patient had previously visited Japan or South Korea and presented with rash and eschar with or without fever, we should consider JSF as a potential diagnosis.

Keywords: infectious disease, Japanese spotted fever, Rickettsial disease, Rickettsia japonica

Procedia PDF Downloads 223
18545 New Results on Exponential Stability of Hybrid Systems

Authors: Grienggrai Rajchakit

Abstract:

This paper is concerned with the exponential stability of switched linear systems with interval time-varying delays. The time delay is any continuous function belonging to a given interval, in which the lower bound of delay is not restricted to zero. By constructing a suitable augmented Lyapunov-Krasovskii functional combined with Leibniz-Newton's formula, a switching rule for the exponential stability of switched linear systems with interval time-varying delays and new delay-dependent sufficient conditions for the exponential stability of the systems are first established in terms of LMIs. Finally, some examples are exploited to illustrate the effectiveness of the proposed schemes.

Keywords: exponential stability, hybrid systems, time-varying delays, lyapunov-krasovskii functional, leibniz-newton's formula

Procedia PDF Downloads 540
18544 An Approaching Index to Evaluate a forward Collision Probability

Authors: Yuan-Lin Chen

Abstract:

This paper presents an approaching forward collision probability index (AFCPI) for alerting and assisting driver in keeping safety distance to avoid the forward collision accident in highway driving. The time to collision (TTC) and time headway (TH) are used to evaluate the TTC forward collision probability index (TFCPI) and the TH forward collision probability index (HFCPI), respectively. The Mamdani fuzzy inference algorithm is presented combining TFCPI and HFCPI to calculate the approaching collision probability index of the vehicle. The AFCPI is easier to understand for the driver who did not even have any professional knowledge in vehicle professional field. At the same time, the driver’s behavior is taken into account for suiting each driver. For the approaching index, the value 0 is indicating the 0% probability of forward collision, and the values 0.5 and 1 are indicating the 50% and 100% probabilities of forward collision, respectively. The AFCPI is useful and easy-to-understand for alerting driver to avoid the forward collision accidents when driving in highway.

Keywords: approaching index, forward collision probability, time to collision, time headway

Procedia PDF Downloads 287