Search results for: lead time reduction
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 24317

Search results for: lead time reduction

20657 Time of Death Determination in Medicolegal Death Investigations

Authors: Michelle Rippy

Abstract:

Medicolegal death investigation historically is a field that does not receive much research attention or advancement, as all of the subjects are deceased. Public health threats, drug epidemics and contagious diseases are typically recognized in decedents first, with thorough and accurate death investigations able to assist in epidemiology research and prevention programs. One vital component of medicolegal death investigation is determining the decedent’s time of death. An accurate time of death can assist in corroborating alibies, determining sequence of death in multiple casualty circumstances and provide vital facts in civil situations. Popular television portrays an unrealistic forensic ability to provide the exact time of death to the minute for someone found deceased with no witnesses present. The actuality of unattended decedent time of death determination can generally only be narrowed to a 4-6 hour window. In the mid- to late-20th century, liver temperatures were an invasive action taken by death investigators to determine the decedent’s core temperature. The core temperature was programmed into an equation to determine an approximate time of death. Due to many inconsistencies with the placement of the thermometer and other variables, the accuracy of the liver temperatures was dispelled and this once common place action lost scientific support. Currently, medicolegal death investigators utilize three major after death or post-mortem changes at a death scene. Many factors are considered in the subjective determination as to the time of death, including the cooling of the decedent, stiffness of the muscles, release of blood internally, clothing, ambient temperature, disease and recent exercise. Current research is utilizing non-invasive hospital grade tympanic thermometers to measure the temperature in the each of the decedent’s ears. This tool can be used at the scene and in conjunction with scene indicators may provide a more accurate time of death. The research is significant and important to investigations and can provide an area of accuracy to a historically inaccurate area, considerably improving criminal and civil death investigations. The goal of the research is to provide a scientific basis to unwitnessed deaths, instead of the art that the determination currently is. The research is currently in progress with expected termination in December 2018. There are currently 15 completed case studies with vital information including the ambient temperature, decedent height/weight/sex/age, layers of clothing, found position, if medical intervention occurred and if the death was witnessed. This data will be analyzed with the multiple variables studied and available for presentation in January 2019.

Keywords: algor mortis, forensic pathology, investigations, medicolegal, time of death, tympanic

Procedia PDF Downloads 108
20656 Visual Thinking Routines: A Mixed Methods Approach Applied to Student Teachers at the American University in Dubai

Authors: Alain Gholam

Abstract:

Visual thinking routines are principles based on several theories, approaches, and strategies. Such routines promote thinking skills, call for collaboration and sharing of ideas, and above all, make thinking and learning visible. Visual thinking routines were implemented in the teaching methodology graduate course at the American University in Dubai. The study used mixed methods. It was guided by the following two research questions: 1). To what extent do visual thinking inspire learning in the classroom, and make time for students’ questions, contributions, and thinking? 2). How do visual thinking routines inspire learning in the classroom and make time for students’ questions, contributions, and thinking? Eight student teachers enrolled in the teaching methodology course at the American University in Dubai (Spring 2017) participated in the following study. First, they completed a survey that measured to what degree they believed visual thinking routines inspired learning in the classroom and made time for students’ questions, contributions, and thinking. In order to build on the results from the quantitative phase, the student teachers were next involved in a qualitative data collection phase, where they had to answer the question: How do visual thinking routines inspire learning in the classroom and make time for students’ questions, contributions, and thinking? Results revealed that the implementation of visual thinking routines in the classroom strongly inspire learning in the classroom and make time for students’ questions, contributions, and thinking. In addition, student teachers explained how visual thinking routines allow for organization, variety, thinking, and documentation. As with all original, new, and unique resources, visual thinking routines are not free of challenges. To make the most of this useful and valued resource, educators, need to comprehend, model and spread an awareness of the effective ways of using such routines in the classroom. It is crucial that such routines become part of the curriculum to allow for and document students’ questions, contributions, and thinking.

Keywords: classroom display, student engagement, thinking classroom, visual thinking routines

Procedia PDF Downloads 221
20655 Data Model to Predict Customize Skin Care Product Using Biosensor

Authors: Ashi Gautam, Isha Shukla, Akhil Seghal

Abstract:

Biosensors are analytical devices that use a biological sensing element to detect and measure a specific chemical substance or biomolecule in a sample. These devices are widely used in various fields, including medical diagnostics, environmental monitoring, and food analysis, due to their high specificity, sensitivity, and selectivity. In this research paper, a machine learning model is proposed for predicting the suitability of skin care products based on biosensor readings. The proposed model takes in features extracted from biosensor readings, such as biomarker concentration, skin hydration level, inflammation presence, sensitivity, and free radicals, and outputs the most appropriate skin care product for an individual. This model is trained on a dataset of biosensor readings and corresponding skin care product information. The model's performance is evaluated using several metrics, including accuracy, precision, recall, and F1 score. The aim of this research is to develop a personalised skin care product recommendation system using biosensor data. By leveraging the power of machine learning, the proposed model can accurately predict the most suitable skin care product for an individual based on their biosensor readings. This is particularly useful in the skin care industry, where personalised recommendations can lead to better outcomes for consumers. The developed model is based on supervised learning, which means that it is trained on a labeled dataset of biosensor readings and corresponding skin care product information. The model uses these labeled data to learn patterns and relationships between the biosensor readings and skin care products. Once trained, the model can predict the most suitable skin care product for an individual based on their biosensor readings. The results of this study show that the proposed machine learning model can accurately predict the most appropriate skin care product for an individual based on their biosensor readings. The evaluation metrics used in this study demonstrate the effectiveness of the model in predicting skin care products. This model has significant potential for practical use in the skin care industry for personalised skin care product recommendations. The proposed machine learning model for predicting the suitability of skin care products based on biosensor readings is a promising development in the skin care industry. The model's ability to accurately predict the most appropriate skin care product for an individual based on their biosensor readings can lead to better outcomes for consumers. Further research can be done to improve the model's accuracy and effectiveness.

Keywords: biosensors, data model, machine learning, skin care

Procedia PDF Downloads 83
20654 Impact of 6-Week Brain Endurance Training on Cognitive and Cycling Performance in Highly Trained Individuals

Authors: W. Staiano, S. Marcora

Abstract:

Introduction: It has been proposed that acute negative effect of mental fatigue (MF) could potentially become a training stimulus for the brain (Brain endurance training (BET)) to adapt and improve its ability to attenuate MF states during sport competitions. Purpose: The aim of this study was to test the efficacy of 6 weeks of BET on cognitive and cycling tests in a group of well-trained subjects. We hypothesised that combination of BET and standard physical training (SPT) would increase cognitive capacity and cycling performance by reducing rating of perceived exertion (RPE) and increase resilience to fatigue more than SPT alone. Methods: In a randomized controlled trial design, 26 well trained participants, after a familiarization session, cycled to exhaustion (TTE) at 80% peak power output (PPO) and, after 90 min rest, at 65% PPO, before and after random allocation to a 6 week BET or active placebo control. Cognitive performance was measured using 30 min of STROOP coloured task performed before cycling performance. During the training, BET group performed a series of cognitive tasks for a total of 30 sessions (5 sessions per week) with duration increasing from 30 to 60 min per session. Placebo engaged in a breathing relaxation training. Both groups were monitored for physical training and were naïve to the purpose of the study. Physiological and perceptual parameters of heart rate, lactate (LA) and RPE were recorded during cycling performances, while subjective workload (NASA TLX scale) was measured during the training. Results: Group (BET vs. Placebo) x Test (Pre-test vs. Post-test) mixed model ANOVA’s revealed significant interaction for performance at 80% PPO (p = .038) or 65% PPO (p = .011). In both tests, groups improved their TTE performance; however, BET group improved significantly more compared to placebo. No significant differences were found for heart rate during the TTE cycling tests. LA did not change significantly at rest in both groups. However, at completion of 65% TTE, it was significantly higher (p = 0.043) in the placebo condition compared to BET. RPE measured at ISO-time in BET was significantly lower (80% PPO, p = 0.041; 65% PPO p= 0.021) compared to placebo. Cognitive results in the STROOP task showed that reaction time in both groups decreased at post-test. However, BET decreased significantly (p = 0.01) more compared to placebo despite no differences accuracy. During training sessions, participants in the BET showed, through NASA TLX questionnaires, constantly significantly higher (p < 0.01) mental demand rates compared to placebo. No significant differences were found for physical demand. Conclusion: The results of this study provide evidences that combining BET and SPT seems to be more effective than SPT alone in increasing cognitive and cycling performance in well trained endurance participants. The cognitive overload produced during the 6-week training of BET can induce a reduction in perception of effort at a specific power, and thus improving cycling performance. Moreover, it provides evidence that including neurocognitive interventions will benefit athletes by increasing their mental resilience, without affecting their physical training load and routine.

Keywords: cognitive training, perception of effort, endurance performance, neuro-performance

Procedia PDF Downloads 112
20653 Water Footprint for the Palm Oil Industry in Malaysia

Authors: Vijaya Subramaniam, Loh Soh Kheang, Astimar Abdul Aziz

Abstract:

Water footprint (WFP) has gained importance due to the increase in water scarcity in the world. This study analyses the WFP for an agriculture sector, i.e., the oil palm supply chain, which produces oil palm fresh fruit bunch (FFB), crude palm oil, palm kernel, and crude palm kernel oil. The water accounting and vulnerability evaluation (WAVE) method was used. This method analyses the water depletion index (WDI) based on the local blue water scarcity. The main contribution towards the WFP at the plantation was the production of FFB from the crop itself at 0.23m³/tonne FFB. At the mill, the burden shifts to the water added during the process, which consists of the boiler and process water, which accounted for 6.91m³/tonne crude palm oil. There was a 33% reduction in the WFP when there was no dilution or water addition after the screw press at the mill. When allocation was performed, the WFP reduced by 42% as the burden was shared with the palm kernel and palm kernel shell. At the kernel crushing plant (KCP), the main contributor towards the WFP 4.96 m³/tonne crude palm kernel oil which came from the palm kernel which carried the burden from upstream followed by electricity, 0.33 m³/tonne crude palm kernel oil used for the process and 0.08 m³/tonne crude palm kernel oil for transportation of the palm kernel. A comparison was carried out for mills with biogas capture versus no biogas capture, and the WFP had no difference for both scenarios. The comparison when the KCPs operate in the proximity of mills as compared to those operating in the proximity of ports only gave a reduction of 6% for the WFP. Both these scenarios showed no difference and insignificant difference, which differed from previous life cycle assessment studies on the carbon footprint, which showed significant differences. This shows that findings change when only certain impact categories are focused on. It can be concluded that the impact from the water used by the oil palm tree is low due to the practice of no irrigation at the plantations and the high availability of water from rainfall in Malaysia. This reiterates the importance of planting oil palm trees in regions with high rainfall all year long, like the tropics. The milling stage had the most significant impact on the WFP. Mills should avoid dilution to reduce this impact.

Keywords: life cycle assessment, water footprint, crude palm oil, crude palm kernel oil, WAVE method

Procedia PDF Downloads 157
20652 Application of Low-order Modeling Techniques and Neural-Network Based Models for System Identification

Authors: Venkatesh Pulletikurthi, Karthik B. Ariyur, Luciano Castillo

Abstract:

The system identification from the turbulence wakes will lead to the tactical advantage to prepare and also, to predict the trajectory of the opponents’ movements. A low-order modeling technique, POD, is used to predict the object based on the wake pattern and compared with pre-trained image recognition neural network (NN) to classify the wake patterns into objects. It is demonstrated that low-order modeling, POD, is able to predict the objects better compared to pretrained NN by ~30%.

Keywords: the bluff body wakes, low-order modeling, neural network, system identification

Procedia PDF Downloads 171
20651 Deep Neck Infection Associated with Peritoneal Sepsis: A Rare Death Case

Authors: Sait Ozsoy, Asude Gokmen, Mehtap Yondem, Hanife A. Alkan, Gulnaz T. Javan

Abstract:

Deep neck infection often develops due to upper respiratory tract and odontogenic infections. Gastrointestinal System perforation can occur for many reasons and is in need of the early diagnosis and prompt surgical treatment. In both cases late or incorrect diagnosis may lead to increase morbidity and high mortality. A patient with a diagnosis of deep neck abscess died while under treatment due to sepsis and multiple organ failure. Autopsy finding showed duodenal ulcer and this is reported in the literature.

Keywords: peptic ulcer perforation, peritonitis, retropharyngeal abscess, sepsis

Procedia PDF Downloads 487
20650 Fill Rate Window as a Criterion for Spares Allocation

Authors: Michael Dreyfuss, Yahel Giat

Abstract:

Limited battery range and long recharging times are the greatest obstacles to the successful adoption of electric cars. One of the suggestions to overcome these problems is that carmakers retain ownership of batteries and provide battery swapping service so that customers exchange their depleted batteries for recharged batteries. Motivated by this example, we consider the problem of optimal spares allocation in an exchangeable-item, multi-location repair system. We generalize the standard service measures of fill rate and average waiting time to reflect the fact that customers penalize the service provider only if they have to wait more than a ‘tolerable’ time window. These measures are denoted as the window fill rate and the truncated waiting time, respectively. We find that the truncated waiting time is convex and therefore a greedy algorithm solves the spares allocation problem efficiently. We show that the window fill rate is generally S-shaped and describe an efficient algorithm to find a near-optimal solution and detail a priori and a posteriori upper bounds to the distance from optimum. The theory is complemented with a large scale numerical example demonstrating the spare battery allocation in battery swapping stations.

Keywords: convex-concave optimization, exchangeable item, M/G/infinity, multiple location, repair system, spares allocation, window fill rate

Procedia PDF Downloads 484
20649 Miniaturization of I-Slot Antenna with Improved Efficiency and Gain

Authors: Mondher Labidi, Fethi Choubani

Abstract:

In this paper, novel miniaturization technique of antenna is proposed using I-slot. Using this technique, gain of antenna can increased for 4dB (antenna only) to 6.6dB for the proposed I-slot antenna and a frequency shift of about 0.45 GHz to 1 GHz is obtained. Also a reduction of the shape size of the antenna is achieved (about 38 %) to operate in the Wi-Fi (2.45 GHz) band.RF Moreover the frequency shift can be controlled by changing the place or the length of the I-slot. Finally the proposed miniature antenna with an improved radiation efficiency and gain was built and tested.

Keywords: slot antenna, miniaturization, RF, electrical equivalent circuit (EEC)

Procedia PDF Downloads 277
20648 Real-Time Monitoring of Drinking Water Quality Using Advanced Devices

Authors: Amani Abdallah, Isam Shahrour

Abstract:

The quality of drinking water is a major concern of public health. The control of this quality is generally performed in the laboratory, which requires a long time. This type of control is not adapted for accidental pollution from sudden events, which can have serious consequences on population health. Therefore, it is of major interest to develop real-time innovative solutions for the detection of accidental contamination in drinking water systems This paper presents researches conducted within the SunRise Demonstrator for ‘Smart and Sustainable Cities’ with a particular focus on the supervision of the water quality. This work aims at (i) implementing a smart water system in a large water network (Campus of the University Lille1) including innovative equipment for real-time detection of abnormal events, such as those related to the contamination of drinking water and (ii) develop a numerical modeling of the contamination diffusion in the water distribution system. The first step included verification of the water quality sensors and their effectiveness on a network prototype of 50m length. This part included the evaluation of the efficiency of these sensors in the detection both bacterial and chemical contamination events in drinking water distribution systems. An on-line optical sensor integral with a laboratory-scale distribution system (LDS) was shown to respond rapidly to changes in refractive index induced by injected loads of chemical (cadmium, mercury) and biological contaminations (Escherichia coli). All injected substances were detected by the sensor; the magnitude of the response depends on the type of contaminant introduced and it is proportional to the injected substance concentration.

Keywords: distribution system, drinking water, refraction index, sensor, real-time

Procedia PDF Downloads 342
20647 Synthesis and Characterisation of Bi-Substituted Magnetite Nanoparticles by Mechanochemical Processing (MCP)

Authors: Morteza Mohri Esfahani, Amir S. H. Rozatian, Morteza Mozaffari

Abstract:

Single phase magnetite nanoparticles and Bi-substituted ones were prepared by mechanochemical processing (MCP). The effects of Bi-substitution on the structural and magnetic properties of the nanoparticles were studied by X-ray Diffraction (XRD) and magnetometry techniques, respectively. The XRD results showed that all samples have spinel phase and by increasing Bi content, the main diffraction peaks were shifted to higher angles, which means the lattice parameter decreases from 0.843 to 0.838 nm and then increases to 0.841 nm. Also, the results revealed that increasing Bi content lead to a decrease in saturation magnetization (Ms) from 74.9 to 48.8 emu/g and an increase in coercivity (Hc) from 96.8 to 137.1 Oe.

Keywords: bi-substituted magnetite nanoparticles, mechanochemical processing, X-ray diffraction, magnetism

Procedia PDF Downloads 525
20646 The Narrative Coherence of Autistic Children’s Accounts of an Experienced Event over Time

Authors: Fuming Yang, Telma Sousa Almeida, Xinyu Li, Yunxi Deng, Heying Zhang, Michael E. Lamb

Abstract:

Twenty-seven children aged 6-15 years with autism spectrum disorder (ASD) and 32 typically developing children were questioned about their participation in a set of activities after a two-week delay and again after a two-month delay, using a best-practice interview protocol. This paper assessed the narrative coherence of children’s reports based on key story grammar elements and temporal features included in their accounts of the event. Results indicated that, over time, both children with ASD and typically developing (TD) children decreased their narrative coherence. Children with ASD were no different from TD peers with regards to story length and syntactic complexity. However, they showed significantly less coherence than TD children. They were less likely to use the gist of the story to organize their narrative coherence. Interviewer prompts influenced children’s narrative coherence. The findings indicated that children with ASD could provide meaningful and reliable testimony about an event they personally experienced, but the narrative coherence of their reports deteriorates over time and is affected by interviewer prompts.

Keywords: autism spectrum disorders, delay, eyewitness testimony, narrative coherence

Procedia PDF Downloads 270
20645 Comparison of Frequency-Domain Contention Schemes in Wireless LANs

Authors: Li Feng

Abstract:

In IEEE 802.11 networks, it is well known that the traditional time-domain contention often leads to low channel utilization. The first frequency-domain contention scheme, the time to frequency (T2F), has recently been proposed to improve the channel utilization and has attracted a great deal of attention. In this paper, we survey the latest research progress on the weighed frequency-domain contention. We present the basic ideas, work principles of these related schemes and point out their differences. This paper is very useful for further study on frequency-domain contention.

Keywords: 802.11, wireless LANs, frequency-domain contention, T2F

Procedia PDF Downloads 446
20644 Achievements of Healthcare Services Vis-À-Vis the Millennium Development Goals Targets: Evidence from Pakistan

Authors: Saeeda Batool, Ather Maqsood Ahmed

Abstract:

This study investigates the impact of public healthcare facilities and socio-economic circumstances on the status of child health in Pakistan. The complete analysis is carried out in correspondence with fourth and sixth millennium development goals. Further, the health variables chosen are also inherited from targeted indicators of the mentioned goals (MDGs). Trends in the Human Opportunity Index (HOI) for both health inequalities and coverage are analyzed using the Pakistan Social and Living Standards Measurement (PLSM) data set for 2001-02 to 2012-13 at the national and provincial level. To reveal the relative importance of each circumstance in achieving the targeted values for child health, Shorrocks decomposition is applied on HOI. The annual point average growth rate of HOI is used to simulate the time period for the achievement of target set by MDGs and universal access also. The results indicate an improvement in HOI for a reduction in child mortality rates from 52.1% in 2001-02 to 67.3% in 2012-13, which confirms the availability of healthcare opportunities to a larger segment of society. Similarly, immunization against measles and other diseases such as Diphtheria, Polio, Bacillus Calmette-Guerin (BCG), and Hepatitis has also registered an improvement from 51.6% to 69.9% during the period of study at the national level. On a positive note, no gender disparity has been found for child health indicators and that health outcome is mostly affected by the parental and geographical features and availability of health infrastructure. However, the study finds that this achievement has been uneven across provinces. Pakistan is not only lagging behind in achieving its health goals, disappointingly with the current rate of health care provision, but it will take many additional years to achieve its targets.

Keywords: socio-economic circumstances, unmet MDGs, public healthcare services, child and infant mortality

Procedia PDF Downloads 217
20643 Optimization of the Numerical Fracture Mechanics

Authors: H. Hentati, R. Abdelmoula, Li Jia, A. Maalej

Abstract:

In this work, we present numerical simulations of the quasi-static crack propagation based on the variation approach. We perform numerical simulations of a piece of brittle material without initial crack. An alternate minimization algorithm is used. Based on these numerical results, we determine the influence of numerical parameters on the location of crack. We show the importance of trying to optimize the time of numerical computation and we present the first attempt to develop a simple numerical method to optimize this time.

Keywords: fracture mechanics, optimization, variation approach, mechanic

Procedia PDF Downloads 593
20642 The Effect of Penalizing Wrong Answers in the Computerized Modified Multiple Choice Testing System

Authors: Min Hae Song, Jooyong Park

Abstract:

Even though assessment using information and communication technology will most likely lead the future of educational assessment, there is little research on this topic. Computerized assessment will not only cut costs but also measure students' performance in ways not possible before. In this context, this study introduces a tool which can overcome the problems of multiple choice tests. Multiple-choice tests (MC) are efficient in automatic grading, however structural problems of multiple-choice tests allow students to find the correct answer from options even though they do not know the answer. A computerized modified multiple-choice testing system (CMMT) was developed using the interactivity of computers, that presents questions first, and options later for a short time when the student requests for them. This study was conducted to find out whether penalizing for wrong answers in CMMT could lower random guessing. In this study, we checked whether students knew the answers by having them respond to the short-answer tests before choosing the given options in CMMT or MC format. Ninety-four students were tested with the directions that they will be penalized for wrong answers, but not for no response. There were 4 experimental conditions: two conditions of high or low percentage of penalizing, each in traditional multiple-choice or CMMT format. In the low penalty condition, the penalty rate was the probability of getting the correct answer by random guessing. In the high penalty condition, students were penalized at twice the percentage of the low penalty condition. The results showed that the number of no response was significantly higher for the CMMT format and the number of random guesses was significantly lower for the CMMT format. There were no significant between the two penalty conditions. This result may be due to the fact that the actual score difference between the two conditions was too small. In the discussion, the possibility of applying CMMT format tests while penalizing wrong answers in actual testing settings was addressed.

Keywords: computerized modified multiple choice test format, multiple-choice test format, penalizing, test format

Procedia PDF Downloads 162
20641 A Hybrid Algorithm Based on Greedy Randomized Adaptive Search Procedure and Chemical Reaction Optimization for the Vehicle Routing Problem with Hard Time Windows

Authors: Imen Boudali, Marwa Ragmoun

Abstract:

The Vehicle Routing Problem with Hard Time Windows (VRPHTW) is a basic distribution management problem that models many real-world problems. The objective of the problem is to deliver a set of customers with known demands on minimum-cost vehicle routes while satisfying vehicle capacity and hard time windows for customers. In this paper, we propose to deal with our optimization problem by using a new hybrid stochastic algorithm based on two metaheuristics: Chemical Reaction Optimization (CRO) and Greedy Randomized Adaptive Search Procedure (GRASP). The first method is inspired by the natural process of chemical reactions enabling the transformation of unstable substances with excessive energy to stable ones. During this process, the molecules interact with each other through a series of elementary reactions to reach minimum energy for their existence. This property is embedded in CRO to solve the VRPHTW. In order to enhance the population diversity throughout the search process, we integrated the GRASP in our method. Simulation results on the base of Solomon’s benchmark instances show the very satisfactory performances of the proposed approach.

Keywords: Benchmark Problems, Combinatorial Optimization, Vehicle Routing Problem with Hard Time Windows, Meta-heuristics, Hybridization, GRASP, CRO

Procedia PDF Downloads 403
20640 Left Posterior Pericardiotomy in the Prevention of Post-Operative Atrial Fibrillation and Cardiac Tamponade: A Retrospective Study of 2118 Isolated Coronary Artery Bypass Graft Patients

Authors: Ayeshmanthe Rathnayake, Siew Goh, Carmel Fenton, Ashutosh Hardikar

Abstract:

Post-Operative Atrial Fibrillation (POAF) is the most frequent complication of cardiac surgery and is associated with reduced survival, increased rates of cognitive changes and cerebrovascular accident, heart failure, renal dysfunction, infection and length of stay, and hospital costs. Cardiac tamponade, although less common, carries high morbidity and mortality. Shed mediastinal blood in the pericardial space is a major source of intrapericardial oxidative stress and inflammation that triggers POAF. The utilisation of a left posterior pericardiotomy aims to shunt blood from the pericardium into the pleural space and have a role in the prevention of POAF as well as cardiac tamponade. 2118 patients had undergone isolated Coronary Artery Bypass Graft (CABG) at Royal Hobart Hospital from 2008-2021. They were divided into pericardiotomy vs control group. Patient baseline demographics, intraoperative data, and post-operative outcomes were reviewed retrospectively. Total incidence of new POAF and cardiac tamponade was 26.1% and 0.75%, respectively. Primary outcome of both the incidence of POAF(22.9% vs27.8%OR 0.77 p<0.05) and Cardiac Tamponade (0% vs 1.1% OR 0.85 p<0.05) were less in the pericardiotomy group.Increasing age, BMI, poor left ventricular function (EF <30%), and return to theatre were independent predictors of developing POAF. There were similar rates of return to theatre for bleeding however, no cases of tamponade in the pericardiotomy group. There were no complications attributable to left posterior pericardiotomy and the time added to the duration of surgery was minimal. Left posterior pericardiotomy is associated with a significant reduction in the incidence of POAFand cardiac tamponade and issafe and efficient.

Keywords: cardiac surgery, pericardiotomy, post-operative atrial fibrillation, cardiac tamponade

Procedia PDF Downloads 83
20639 Numerical Evolution Methods of Rational Form for Diffusion Equations

Authors: Said Algarni

Abstract:

The purpose of this study was to investigate selected numerical methods that demonstrate good performance in solving PDEs. We adapted alternative method that involve rational polynomials. Padé time stepping (PTS) method, which is highly stable for the purposes of the present application and is associated with lower computational costs, was applied. Furthermore, PTS was modified for our study which focused on diffusion equations. Numerical runs were conducted to obtain the optimal local error control threshold.

Keywords: Padé time stepping, finite difference, reaction diffusion equation, PDEs

Procedia PDF Downloads 289
20638 Managing Education through, Effective School Community Relationships/Participation for National Security

Authors: Shehu S. Janguza

Abstract:

The need for national security cannot be over Emphasis, which should be pursued by any means. Thus the need for effective management of education through effective school community Relationship/participation. In preparing and implementing only effort to promote community involvement in manning Education, it is importance to understand the whole picture of community participation, how it works, what forms are used, what benefit it can yield and what we should expect in the process of carrying out the efforts finally emphasis will be made on how effective school community relationship/participation and lead to national security.

Keywords: community participation, managing, school community, national security

Procedia PDF Downloads 587
20637 Enhancing Scalability in Ethereum Network Analysis: Methods and Techniques

Authors: Stefan K. Behfar

Abstract:

The rapid growth of the Ethereum network has brought forth the urgent need for scalable analysis methods to handle the increasing volume of blockchain data. In this research, we propose efficient methodologies for making Ethereum network analysis scalable. Our approach leverages a combination of graph-based data representation, probabilistic sampling, and parallel processing techniques to achieve unprecedented scalability while preserving critical network insights. Data Representation: We develop a graph-based data representation that captures the underlying structure of the Ethereum network. Each block transaction is represented as a node in the graph, while the edges signify temporal relationships. This representation ensures efficient querying and traversal of the blockchain data. Probabilistic Sampling: To cope with the vastness of the Ethereum blockchain, we introduce a probabilistic sampling technique. This method strategically selects a representative subset of transactions and blocks, allowing for concise yet statistically significant analysis. The sampling approach maintains the integrity of the network properties while significantly reducing the computational burden. Graph Convolutional Networks (GCNs): We incorporate GCNs to process the graph-based data representation efficiently. The GCN architecture enables the extraction of complex spatial and temporal patterns from the sampled data. This combination of graph representation and GCNs facilitates parallel processing and scalable analysis. Distributed Computing: To further enhance scalability, we adopt distributed computing frameworks such as Apache Hadoop and Apache Spark. By distributing computation across multiple nodes, we achieve a significant reduction in processing time and enhanced memory utilization. Our methodology harnesses the power of parallelism, making it well-suited for large-scale Ethereum network analysis. Evaluation and Results: We extensively evaluate our methodology on real-world Ethereum datasets covering diverse time periods and transaction volumes. The results demonstrate its superior scalability, outperforming traditional analysis methods. Our approach successfully handles the ever-growing Ethereum data, empowering researchers and developers with actionable insights from the blockchain. Case Studies: We apply our methodology to real-world Ethereum use cases, including detecting transaction patterns, analyzing smart contract interactions, and predicting network congestion. The results showcase the accuracy and efficiency of our approach, emphasizing its practical applicability in real-world scenarios. Security and Robustness: To ensure the reliability of our methodology, we conduct thorough security and robustness evaluations. Our approach demonstrates high resilience against adversarial attacks and perturbations, reaffirming its suitability for security-critical blockchain applications. Conclusion: By integrating graph-based data representation, GCNs, probabilistic sampling, and distributed computing, we achieve network scalability without compromising analytical precision. This approach addresses the pressing challenges posed by the expanding Ethereum network, opening new avenues for research and enabling real-time insights into decentralized ecosystems. Our work contributes to the development of scalable blockchain analytics, laying the foundation for sustainable growth and advancement in the domain of blockchain research and application.

Keywords: Ethereum, scalable network, GCN, probabilistic sampling, distributed computing

Procedia PDF Downloads 62
20636 Optimization of Process Parameters in Wire Electrical Discharge Machining of Inconel X-750 for Dimensional Deviation Using Taguchi Technique

Authors: Mandeep Kumar, Hari Singh

Abstract:

The effective optimization of machining process parameters affects dramatically the cost and production time of machined components as well as the quality of the final products. This paper presents the optimization aspects of a Wire Electrical Discharge Machining operation using Inconel X-750 as work material. The objective considered in this study is minimization of the dimensional deviation. Six input process parameters of WEDM namely spark gap voltage, pulse-on time, pulse-off time, wire feed rate, peak current and wire tension, were chosen as variables to study the process performance. Taguchi's design of experiments methodology has been used for planning and designing the experiments. The analysis of variance was carried out for raw data as well as for signal to noise ratio. Four input parameters and one two-factor interaction have been found to be statistically significant for their effects on the response of interest. The confirmation experiments were also performed for validating the predicted results.

Keywords: ANOVA, DOE, inconel, machining, optimization

Procedia PDF Downloads 195
20635 Assessment of Physical Characteristics of Maize (Zea Mays) Stored in Metallic Silos

Authors: B. A. Alabadan, E. S. Ajayi, C. A. Okolo

Abstract:

The storage losses recorded globally in maize (Zea mays) especially in the developing countries is worrisome. Certain degenerating changes in the physical characteristics (PC) of the grain occur due to the interaction between the stored maize and the immediate environment especially during long storage period. There has been tremendous reduction in the storage losses since the evolution of metallic silos. This study was carried out to assess the physical quality attributes of maize stored in 2500 MT and 1 MT metallic silos for a period of eight months. The PC evaluated includes percentage moisture content MC, insect damage ID, foreign matters FM, hectolitre weight HC, mould M and germinability VG. The evaluation of data obtained was done using statistical package for social sciences (SPSS 20) for windows evaluation version to determine significant levels and trend of deterioration (P < 0.05) for all the values obtained using Multiple Analysis of Variance (MANOVA) and Duncan’s multivariate test. The result shows that the PC are significant with duration of storage at (P < 0.05) except MI and FM that are significant at (P > 0.05) irrespective of the size of the metallic silos. The average mean deviation for physical properties from the control in respect to duration of storage are as follows: MC 10.0 ±0.00%, HC 72.9 ± 0.44% ID 0.29 ± 0.00%, BG 0.55±0.05%, MI 0.00 ± 0.65%, FM 0.80± 0.20%, VG 100 ± 0.03%. The variables that were found to be significant (p < 0.05) with the position of grain in the bulk are VG, MI and ID while others are insignificant at (p > 0.05). Variables were all significant (p < 0.05) with the duration of storage with (0.00) significant levels, irrespective of the size of the metallic silos, but were insignificant with the position of the grain in the bulk (p > 0.05). From the results, it can be concluded that there is a slight decrease of the following variables, with time, HC, MC, and V, probably due to weather fluctuations and grain respiration, while FM, BG, ID and M were found to increase slightly probably due to insect activity in the bigger silos and loss of moisture. The size of metallic silos has no remarkable influence on the PC of stored maize (Zea mays). Germinability was found to be better with the 1 MT silos probably due to its hermetic nature. Smaller size metallic silos are preferred for storage of seeds but bigger silos largely depend on the position of the grains in the bulk.

Keywords: maize, storage, silo, physical characteristics

Procedia PDF Downloads 292
20634 Relativistic Effects of Rotation

Authors: Yin Rui, Yin Ming, Wang Yang

Abstract:

For a rotational reference frame of the theory of special relativity, the critical radius is defined as the distance from the axis to the point where the tangential velocity is equal to the speed of light, and the critical cylinder as the set of all points separated from the axis by this critical radius. Based on these terms, two relativistic effects of rotation are discovered: (i) the tangential velocity in the region of Outside Critical Cylinder (OCC) is not superluminal due to the existence of space-time exchange; (ii) some of the physical quantities of the rotational body have an opposite mathematic sign at OCC versus those at Inside Critical Cylinder (ICC), which is termed as the Critical Cylindrical Effect (CCE). The laboratory experiments demonstrate that the repulsive force exerted on an anion by electrons will change to an attractive force by the electrons in precession while the anion is at OCC of the precession. Thirty-six screenshots from four experimental videos are provided. Theoretical proofs for both space-time exchange and CCE are then presented. The CCEs of field force are also discussed.

Keywords: critical radius, critical cylindrical effect, special relativity, space-time exchange

Procedia PDF Downloads 59
20633 Quantifying Stability of Online Communities and Its Impact on Disinformation

Authors: Victor Chomel, Maziyar Panahi, David Chavalarias

Abstract:

Misinformation has taken an increasingly worrying place in social media. Propagation patterns are closely linked to the structure of communities. This study proposes a method of community analysis based on a combination of centrality indicators for the network and its main communities. The objective is to establish a link between the stability of the communities over time, the social ascension of its members internally, and the propagation of information in the community. To this end, data from the debates about global warming and political communities on Twitter have been collected, and several tens of millions of tweets and retweets have helped us better understand the structure of these communities. The quantification of this stability allows for the study of the propagation of information of any kind, including disinformation. Our results indicate that the most stable communities over time are the ones that enable the establishment of nodes capturing a large part of the information and broadcasting its opinions. Conversely, communities with a high turnover and social ascendancy only stabilize themselves strongly in the face of adversity and external events but seem to offer a greater diversity of opinions most of the time.

Keywords: community analysis, disinformation, misinformation, Twitter

Procedia PDF Downloads 132
20632 A Method of Detecting the Difference in Two States of Brain Using Statistical Analysis of EEG Raw Data

Authors: Digvijaysingh S. Bana, Kiran R. Trivedi

Abstract:

This paper introduces various methods for the alpha wave to detect the difference between two states of brain. One healthy subject participated in the experiment. EEG was measured on the forehead above the eye (FP1 Position) with reference and ground electrode are on the ear clip. The data samples are obtained in the form of EEG raw data. The time duration of reading is of one minute. Various test are being performed on the alpha band EEG raw data.The readings are performed in different time duration of the entire day. The statistical analysis is being carried out on the EEG sample data in the form of various tests.

Keywords: electroencephalogram(EEG), biometrics, authentication, EEG raw data

Procedia PDF Downloads 453
20631 Multi-Objectives Genetic Algorithm for Optimizing Machining Process Parameters

Authors: Dylan Santos De Pinho, Nabil Ouerhani

Abstract:

Energy consumption of machine-tools is becoming critical for machine-tool builders and end-users because of economic, ecological and legislation-related reasons. Many machine-tool builders are seeking for solutions that allow the reduction of energy consumption of machine-tools while preserving the same productivity rate and the same quality of machined parts. In this paper, we present the first results of a project conducted jointly by academic and industrial partners to reduce the energy consumption of a Swiss-Type lathe. We employ genetic algorithms to find optimal machining parameters – the set of parameters that lead to the best trade-off between energy consumption, part quality and tool lifetime. Three main machining process parameters are considered in our optimization technique, namely depth of cut, spindle rotation speed and material feed rate. These machining process parameters have been identified as the most influential ones in the configuration of the Swiss-type machining process. A state-of-the-art multi-objective genetic algorithm has been used. The algorithm combines three fitness functions, which are objective functions that permit to evaluate a set of parameters against the three objectives: energy consumption, quality of the machined parts, and tool lifetime. In this paper, we focus on the investigation of the fitness function related to energy consumption. Four different energy consumption related fitness functions have been investigated and compared. The first fitness function refers to the Kienzle cutting force model. The second fitness function uses the Material Removal Rate (RMM) as an indicator of energy consumption. The two other fitness functions are non-deterministic, learning-based functions. One fitness function uses a simple Neural Network to learn the relation between the process parameters and the energy consumption from experimental data. Another fitness function uses Lasso regression to determine the same relation. The goal is, then, to find out which fitness functions predict best the energy consumption of a Swiss-Type machining process for the given set of machining process parameters. Once determined, these functions may be used for optimization purposes – determine the optimal machining process parameters leading to minimum energy consumption. The performance of the four fitness functions has been evaluated. The Tornos DT13 Swiss-Type Lathe has been used to carry out the experiments. A mechanical part including various Swiss-Type machining operations has been selected for the experiments. The evaluation process starts with generating a set of CNC (Computer Numerical Control) programs for machining the part at hand. Each CNC program considers a different set of machining process parameters. During the machining process, the power consumption of the spindle is measured. All collected data are assigned to the appropriate CNC program and thus to the set of machining process parameters. The evaluation approach consists in calculating the correlation between the normalized measured power consumption and the normalized power consumption prediction for each of the four fitness functions. The evaluation shows that the Lasso and Neural Network fitness functions have the highest correlation coefficient with 97%. The fitness function “Material Removal Rate” (MRR) has a correlation coefficient of 90%, whereas the Kienzle-based fitness function has a correlation coefficient of 80%.

Keywords: adaptive machining, genetic algorithms, smart manufacturing, parameters optimization

Procedia PDF Downloads 139
20630 Chemical Treatment of Wastewater through Biosorption for the Removal of Toxic Metals

Authors: Shafiq Alam, Manjunathan Ulaganathan

Abstract:

Water/wastewater often contains heavy/toxic metals, such as lead, copper, zinc and arsenic as well as harmful elements, such as antimony, selenium and fluoride. It may also contains radioactive elements, such as cesium and strontium. If they are not removed from water/wastewater then the environment and human health can be negatively impacted. Extensive research has been carried out to remove such harmful metals/elements from water/wastewater through biosorption using biomaterials (bioadsorbents). This presentation will give an overview of the research on preparation of bioadsorbents from biomass wastes and their use for the removal of harmful metals/elements from aqueous media.

Keywords: biosorption, environmental, toxic metals, wastewater

Procedia PDF Downloads 267
20629 Discrete Estimation of Spectral Density for Alpha Stable Signals Observed with an Additive Error

Authors: R. Sabre, W. Horrigue, J. C. Simon

Abstract:

This paper is interested in two difficulties encountered in practice when observing a continuous time process. The first is that we cannot observe a process over a time interval; we only take discrete observations. The second is the process frequently observed with a constant additive error. It is important to give an estimator of the spectral density of such a process taking into account the additive observation error and the choice of the discrete observation times. In this work, we propose an estimator based on the spectral smoothing of the periodogram by the polynomial Jackson kernel reducing the additive error. In order to solve the aliasing phenomenon, this estimator is constructed from observations taken at well-chosen times so as to reduce the estimator to the field where the spectral density is not zero. We show that the proposed estimator is asymptotically unbiased and consistent. Thus we obtain an estimate solving the two difficulties concerning the choice of the instants of observations of a continuous time process and the observations affected by a constant error.

Keywords: spectral density, stable processes, aliasing, periodogram

Procedia PDF Downloads 129
20628 Monte Carlo Simulations of LSO/YSO for Dose Evaluation in Photon Beam Radiotherapy

Authors: H. Donya

Abstract:

Monte Carlo (MC) techniques play a fundamental role in radiotherapy. A two non-water-equivalent of different media were used to evaluate the dose in water. For such purpose, Lu2SiO5 (LSO) and Y2SiO5 (YSO) orthosilicates scintillators are chosen for MC simulation using Penelope code. To get higher efficiency in dose calculation, variance reduction techniques are discussed. Overall results of this investigation ensured that the LSO/YSO bi-media a good combination to tackle over-response issue in dynamic photon radiotherapy.

Keywords: Lu2SiO5 (LSO) and Y2SiO5 (YSO) orthosilicates, Monte Carlo, correlated sampling, radiotherapy

Procedia PDF Downloads 389