Search results for: multivariate time series data
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 37816

Search results for: multivariate time series data

36736 Increases in Serum Erythropoietin Hormone in Recreational Breath-Hold Divers Following a Series of Repeated Apnoeas: Apnoea beyond Freediving

Authors: Antonis Elia, Theo Loizou, Gladys Onambele-Pearson, Matthew Barlow, Georgina Stebbings

Abstract:

Hypoxic conditions have been reported to enhance red blood cell production in both acclimatised low-landers and altitude adapted populations. This process is mediated by the erythropoietin hormone, which is released predominantly by the hypoxic kidney. A higher haemoglobin concentration was previously reported in elite breath-hold divers when compared to elite-skiers and untrained individuals. Therefore, the present study aimed to investigate whether apnoea induced hypoxia could induce a significant increase in serum erythropoietin concentration in recreational breath-hold divers which would provide an explanation to the higher haemoglobin levels observed in elite breath-hold divers. Identifying whether apnoea induced hypoxia induces a significant increase in serum erythropoietin might suggest that apnoea can be used as an alternative acclimatisation method to high altitude exposure. Seven healthy, recreational male breath-hold divers performed two sets of five 180 second breath-holds with a ten-minute supine rest between each set and a two-minute seated rest between each apnoea. During each breath-hold, participant’s heart rate and peripheral oxygen saturation levels were recorded every subsequent 10 seconds until the end of the 180 second breath-hold. After each 180 second breath-hold a capillary blood sample was collected from the finger to identify circulating haemoglobin levels. Following completion of the apnoeic protocol, three blood samples were collected at 30, 90 and 180 minutes to measure circulating erythropoietin levels. A significant interaction between erythropoietin and time was observed (F(3,18)= 4.72, p < 0.001), with significant increases in erythropoietin evident at 30 (t(6)= -5.035, p < 0.0590 (t(6)= -6.162, p < 0.05) and 180 (t(6)= - 7.232, p < 0.001) minutes post the last apnoea when compared to baseline. Corresponding average increases when compared to baseline were 16% at 30, 23% at 90 and 40% at 180 minutes post the last apnoea. A significant interaction between haemoglobin and time was observed (F(78,84)= 20.814, p < 0.001), with significant increases in haemoglobin evident at the fifth (t(29)= -1.124, p < 0.001), ninth (t(29)= -1.357, p < 0.001) and tenth (t(29)= -1.211, p < 0.05) apnoeas when compared to baseline. A significant interaction between peripheral oxygen saturation and time was observed (F(10,60)= 408.23, p < 0.001). The present study demonstrates that a series of ten 180 second breath-holds is sufficient to induce a significant increase in the circulating erythropoietin concentration of recreational breath hold divers. These observations may suggest that apnoea induced hypoxia may be used as an alternative acclimatisation method to high altitude exposure.

Keywords: apnoea, breath-holding, diving reflex, erythropoietin, haemoglobin

Procedia PDF Downloads 162
36735 Optimization of Real Time Measured Data Transmission, Given the Amount of Data Transmitted

Authors: Michal Kopcek, Tomas Skulavik, Michal Kebisek, Gabriela Krizanova

Abstract:

The operation of nuclear power plants involves continuous monitoring of the environment in their area. This monitoring is performed using a complex data acquisition system, which collects status information about the system itself and values of many important physical variables e.g. temperature, humidity, dose rate etc. This paper describes a proposal and optimization of communication that takes place in teledosimetric system between the central control server responsible for the data processing and storing and the decentralized measuring stations, which are measuring the physical variables. Analyzes of ongoing communication were performed and consequently the optimization of the system architecture and communication was done.

Keywords: communication protocol, transmission optimization, data acquisition, system architecture

Procedia PDF Downloads 497
36734 Design and Development of a Computerized Medical Record System for Hospitals in Remote Areas

Authors: Grace Omowunmi Soyebi

Abstract:

A computerized medical record system is a collection of medical information about a person that is stored on a computer. One principal problem of most hospitals in rural areas is using the file management system for keeping records. A lot of time is wasted when a patient visits the hospital, probably in an emergency, and the nurse or attendant has to search through voluminous files before the patient's file can be retrieved, this may cause an unexpected to happen to the patient. This Data Mining application is to be designed using a Structured System Analysis and design method which will help in a well-articulated analysis of the existing file management system, feasibility study, and proper documentation of the Design and Implementation of a Computerized medical record system. This Computerized system will replace the file management system and help to quickly retrieve a patient's record with increased data security, access clinical records for decision-making, and reduce the time range at which a patient gets attended to.

Keywords: programming, computing, data, innovation

Procedia PDF Downloads 102
36733 A Hybrid Data Mining Algorithm Based System for Intelligent Defence Mission Readiness and Maintenance Scheduling

Authors: Shivam Dwivedi, Sumit Prakash Gupta, Durga Toshniwal

Abstract:

It is a challenging task in today’s date to keep defence forces in the highest state of combat readiness with budgetary constraints. A huge amount of time and money is squandered in the unnecessary and expensive traditional maintenance activities. To overcome this limitation Defence Intelligent Mission Readiness and Maintenance Scheduling System has been proposed, which ameliorates the maintenance system by diagnosing the condition and predicting the maintenance requirements. Based on new data mining algorithms, this system intelligently optimises mission readiness for imminent operations and maintenance scheduling in repair echelons. With modified data mining algorithms such as Weighted Feature Ranking Genetic Algorithm and SVM-Random Forest Linear ensemble, it improves the reliability, availability and safety, alongside reducing maintenance cost and Equipment Out of Action (EOA) time. The results clearly conclude that the introduced algorithms have an edge over the conventional data mining algorithms. The system utilizing the intelligent condition-based maintenance approach improves the operational and maintenance decision strategy of the defence force.

Keywords: condition based maintenance, data mining, defence maintenance, ensemble, genetic algorithms, maintenance scheduling, mission capability

Procedia PDF Downloads 274
36732 The Lopsided Burden of Non-Communicable Diseases in India: Evidences from the Decade 2004-2014

Authors: Kajori Banerjee, Laxmi Kant Dwivedi

Abstract:

India is a part of the ongoing globalization, contemporary convergence, industrialization and technical advancement that is taking place world-wide. Some of the manifestations of this evolution is rapid demographic, socio-economic, epidemiological and health transition. There has been a considerable increase in non-communicable diseases due to change in lifestyle. This study aims to assess the direction of burden of disease and compare the pressure of infectious diseases against cardio-vascular, endocrine, metabolic and nutritional diseases. The change in prevalence in a ten-year period (2004-2014) is further decomposed to determine the net contribution of various socio-economic and demographic covariates. The present study uses the recent 71st (2014) and 60th (2004) rounds of National Sample Survey. The pressure of infectious diseases against cardio-vascular (CVD), endocrine, metabolic and nutritional (EMN) diseases during 2004-2014 is calculated by Prevalence Rates (PR), Hospitalization Rates (HR) and Case Fatality Rates (CFR). The prevalence of non-communicable diseases are further used as a dependent variable in a logit regression to find the effect of various social, economic and demographic factors on the chances of suffering from the particular disease. Multivariate decomposition technique further assists in determining the net contribution of socio-economic and demographic covariates. This paper upholds evidences of stagnation of the burden of communicable diseases (CD) and rapid increase in the burden of non-communicable diseases (NCD) uniformly for all population sub-groups in India. CFR for CVD has increased drastically in 2004-2014. Logit regression indicates the chances of suffering from CVD and EMN is significantly higher among the urban residents, older ages, females, widowed/ divorced and separated individuals. Decomposition displays ample proof that improvement in quality of life markers like education, urbanization, longevity of life has positively contributed in increasing the NCD prevalence rate. In India’s current epidemiological phase, compression theory of morbidity is in action as a significant rise in the probability of contracting the NCDs over the time period among older ages is observed. Age is found to play a vital contributor in increasing the probability of having CVD and EMN over the study decade 2004-2014 in the nationally representative sample of National Sample Survey.

Keywords: cardio-vascular disease, case-fatality rate, communicable diseases, hospitalization rate, multivariate decomposition, non-communicable diseases, prevalence rate

Procedia PDF Downloads 292
36731 Mobile Crowdsensing Scheme by Predicting Vehicle Mobility Using Deep Learning Algorithm

Authors: Monojit Manna, Arpan Adhikary

Abstract:

In Mobile cloud sensing across the globe, an emerging paradigm is selected by the user to compute sensing tasks. In urban cities current days, Mobile vehicles are adapted to perform the task of data sensing and data collection for universality and mobility. In this work, we focused on the optimality and mobile nodes that can be selected in order to collect the maximum amount of data from urban areas and fulfill the required data in the future period within a couple of minutes. We map out the requirement of the vehicle to configure the maximum data optimization problem and budget. The Application implementation is basically set up to generalize a realistic online platform in which real-time vehicles are moving apparently in a continuous manner. The data center has the authority to select a set of vehicles immediately. A deep learning-based scheme with the help of mobile vehicles (DLMV) will be proposed to collect sensing data from the urban environment. From the future time perspective, this work proposed a deep learning-based offline algorithm to predict mobility. Therefore, we proposed a greedy approach applying an online algorithm step into a subset of vehicles for an NP-complete problem with a limited budget. Real dataset experimental extensive evaluations are conducted for the real mobility dataset in Rome. The result of the experiment not only fulfills the efficiency of our proposed solution but also proves the validity of DLMV and improves the quantity of collecting the sensing data compared with other algorithms.

Keywords: mobile crowdsensing, deep learning, vehicle recruitment, sensing coverage, data collection

Procedia PDF Downloads 57
36730 Presenting a Model Based on Artificial Neural Networks to Predict the Execution Time of Design Projects

Authors: Hamed Zolfaghari, Mojtaba Kord

Abstract:

After feasibility study the design phase is started and the rest of other phases are highly dependent on this phase. forecasting the duration of design phase could do a miracle and would save a lot of time. This study provides a fast and accurate Machine learning (ML) and optimization framework, which allows a quick duration estimation of project design phase, hence improving operational efficiency and competitiveness of a design construction company. 3 data sets of three years composed of daily time spent for different design projects are used to train and validate the ML models to perform multiple projects. Our study concluded that Artificial Neural Network (ANN) performed an accuracy of 0.94.

Keywords: time estimation, machine learning, Artificial neural network, project design phase

Procedia PDF Downloads 65
36729 Real-Time Mine Safety System with the Internet of Things

Authors: Şakir Bingöl, Bayram İslamoğlu, Ebubekir Furkan Tepeli, Fatih Mehmet Karakule, Fatih Küçük, Merve Sena Arpacık, Mustafa Taha Kabar, Muhammet Metin Molak, Osman Emre Turan, Ömer Faruk Yesir, Sıla İnanır

Abstract:

This study introduces an IoT-based real-time safety system for mining, addressing global safety challenges. The wearable device, seamlessly integrated into miners' jackets, employs LoRa technology for communication and offers real-time monitoring of vital health and environmental data. Unique features include an LCD panel for immediate information display and sound-based location tracking for emergency response. The methodology involves sensor integration, data transmission, and ethical testing. Validation confirms the system's effectiveness in diverse mining scenarios. The study calls for ongoing research to adapt the system to different mining contexts, emphasizing its potential to significantly enhance safety standards in the industry.

Keywords: mining safety, internet of things, wearable technology, LoRa, RFID tracking, real-time safety system, safety alerts, safety measures

Procedia PDF Downloads 34
36728 Dynamic Change of Floods Disaster Monitoring for River Central Bar by Remote Sensing Time-Series Images

Authors: Zuoji Huang, Jinyan Sun, Chunlin Wang, Haiming Qian, Nan Xu

Abstract:

The spatial extent and area of central river bars can always vary due to the impact of water level, sediment supply and human activities. In 2016, a catastrophic flood disaster caused by sustained and heavy rainfall happened in the middle and lower Yangtze River. The flood led to the most serious economic and social loss since 1954, and strongly affected the central river bar. It is essential to continuously monitor the dynamics change of central bars because it can avoid frequent field measurements in central bars before and after the flood disaster and is helpful for flood warning. This paper focused on the dynamic change of central bars of Phoenix bar and Changsha bar in the Yangtze River in 2016. In this study, GF-1 (GaoFen-1) WFV(wide field view) data was employed owing to its high temporal frequency and high spatial resolution. A simple NDWI (Normalized Difference Water Index) method was utilized for river central bar mapping. Human-checking was then performed to ensure the mapping quality. The relationship between the area of central bars and the measured water level was estimated using four mathematical models. Furthermore, a risk assessment index was proposed to map the spatial pattern of inundation risk of central bars. The results indicate a good ability of the GF-1 WFV imagery with a 16-m spatial resolution to characterize the seasonal variation of central river bars and to capture the impact of a flood disaster on the area of central bars. This paper observed a significant negative but nonlinear relationship between the water level and the area of central bars, and found that the cubic function fits best among four models (R² = 0.9839, P < 0.000001, RMSE = 0.4395). The maximum of the inundated area of central bars appeared during the rainy season on July 8, 2016, and the minimum occurred during the dry season on December 28, 2016, which are consistent with the water level measured by the hydrological station. The results derived from GF-1 data could provide a useful reference for decision-making of real-time disaster early warning and post-disaster reconstruction.

Keywords: central bars, dynamic change, water level, the Yangtze river

Procedia PDF Downloads 226
36727 Black-Hole Dimension: A Distinct Methodology of Understanding Time, Space and Data in Architecture

Authors: Alp Arda

Abstract:

Inspired by Nolan's ‘Interstellar’, this paper delves into speculative architecture, asking, ‘What if an architect could traverse time to study a city?’ It unveils the ‘Black-Hole Dimension,’ a groundbreaking concept that redefines urban identities beyond traditional boundaries. Moving past linear time narratives, this approach draws from the gravitational dynamics of black holes to enrich our understanding of urban and architectural progress. By envisioning cities and structures as influenced by black hole-like forces, it enables an in-depth examination of their evolution through time and space. The Black-Hole Dimension promotes a temporal exploration of architecture, treating spaces as narratives of their current state interwoven with historical layers. It advocates for viewing architectural development as a continuous, interconnected journey molded by cultural, economic, and technological shifts. This approach not only deepens our understanding of urban evolution but also empowers architects and urban planners to create designs that are both adaptable and resilient. Echoing themes from popular culture and science fiction, this methodology integrates the captivating dynamics of time and space into architectural analysis, challenging established design conventions. The Black-Hole Dimension champions a philosophy that welcomes unpredictability and complexity, thereby fostering innovation in design. In essence, the Black-Hole Dimension revolutionizes architectural thought by emphasizing space-time as a fundamental dimension. It reimagines our built environments as vibrant, evolving entities shaped by the relentless forces of time, space, and data. This groundbreaking approach heralds a future in architecture where the complexity of reality is acknowledged and embraced, leading to the creation of spaces that are both responsive to their temporal context and resilient against the unfolding tapestry of time.

Keywords: black-hole, timeline, urbanism, space and time, speculative architecture

Procedia PDF Downloads 47
36726 Sensor Data Analysis for a Large Mining Major

Authors: Sudipto Shanker Dasgupta

Abstract:

One of the largest mining companies wanted to look at health analytics for their driverless trucks. These trucks were the key to their supply chain logistics. The automated trucks had multi-level sub-assemblies which would send out sensor information. The use case that was worked on was to capture the sensor signal from the truck subcomponents and analyze the health of the trucks from repair and replacement purview. Open source software was used to stream the data into a clustered Hadoop setup in Amazon Web Services cloud and Apache Spark SQL was used to analyze the data. All of this was achieved through a 10 node amazon 32 core, 64 GB RAM setup real-time analytics was achieved on ‘300 million records’. To check the scalability of the system, the cluster was increased to 100 node setup. This talk will highlight how Open Source software was used to achieve the above use case and the insights on the high data throughput on a cloud set up.

Keywords: streaming analytics, data science, big data, Hadoop, high throughput, sensor data

Procedia PDF Downloads 387
36725 QSAR, Docking and E-pharmacophore Approach on Novel Series of HDAC Inhibitors with Thiophene Linker as Anticancer Agents

Authors: Harish Rajak, Preeti Patel

Abstract:

HDAC inhibitors can reactivate gene expression and inhibit the growth and survival of cancer cells. The 3D-QSAR and Pharmacophore modeling studies were performed to identify important pharmacophoric features and correlate 3D-chemical structure with biological activity. The pharmacophore hypotheses were developed using e-pharmacophore script and phase module. Pharmacophore hypothesis represents the 3D arrangement of molecular features necessary for activity. A series of 55 compounds with well-assigned HDAC inhibitory activity was used for 3D-QSAR model development. Best 3D-QSAR model, which is a five PLS factor model with good statistics and predictive ability, acquired Q2 (0.7293), R2 (0.9811) and standard deviation (0.0952). Molecular docking were performed using Histone Deacetylase protein (PDB ID: 1t69) and prepared series of hydroxamic acid based HDAC inhibitors. Docking study of compound 43 show significant binding interactions Ser 276 and oxygen atom of dioxine cap region, Gly 151 and amino group and Asp 267 with carboxyl group of CONHOH, which are essential for anticancer activity. On docking, most of the compounds exhibited better glide score values between -8 to -10.5. We have established structure activity correlation using docking, energetic based pharmacophore modelling, pharmacophore and atom based 3D QSAR model. The results of these studies were further used for the design and testing of new HDAC analogs.

Keywords: Docking, e-pharmacophore, HDACIs, QSAR, Suberoylanilidehydroxamic acid.

Procedia PDF Downloads 281
36724 Polygenetic Iron Mineralization in the Baba-Ali and Galali Deposits, Further Evidences from Stable (S, O, H) Isotope Data, NW Hamedan, Iran

Authors: Ghodratollah Rostami Paydar

Abstract:

The Baba-Ali and Galali iron deposits are located in northwest Hamedan and the Iranian Sanandaj-Sirjan geological structural zone. The host rocks of these deposits are metavolcanosedimentary successions of Songhor stratigraphic series with permo-trriassic age. Field investigation, ore geometry, textures and structures and paragenetic sequence of minerals, all indicate that the ore minerals are crystallized in four stages: primary volcanosedimentary stage, secondary regional metamorphism with formation of ductile shear zones, contact metamorphism and metasomatism stage and the finally late hydrothermal mineralization within uplift and exposure. Totally 29 samples of sulfide, oxide-silicate and carbonate minerals of iron orees and gangue has been purified for stable isotope analysis. The isotope ratio data assure that occurrence of dynamothermal metamorphism in these areas typically involves a lengthy period of time, which results in a tendency toward isotopic homogenization specifically in O and H stable isotopes and showing the role of metamorphic waters in mineralization process. Measurement of δ34S (CDT) in first generation of pyrite is higher than another ones, so it confirms the volcanogenic origin of primary iron mineralization. δ13C data measurements in Galali carbonate country rocks show a marine origin. δ18O in magnetite and skarn forming silicates, δ18O and δ13C in limestone and skarn calcite and δ34S in sulphides are all consistent with the interaction of a magmatic-equilibrated fluid with Galali limestone, and a dominantly magmatic source for S. All these data imply skarn formation and mineralisation in a magmatic-hydrothermal system that maintained high salinity to relatively late stages resulting in the formation of the regional Na metasomatic alteration halo. Late stage hydrothermal quartz-calcite veinlets are important for gold mineralization, but the economic evaluation is required to detailed geochemical studies.

Keywords: iron, polygenetic, stable isotope, BabaAli, Galali

Procedia PDF Downloads 287
36723 Development of Academic Software for Medial Axis Determination of Porous Media from High-Resolution X-Ray Microtomography Data

Authors: S. Jurado, E. Pazmino

Abstract:

Determination of the medial axis of a porous media sample is a non-trivial problem of interest for several disciplines, e.g., hydrology, fluid dynamics, contaminant transport, filtration, oil extraction, etc. However, the computational tools available for researchers are limited and restricted. The primary aim of this work was to develop a series of algorithms to extract porosity, medial axis structure, and pore-throat size distributions from porous media domains. A complementary objective was to provide the algorithms as free computational software available to the academic community comprising researchers and students interested in 3D data processing. The burn algorithm was tested on porous media data obtained from High-Resolution X-Ray Microtomography (HRXMT) and idealized computer-generated domains. The real data and idealized domains were discretized in voxels domains of 550³ elements and binarized to denote solid and void regions to determine porosity. Subsequently, the algorithm identifies the layer of void voxels next to the solid boundaries. An iterative process removes or 'burns' void voxels in sequence of layer by layer until all the void space is characterized. Multiples strategies were tested to optimize the execution time and use of computer memory, i.e., segmentation of the overall domain in subdomains, vectorization of operations, and extraction of single burn layer data during the iterative process. The medial axis determination was conducted identifying regions where burnt layers collide. The final medial axis structure was refined to avoid concave-grain effects and utilized to determine the pore throat size distribution. A graphic user interface software was developed to encompass all these algorithms, including the generation of idealized porous media domains. The software allows input of HRXMT data to calculate porosity, medial axis, and pore-throat size distribution and provide output in tabular and graphical formats. Preliminary tests of the software developed during this study achieved medial axis, pore-throat size distribution and porosity determination of 100³, 320³ and 550³ voxel porous media domains in 2, 22, and 45 minutes, respectively in a personal computer (Intel i7 processor, 16Gb RAM). These results indicate that the software is a practical and accessible tool in postprocessing HRXMT data for the academic community.

Keywords: medial axis, pore-throat distribution, porosity, porous media

Procedia PDF Downloads 98
36722 Design of a Real Time Heart Sounds Recognition System

Authors: Omer Abdalla Ishag, Magdi Baker Amien

Abstract:

Physicians used the stethoscope for listening patient heart sounds in order to make a diagnosis. However, the determination of heart conditions by acoustic stethoscope is a difficult task so it requires special training of medical staff. This study developed an accurate model for analyzing the phonocardiograph signal based on PC and DSP processor. The system has been realized into two phases; offline and real time phase. In offline phase, 30 cases of heart sounds files were collected from medical students and doctor's world website. For experimental phase (real time), an electronic stethoscope has been designed, implemented and recorded signals from 30 volunteers, 17 were normal cases and 13 were various pathologies cases, these acquired 30 signals were preprocessed using an adaptive filter to remove lung sounds. The background noise has been removed from both offline and real data, using wavelet transform, then graphical and statistics features vector elements were extracted, finally a look-up table was used for classification heart sounds cases. The obtained results of the implemented system showed accuracy of 90%, 80% and sensitivity of 87.5%, 82.4% for offline data, and real data respectively. The whole system has been designed on TMS320VC5509a DSP Platform.

Keywords: code composer studio, heart sounds, phonocardiograph, wavelet transform

Procedia PDF Downloads 423
36721 Modeling Usage Patterns of Mobile App Service in App Market Using Hidden Markov Model

Authors: Yangrae Cho, Jinseok Kim, Yongtae Park

Abstract:

Mobile app service ecosystem has been abruptly emerged, explosively grown, and dynamically transformed. In contrast with product markets in which product sales directly cause increment in firm’s income, customer’s usage is less visible but more valuable in service market. Especially, the market situation with cutthroat competition in mobile app store makes securing and keeping of users as vital. Although a few service firms try to manage their apps’ usage patterns by fitting on S-curve or applying other forecasting techniques, the time series approaches based on past sequential data are subject to fundamental limitation in the market where customer’s attention is being moved unpredictably and dynamically. We therefore propose a new conceptual approach for detecting usage pattern of mobile app service with Hidden Markov Model (HMM) which is based on the dual stochastic structure and mainly used to clarify unpredictable and dynamic sequential patterns in voice recognition or stock forecasting. Our approach could be practically utilized for app service firms to manage their services’ lifecycles and academically expanded to other markets.

Keywords: mobile app service, usage pattern, Hidden Markov Model, pattern detection

Procedia PDF Downloads 311
36720 Annexing the Strength of Information and Communication Technology (ICT) for Real-time TB Reporting Using TB Situation Room (TSR) in Nigeria: Kano State Experience

Authors: Ibrahim Umar, Ashiru Rajab, Sumayya Chindo, Emmanuel Olashore

Abstract:

INTRODUCTION: Kano is the most populous state in Nigeria and one of the two states with the highest TB burden in the country. The state notifies an average of 8,000+ TB cases quarterly and has the highest yearly notification of all the states in Nigeria from 2020 to 2022. The contribution of the state TB program to the National TB notification varies from 9% to 10% quarterly between the first quarter of 2022 and second quarter of 2023. The Kano State TB Situation Room is an innovative platform for timely data collection, collation and analysis for informed decision in health system. During the 2023 second National TB Testing week (NTBTW) Kano TB program aimed at early TB detection, prevention and treatment. The state TB Situation room provided avenue to the state for coordination and surveillance through real time data reporting, review, analysis and use during the NTBTW. OBJECTIVES: To assess the role of innovative information and communication technology platform for real-time TB reporting during second National TB Testing week in Nigeria 2023. To showcase the NTBTW data cascade analysis using TSR as innovative ICT platform. METHODOLOGY: The State TB deployed a real-time virtual dashboard for NTBTW reporting, analysis and feedback. A data room team was set up who received realtime data using google link. Data received was analyzed using power BI analytic tool with statistical alpha level of significance of <0.05. RESULTS: At the end of the week-long activity and using the real-time dashboard with onsite mentorship of the field workers, the state TB program was able to screen a total of 52,054 people were screened for TB from 72,112 individuals eligible for screening (72% screening rate). A total of 9,910 presumptive TB clients were identified and evaluated for TB leading to diagnosis of 445 TB patients with TB (5% yield from presumptives) and placement of 435 TB patients on treatment (98% percentage enrolment). CONCLUSION: The TB Situation Room (TBSR) has been a great asset to Kano State TB Control Program in meeting up with the growing demand for timely data reporting in TB and other global health responses. The use of real time surveillance data during the 2023 NTBTW has in no small measure improved the TB response and feedback in Kano State. Scaling up this intervention to other disease areas, states and nations is a positive step in the right direction towards global TB eradication.

Keywords: tuberculosis (tb), national tb testing week (ntbtw), tb situation rom (tsr), information communication technology (ict)

Procedia PDF Downloads 48
36719 The Study on Corpse Floating Time in Shanghai Region of China

Authors: Hang Meng, Wen-Bin Liu, Bi Xiao, Kai-Jun Ma, Jian-Hui Xie, Geng Fei, Tian-Ye Zhang, Lu-Yi Xu, Dong-Chuan Zhang

Abstract:

The victims in water are often found in the coastal region, along river region or the region with lakes. In China, the examination for the bodies of victims in the water is conducted by forensic doctors working in the public security bureau. Because the enter water time for most of the victims are not clear, and often lack of monitor images and other information, so to find out the corpse enter water time for victims is very difficult. After the corpse of the victim enters the water, it sinks first, then corruption gas produces, which can make the density of the corpse to be less than water, and thus rise again. So the factor that determines the corpse floating time is temperature. On the basis of the temperature data obtained in Shanghai region of China (Shanghai is a north subtropical marine monsoon climate, with an average annual temperature of about 17.1℃. The hottest month is July, the average monthly temperature is 28.6℃, and the coldest month is January, the average monthly temperature is 4.8℃). This study selected about 100 cases with definite corpse enter water time and corpse floating time, analyzed the cases and obtained the empirical law of the corpse floating time. For example, in the Shanghai region, on June 15th and October 15th, the corpse floating time is about 1.5 days. In early December, the bodies who entered the water will go up around January 1st of the following year, and the bodies who enter water in late December will float in March of next year. The results of this study can be used to roughly estimate the water enter time of the victims in Shanghai. Forensic doctors around the world can also draw on the results of this study to infer the time when the corpses of the victims in the water go up.

Keywords: corpse enter water time, corpse floating time, drowning, forensic pathology, victims in the water

Procedia PDF Downloads 176
36718 Designing Price Stability Model of Red Cayenne Pepper Price in Wonogiri District, Centre Java, Using ARCH/GARCH Method

Authors: Fauzia Dianawati, Riska W. Purnomo

Abstract:

Food and agricultural sector become the biggest sector contributing to inflation in Indonesia. Especially in Wonogiri district, red cayenne pepper was the biggest sector contributing to inflation on 2016. A national statistic proved that in recent five years red cayenne pepper has the highest average level of fluctuation among all commodities. Some factors, like supply chain, price disparity, production quantity, crop failure, and oil price become the possible factor causes high volatility level in red cayenne pepper price. Therefore, this research tries to find the key factor causing fluctuation on red cayenne pepper by using ARCH/GARCH method. The method could accommodate the presence of heteroscedasticity in time series data. At the end of the research, it is statistically found that the second level of supply chain becomes the biggest part contributing to inflation with 3,35 of coefficient in fluctuation forecasting model of red cayenne pepper price. This model could become a reference to the government to determine the appropriate policy in maintaining the price stability of red cayenne pepper.

Keywords: ARCH/GARCH, forecasting, red cayenne pepper, volatility, supply chain

Procedia PDF Downloads 166
36717 Research of Data Cleaning Methods Based on Dependency Rules

Authors: Yang Bao, Shi Wei Deng, WangQun Lin

Abstract:

This paper introduces the concept and principle of data cleaning, analyzes the types and causes of dirty data, and proposes several key steps of typical cleaning process, puts forward a well scalability and versatility data cleaning framework, in view of data with attribute dependency relation, designs several of violation data discovery algorithms by formal formula, which can obtain inconsistent data to all target columns with condition attribute dependent no matter data is structured (SQL) or unstructured (NoSQL), and gives 6 data cleaning methods based on these algorithms.

Keywords: data cleaning, dependency rules, violation data discovery, data repair

Procedia PDF Downloads 540
36716 The Prognostic Prediction Value of Positive Lymph Nodes Numbers for the Hypopharyngeal Squamous Cell Carcinoma

Authors: Wendu Pang, Yaxin Luo, Junhong Li, Yu Zhao, Danni Cheng, Yufang Rao, Minzi Mao, Ke Qiu, Yijun Dong, Fei Chen, Jun Liu, Jian Zou, Haiyang Wang, Wei Xu, Jianjun Ren

Abstract:

We aimed to compare the prognostic prediction value of positive lymph node number (PLNN) to the American Joint Committee on Cancer (AJCC) tumor, lymph node, and metastasis (TNM) staging system for patients with hypopharyngeal squamous cell carcinoma (HPSCC). A total of 826 patients with HPSCC from the Surveillance, Epidemiology, and End Results database (2004–2015) were identified and split into two independent cohorts: training (n=461) and validation (n=365). Univariate and multivariate Cox regression analyses were used to evaluate the prognostic effects of PLNN in patients with HPSCC. We further applied six Cox regression models to compare the survival predictive values of the PLNN and AJCC TNM staging system. PLNN showed a significant association with overall survival (OS) and cancer-specific survival (CSS) (P < 0.001) in both univariate and multivariable analyses, and was divided into three groups (PLNN 0, PLNN 1-5, and PLNN>5). In the training cohort, multivariate analysis revealed that the increased PLNN of HPSCC gave rise to significantly poor OS and CSS after adjusting for age, sex, tumor size, and cancer stage; this trend was also verified by the validation cohort. Additionally, the survival model incorporating a composite of PLNN and TNM classification (C-index, 0.705, 0.734) performed better than the PLNN and AJCC TNM models. PLNN can serve as a powerful survival predictor for patients with HPSCC and is a surrogate supplement for cancer staging systems.

Keywords: hypopharyngeal squamous cell carcinoma, positive lymph nodes number, prognosis, prediction models, survival predictive values

Procedia PDF Downloads 124
36715 Temporal Axis in Japanese: The Paradox of a Metaphorical Orientation in Time

Authors: Tomoko Usui

Abstract:

In the field of linguistics, it has been said that concepts associated with space and motion systematically contribute structure to the temporal concept. This is the conceptual metaphor theory. conceptual metaphors typically employ a more abstract concept (time) as their target and a more concrete or physical concept as their source (space). This paper will examine two major temporal conceptual metaphors: Ego-centered Moving Time Metaphor and Time-RP Metaphor. Moving time generally receives a front-back orientation, however, Japanese shows a different orientation given to time. By means of Ego perspective, this paper will illustrate the paradox of a metaphorical orientation in time.

Keywords: Ego-centered Moving Time Metaphor, Japanese saki, temporal metaphors, Time RP Metaphor

Procedia PDF Downloads 477
36714 Optimal and Best Timing for Capturing Satellite Thermal Images of Concrete Object

Authors: Toufic Abd El-Latif Sadek

Abstract:

The concrete object represents the concrete areas, like buildings. The best, easy, and efficient extraction of the concrete object from satellite thermal images occurred at specific times during the days of the year, by preventing the gaps in times which give the close and same brightness from different objects. Thus, to achieve the best original data which is the aim of the study and then better extraction of the concrete object and then better analysis. The study was done using seven sample objects, asphalt, concrete, metal, rock, dry soil, vegetation, and water, located at one place carefully investigated in a way that all the objects achieve the homogeneous in acquired data at the same time and same weather conditions. The samples of the objects were on the roof of building at position taking by global positioning system (GPS) which its geographical coordinates is: Latitude= 33 degrees 37 minutes, Longitude= 35 degrees 28 minutes, Height= 600 m. It has been found that the first choice and the best time in February is at 2:00 pm, in March at 4 pm, in April and may at 12 pm, in August at 5:00 pm, in October at 11:00 am. The best time in June and November is at 2:00 pm.

Keywords: best timing, concrete areas, optimal, satellite thermal images

Procedia PDF Downloads 334
36713 Adaptive Swarm Balancing Algorithms for Rare-Event Prediction in Imbalanced Healthcare Data

Authors: Jinyan Li, Simon Fong, Raymond Wong, Mohammed Sabah, Fiaidhi Jinan

Abstract:

Clinical data analysis and forecasting have make great contributions to disease control, prevention and detection. However, such data usually suffer from highly unbalanced samples in class distributions. In this paper, we target at the binary imbalanced dataset, where the positive samples take up only the minority. We investigate two different meta-heuristic algorithms, particle swarm optimization and bat-inspired algorithm, and combine both of them with the synthetic minority over-sampling technique (SMOTE) for processing the datasets. One approach is to process the full dataset as a whole. The other is to split up the dataset and adaptively process it one segment at a time. The experimental results reveal that while the performance improvements obtained by the former methods are not scalable to larger data scales, the later one, which we call Adaptive Swarm Balancing Algorithms, leads to significant efficiency and effectiveness improvements on large datasets. We also find it more consistent with the practice of the typical large imbalanced medical datasets. We further use the meta-heuristic algorithms to optimize two key parameters of SMOTE. Leading to more credible performances of the classifier, and shortening the running time compared with the brute-force method.

Keywords: Imbalanced dataset, meta-heuristic algorithm, SMOTE, big data

Procedia PDF Downloads 424
36712 Authorization of Commercial Communication Satellite Grounds for Promoting Turkish Data Relay System

Authors: Celal Dudak, Aslı Utku, Burak Yağlioğlu

Abstract:

Uninterrupted and continuous satellite communication through the whole orbit time is becoming more indispensable every day. Data relay systems are developed and built for various high/low data rate information exchanges like TDRSS of USA and EDRSS of Europe. In these missions, a couple of task-dedicated communication satellites exist. In this regard, for Turkey a data relay system is attempted to be defined exchanging low data rate information (i.e. TTC) for Earth-observing LEO satellites appointing commercial GEO communication satellites all over the world. First, justification of this attempt is given, demonstrating duration enhancements in the link. Discussion of preference of RF communication is, also, given instead of laser communication. Then, preferred communication GEOs – including TURKSAT4A already belonging to Turkey- are given, together with the coverage enhancements through STK simulations and the corresponding link budget. Also, a block diagram of the communication system is given on the LEO satellite.

Keywords: communication, GEO satellite, data relay system, coverage

Procedia PDF Downloads 418
36711 Comparison of Self-Efficacy and Life Satisfaction in Normal Users and Users with Internet Addiction

Authors: Mansour Abdi, Hadi Molaei Yasavoli

Abstract:

The purpose of this research is to comparison of self- efficacy and life satisfaction in normal users and users with internet addiction. The present study was descriptive and causal-comparative. Therefore, 304 students were selected random sampling method from students of Semnan University and completed questionnaires of internet addiction (young), Self-Efficacy Questionnaire and Life Satisfaction (SWIS). For data analysis was used the Multivariate Analysis of Variance (MANOVA). The results showed that internet addiction users have lower levels of self-efficacy and life satisfaction in comparison with normal users and the difference in p=0/0005 significantly. The findings showed that 78 percent of the variance in the dependent variables of self-efficacy and life satisfaction by grouping variables (internet addiction users and normal) is determined. Finally, considering that the rate of self-efficacy and life satisfaction is effective in the incidence of Internet addiction, it is proposed required measures are taken to enhance self-efficacy and life satisfaction in Internet users.

Keywords: self-efficacy, life satisfaction, users, internet addiction, normal users

Procedia PDF Downloads 469
36710 SisGeo: Support System for the Research of Georeferenced Comparisons Applied to Professional and Academic Devices

Authors: Bruno D. Souza, Gerson G. Cunha, Michael O. Ferreira, Roberto Rosenhaim, Robson C. Santos, Sergio O. Santos

Abstract:

Devices and applications that use satellite-based positioning are becoming more popular day-by-day. Thus, evolution and improvement in this technology are mandatory. Accordingly, satellite georeferenced systems need to accomplish the same evolution rhythm. Either GPS (Global Positioning System) or its similar Russian GLONASS (Global Navigation Satellite System) are system samples that offer us powerful tools to plot coordinates on the earth surface. The development of this research aims the study of several aspects related to use of GPS and GLONASS technologies, given its application and collected data improvement during geodetic data acquisition. So, both relevant theoretic and practical aspects are considered. In this context, at the theoretical part, the main systems' characteristics are shown, observing its similarities and differences. At the practical part, a series of experiences are performed and obtained data packages are compared in order to demonstrate equivalence or differences among them. The evaluation methodology targets both quantitative and qualitative analysis provided by GPS and GPS/GLONASS receptors. Meanwhile, a specific collected data storage system was developed to better compare and analyze them (SisGeo - Georeferenced Research Comparison Support System).

Keywords: satellites, systems, applications, experiments, receivers

Procedia PDF Downloads 232
36709 Improving Cost and Time Control of Construction Projects Management Practices in Nigeria

Authors: Mustapha Yakubu, Ahmed Usman, Hashim Ambursa

Abstract:

This paper presents the findings of a research which sought to investigate techniques used to improve cost and time control of construction projects management practice in Nigeria. However, there is limited research on issues surrounding the practical usage of these techniques. Data were collected through a questionnaire distributed to construction experts through a survey conducted on the 100 construction organisations and 50 construction consultancy firms in the Nigeria aimed at identifying common project cost and time control practices and factors inhibiting effective project control in practice. The study reveals that despite the vast application of control techniques a high proportion of respondents still experienced cost and time overruns on a significant proportion of their projects. Analysis of the survey results concluded that more effort should be geared at the management of the identified top project control inhibiting factors. This paper has outlined some measures for mitigating these inhibiting factors so that the outcome of project time and cost control can be improved in practice.

Keywords: construction project, cost control, Nigeria, time control

Procedia PDF Downloads 287
36708 The Effect of Improvement Programs in the Mean Time to Repair and in the Mean Time between Failures on Overall Lead Time: A Simulation Using the System Dynamics-Factory Physics Model

Authors: Marcel Heimar Ribeiro Utiyama, Fernanda Caveiro Correia, Dario Henrique Alliprandini

Abstract:

The importance of the correct allocation of improvement programs is of growing interest in recent years. Due to their limited resources, companies must ensure that their financial resources are directed to the correct workstations in order to be the most effective and survive facing the strong competition. However, to our best knowledge, the literature about allocation of improvement programs does not analyze in depth this problem when the flow shop process has two capacity constrained resources. This is a research gap which is deeply studied in this work. The purpose of this work is to identify the best strategy to allocate improvement programs in a flow shop with two capacity constrained resources. Data were collected from a flow shop process with seven workstations in an industrial control and automation company, which process 13.690 units on average per month. The data were used to conduct a simulation with the System Dynamics-Factory Physics model. The main variables considered, due to their importance on lead time reduction, were the mean time between failures and the mean time to repair. The lead time reduction was the output measure of the simulations. Ten different strategies were created: (i) focused time to repair improvement, (ii) focused time between failures improvement, (iii) distributed time to repair improvement, (iv) distributed time between failures improvement, (v) focused time to repair and time between failures improvement, (vi) distributed time to repair and between failures improvement, (vii) hybrid time to repair improvement, (viii) hybrid time between failures improvements, (ix) time to repair improvement strategy towards the two capacity constrained resources, (x) time between failures improvement strategy towards the two capacity constrained resources. The ten strategies tested are variations of the three main strategies for improvement programs named focused, distributed and hybrid. Several comparisons among the effect of the ten strategies in lead time reduction were performed. The results indicated that for the flow shop analyzed, the focused strategies delivered the best results. When it is not possible to perform a large investment on the capacity constrained resources, companies should use hybrid approaches. An important contribution to the academy is the hybrid approach, which proposes a new way to direct the efforts of improvements. In addition, the study in a flow shop with two strong capacity constrained resources (more than 95% of utilization) is an important contribution to the literature. Another important contribution is the problem of allocation with two CCRs and the possibility of having floating capacity constrained resources. The results provided the best improvement strategies considering the different strategies of allocation of improvement programs and different positions of the capacity constrained resources. Finally, it is possible to state that both strategies, hybrid time to repair improvement and hybrid time between failures improvement, delivered best results compared to the respective distributed strategies. The main limitations of this study are mainly regarding the flow shop analyzed. Future work can further investigate different flow shop configurations like a varying number of workstations, different number of products or even different positions of the two capacity constrained resources.

Keywords: allocation of improvement programs, capacity constrained resource, hybrid strategy, lead time, mean time to repair, mean time between failures

Procedia PDF Downloads 103
36707 Exploring Influence Range of Tainan City Using Electronic Toll Collection Big Data

Authors: Chen Chou, Feng-Tyan Lin

Abstract:

Big Data has been attracted a lot of attentions in many fields for analyzing research issues based on a large number of maternal data. Electronic Toll Collection (ETC) is one of Intelligent Transportation System (ITS) applications in Taiwan, used to record starting point, end point, distance and travel time of vehicle on the national freeway. This study, taking advantage of ETC big data, combined with urban planning theory, attempts to explore various phenomena of inter-city transportation activities. ETC, one of government's open data, is numerous, complete and quick-update. One may recall that living area has been delimited with location, population, area and subjective consciousness. However, these factors cannot appropriately reflect what people’s movement path is in daily life. In this study, the concept of "Living Area" is replaced by "Influence Range" to show dynamic and variation with time and purposes of activities. This study uses data mining with Python and Excel, and visualizes the number of trips with GIS to explore influence range of Tainan city and the purpose of trips, and discuss living area delimited in current. It dialogues between the concepts of "Central Place Theory" and "Living Area", presents the new point of view, integrates the application of big data, urban planning and transportation. The finding will be valuable for resource allocation and land apportionment of spatial planning.

Keywords: Big Data, ITS, influence range, living area, central place theory, visualization

Procedia PDF Downloads 261