Search results for: multivariate time series data
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 37847

Search results for: multivariate time series data

32387 An Analysis of the Recent Flood Scenario (2017) of the Southern Districts of the State of West Bengal, India

Authors: Soumita Banerjee

Abstract:

The State of West Bengal is mostly watered by innumerable rivers, and they are different in nature in both the northern and the southern part of the state. The southern part of West Bengal is mainly drained with the river Bhagirathi-Hooghly, and its major distributaries and tributaries have divided this major river basin into many subparts like the Ichamati-Bidyadhari, Pagla-Bansloi, Mayurakshi-Babla, Ajay, Damodar, Kangsabati Sub-basin to name a few. These rivers basically drain the Districts of Bankura, Burdwan, Hooghly, Nadia and Purulia, Birbhum, Midnapore, Murshidabad, North 24-Parganas, Kolkata, Howrah and South 24-Parganas. West Bengal has a huge number of flood-prone blocks in the southern part of the state of West Bengal, the responsible factors for flood situation are the shape and size of the catchment area, its steep gradient starting from plateau to flat terrain, the river bank erosion and its siltation, tidal condition especially in the lower Ganga Basin and very low maintenance of the embankments which are mostly used as communication links. Along with these factors, DVC (Damodar Valley Corporation) plays an important role in the generation (with the release of water) and controlling the flood situation. This year the whole Gangetic West Bengal is being flooded due to high intensity and long duration rainfall, and the release of water from the Durgapur Barrage As most of the rivers are interstate in nature at times floods also take place with release of water from the dams of the neighbouring states like Jharkhand. Other than Embankments, there is no such structural measures for combatting flood in West Bengal. This paper tries to analyse the reasons behind the flood situation this year especially with the help of climatic data collected from the Indian Metrological Department, flood related data from the Irrigation and Waterways Department, West Bengal and GPM (General Precipitation Measurement) data for rainfall analysis. Based on the threshold value derived from the calculation of the past available flood data, it is possible to predict the flood events which may occur in the near future and with the help of social media it can be spread out within a very short span of time to aware the mass. On a larger or a governmental scale, heightening the settlements situated on the either banks of the river can yield a better result than building up embankments.

Keywords: dam failure, embankments, flood, rainfall

Procedia PDF Downloads 207
32386 Nanocrystalline Na0.1V2O5.nH2Oxerogel Thin Film for Gas Sensing

Authors: M. S. Al-Assiri, M. M. El-Desoky, A. A. Bahgat

Abstract:

Nanocrystalline thin film of Na0.1V2O5.nH2O xerogel obtained by sol-gel synthesis was used as a gas sensor. Gas sensing properties of different gases such as hydrogen, petroleum and humidity were investigated. Applying XRD and TEM the size of the nanocrystals is found to be 7.5 nm. SEM shows a highly porous structure with submicron meter-sized voids present throughout the sample. FTIR measurement shows different chemical groups identifying the obtained series of gels. The sample was n-type semiconductor according to the thermoelectric power and electrical conductivity. It can be seen that the sensor response curves from 130°C to 150°C show a rapid increase in sensitivity for all types of gas injection, low response values for heating period and the rapid high response values for cooling period. This result may suggest that this material is able to act as gas sensor during the heating and cooling process.

Keywords: sol-gel, thermoelectric power, XRD, TEM, gas sensing

Procedia PDF Downloads 290
32385 Transient Electrical Resistivity and Elastic Wave Velocity of Sand-Cement-Inorganic Binder Mixture

Authors: Kiza Rusati Pacifique, Ki-il Song

Abstract:

The cement milk grout has been used for ground improvement. Due to the environmental issues related to cement, the reduction of cement usage is requesting. In this study, inorganic binder is introduced to reduce the use of cement contents for ground improvement. To evaluate transient electrical and mechanical properties of sand-cement-inorganic binder mixture, two non-destructive testing (NDT) methods, Electrical Resistivity (ER) and Free Free Resonant Column (FFRC) tests were adopted in addition to unconfined compressive strength test. Electrical resistivity, longitudinal wave velocity and damping ratio of sand-cement admixture samples improved with addition of inorganic binders were measured. Experimental tests were performed considering four different mixing ratios and three different cement contents depending on the curing time. Results show that mixing ratio and curing time have considerable effects on electrical and mechanical properties of mixture. Unconfined compressive strength (UCS) decreases as the cement content decreases. However, sufficient grout strength can be obtained with increase of content of inorganic binder. From the results, it is found that the inorganic binder can be used to enhance the mechanical properties of mixture and reduce the cement content. It is expected that data and trends proposed in this study can be used as reference in predicting grouting quality in the field.

Keywords: damping ratio, electrical resistivity, ground improvement, inorganic binder, longitudinal wave velocity, unconfined compression strength

Procedia PDF Downloads 328
32384 Differences Choosing Closed Approach or Open Approach in Rhinoplasty Outcomes

Authors: Alessandro Marano

Abstract:

Aim: The author describes a strategy for choosing between two different rhinoplasty approaches for outcomes treatment. Methods: Series of the case study. There are advantages and disadvantages on both approaches for rhinoplasty. On the side of the open approach, we are be able to better manage the techniques for shaping and restoring nasal structures in rhinoplasty outcomes; on the other side, the closed approach requires more practice and experience to achieve good results. Results: Author’s choice is the closed approach on rhinoplasty outcomes. Anyway, the open approach is most commonly preferred due to superior management and better vision on nasal structures. Conclusions: Both approaches are valid for the treatment of rhinoplasty outcomes, author's preferred approach is closed, with minimally invasive modification focused on restoring outcomes in nasal function and aesthetics.

Keywords: rhinoplasty, aesthetic, face, outcomes

Procedia PDF Downloads 92
32383 Early Prediction of Diseases in a Cow for Cattle Industry

Authors: Ghufran Ahmed, Muhammad Osama Siddiqui, Shahbaz Siddiqui, Rauf Ahmad Shams Malick, Faisal Khan, Mubashir Khan

Abstract:

In this paper, a machine learning-based approach for early prediction of diseases in cows is proposed. Different ML algos are applied to extract useful patterns from the available dataset. Technology has changed today’s world in every aspect of life. Similarly, advanced technologies have been developed in livestock and dairy farming to monitor dairy cows in various aspects. Dairy cattle monitoring is crucial as it plays a significant role in milk production around the globe. Moreover, it has become necessary for farmers to adopt the latest early prediction technologies as the food demand is increasing with population growth. This highlight the importance of state-ofthe-art technologies in analyzing how important technology is in analyzing dairy cows’ activities. It is not easy to predict the activities of a large number of cows on the farm, so, the system has made it very convenient for the farmers., as it provides all the solutions under one roof. The cattle industry’s productivity is boosted as the early diagnosis of any disease on a cattle farm is detected and hence it is treated early. It is done on behalf of the machine learning output received. The learning models are already set which interpret the data collected in a centralized system. Basically, we will run different algorithms on behalf of the data set received to analyze milk quality, and track cows’ health, location, and safety. This deep learning algorithm draws patterns from the data, which makes it easier for farmers to study any animal’s behavioral changes. With the emergence of machine learning algorithms and the Internet of Things, accurate tracking of animals is possible as the rate of error is minimized. As a result, milk productivity is increased. IoT with ML capability has given a new phase to the cattle farming industry by increasing the yield in the most cost-effective and time-saving manner.

Keywords: IoT, machine learning, health care, dairy cows

Procedia PDF Downloads 42
32382 Development and Power Characterization of an IoT Network for Agricultural Imaging Applications

Authors: Jacob Wahl, Jane Zhang

Abstract:

This paper describes the development and characterization of a prototype IoT network for use with agricultural imaging and monitoring applications. The sensor and gateway nodes are designed using the ESP32 SoC with integrated Bluetooth Low Energy 4.2 and Wi-Fi. A development board, the Arducam IoTai ESP32, is used for prototyping, testing, and power measurements. Google’s Firebase is used as the cloud storage site for image data collected by the sensor. The sensor node captures images using the OV2640 2MP camera module and transmits the image data to the gateway via Bluetooth Low Energy. The gateway then uploads the collected images to Firebase via a known nearby Wi-Fi network connection. This image data can then be processed and analyzed by computer vision and machine learning pipelines to assess crop growth or other needs. The sensor node achieves a wireless transmission data throughput of 220kbps while consuming 150mA of current; the sensor sleeps at 162µA. The sensor node device lifetime is estimated to be 682 days on a 6600mAh LiPo battery while acquiring five images per day based on the development board power measurements. This network can be utilized by any application that requires high data rates, low power consumption, short-range communication, and large amounts of data to be transmitted at low-frequency intervals.

Keywords: Bluetooth low energy, ESP32, firebase cloud, IoT, smart farming

Procedia PDF Downloads 124
32381 Optimization the Freeze Drying Conditions of Olive Seeds

Authors: Alev Yüksel Aydar, Tuncay Yılmaz, Melisa Özçeli̇k, Tuba Aydın, Elif Karabaş

Abstract:

In this study, response surface methodology (RSM) was used to obtain the optimum conditions for the freeze-drying of Gemlik variety olive seeds of to achieve the desired quality characteristics. The Box Behnken Design (BBD) was applied with three-variable and three replications in the center point. The effects of the different drying parameters including initial temperature of olive seed, pressure and time for freezing on the DPPH activity, total phenolic contents, and oleuropein absorbance value of the samples were investigated. Temperature (50 – 82 °C), pressure (0.2-0.5 mbar), time (6-10 hours) were chosen as independent variables. The analysis revealed that, while the temperature of the product prior to lyophilization and the drying time had no statistically significant effect on DPPH activity (p>0.05), the pressure was more important than the other two variables , and the quadratic effect of pressure had a significant effect on DPPH activity (p<0.05). The R2 and Adj-R2 values of the DPPH activity model were calculated to be 0.8962 and 0.7045, respectively.

Keywords: olive seed, gemlik variety, DPPH, phenolics, optimization

Procedia PDF Downloads 70
32380 Groupthink: The Dark Side of Team Cohesion

Authors: Farhad Eizakshiri

Abstract:

The potential for groupthink to explain the issues contributing to deterioration of decision-making ability within the unitary team and so to cause poor outcomes attracted a great deal of attention from a variety of disciplines, including psychology, social and organizational studies, political science, and others. Yet what remains unclear is how and why the team members’ strivings for unanimity and cohesion override their motivation to realistically appraise alternative courses of action. In this paper, the findings of a sequential explanatory mixed-methods research containing an experiment with thirty groups of three persons each and interviews with all experimental groups to investigate this issue is reported. The experiment sought to examine how individuals aggregate their views in order to reach a consensual group decision concerning the completion time of a task. The results indicated that groups made better estimates when they had no interaction between members in comparison with the situation that groups collectively agreed on time estimates. To understand the reasons, the qualitative data and informal observations collected during the task were analyzed through conversation analysis, thus leading to four reasons that caused teams to neglect divergent viewpoints and reduce the number of ideas being considered. Reasons found were the concurrence-seeking tendency, pressure on dissenters, self-censorship, and the illusion of invulnerability. It is suggested that understanding the dynamics behind the aforementioned reasons of groupthink will help project teams to avoid making premature group decisions by enhancing careful evaluation of available information and analysis of available decision alternatives and choices.

Keywords: groupthink, group decision, cohesiveness, project teams, mixed-methods research

Procedia PDF Downloads 383
32379 Model Development for Real-Time Human Sitting Posture Detection Using a Camera

Authors: Jheanel E. Estrada, Larry A. Vea

Abstract:

This study developed model to detect proper/improper sitting posture using the built in web camera which detects the upper body points’ location and distances (chin, manubrium and acromion process). It also established relationships of human body frames and proper sitting posture. The models were developed by training some well-known classifiers such as KNN, SVM, MLP, and Decision Tree using the data collected from 60 students of different body frames. Decision Tree classifier demonstrated the most promising model performance with an accuracy of 95.35% and a kappa of 0.907 for head and shoulder posture. Results also showed that there were relationships between body frame and posture through Body Mass Index.

Keywords: posture, spinal points, gyroscope, image processing, ergonomics

Procedia PDF Downloads 313
32378 Food Insecurity and Quality of Life among the Poor Elderly in South Korea

Authors: Jayoung Cho

Abstract:

Poverty has become a social problem in South Korea, given that seven out of ten elderly experience multidimensional poverty. As quality of life is a major social welfare measure of a society, verifying the major factors affecting the quality of life among the elderly in poverty can be used as baseline data for the promotion of welfare. This study aims to investigate the longitudinal relationships between food insecurity and quality of life among the elderly in poverty. In this study, panel regression analysis using 5-year longitudinal panel data were derived from Korea Welfare Panel Study (KWPS, 2011-2015) were used to identify the research question. A total of 1,327 elderly people aged 65 or older with less than 60% of median income was analyzed. The main results of the study are as follows; first, the level of quality of life of the poor elderly was on average of 5, and repeated the increase and decrease over time. Second, food insecurity and quality of life of the elderly in poverty had a longitudinal causal relationship. Furthermore, the statistical significance of food insecurity was the highest despite controlling for major variables affecting the quality of life among the poor elderly. Therefore, political and practical approaches are strongly suggested and considered regarding the food insecurity for the quality of life among the elderly in poverty. In practical intervention, it is necessary to pay attention to food insecurity when assessing the poor elderly. Also, there is a need to build a new delivery system that incorporates segmented health and nutrition-related services. This study has an academic significance in that it brought out the issue of food insecurity of the poor elderly and confirmed the longitudinal relationship between food insecurity and quality of life.

Keywords: food insecurity, longitudinal panel analysis, poor elderly, quality of life

Procedia PDF Downloads 214
32377 Estimation of Soil Moisture at High Resolution through Integration of Optical and Microwave Remote Sensing and Applications in Drought Analyses

Authors: Donglian Sun, Yu Li, Paul Houser, Xiwu Zhan

Abstract:

California experienced severe drought conditions in the past years. In this study, the drought conditions in California are analyzed using soil moisture anomalies derived from integrated optical and microwave satellite observations along with auxiliary land surface data. Based on the U.S. Drought Monitor (USDM) classifications, three typical drought conditions were selected for the analysis: extreme drought conditions in 2007 and 2013, severe drought conditions in 2004 and 2009, and normal conditions in 2005 and 2006. Drought is defined as negative soil moisture anomaly. To estimate soil moisture at high spatial resolutions, three approaches are explored in this study: the universal triangle model that estimates soil moisture from Normalized Difference Vegetation Index (NDVI) and Land Surface Temperature (LST); the basic model that estimates soil moisture under different conditions with auxiliary data like precipitation, soil texture, topography, and surface types; and the refined model that uses accumulated precipitation and its lagging effects. It is found that the basic model shows better agreements with the USDM classifications than the universal triangle model, while the refined model using precipitation accumulated from the previous summer to current time demonstrated the closest agreements with the USDM patterns.

Keywords: soil moisture, high resolution, regional drought, analysis and monitoring

Procedia PDF Downloads 117
32376 Hidden Hot Spots: Identifying and Understanding the Spatial Distribution of Crime

Authors: Lauren C. Porter, Andrew Curtis, Eric Jefferis, Susanne Mitchell

Abstract:

A wealth of research has been generated examining the variation in crime across neighborhoods. However, there is also a striking degree of crime concentration within neighborhoods. A number of studies show that a small percentage of street segments, intersections, or addresses account for a large portion of crime. Not surprisingly, a focus on these crime hot spots can be an effective strategy for reducing community level crime and related ills, such as health problems. However, research is also limited in an important respect. Studies tend to use official data to identify hot spots, such as 911 calls or calls for service. While the use of call data may be more representative of the actual level and distribution of crime than some other official measures (e.g. arrest data), call data still suffer from the 'dark figure of crime.' That is, there is most certainly a degree of error between crimes that occur versus crimes that are reported to the police. In this study, we present an alternative method of identifying crime hot spots, that does not rely on official data. In doing so, we highlight the potential utility of neighborhood-insiders to identify and understand crime dynamics within geographic spaces. Specifically, we use spatial video and geo-narratives to record the crime insights of 36 police, ex-offenders, and residents of a high crime neighborhood in northeast Ohio. Spatial mentions of crime are mapped to identify participant-identified hot spots, and these are juxtaposed with calls for service (CFS) data. While there are bound to be differences between these two sources of data, we find that one location, in particular, a corner store, emerges as a hot spot for all three groups of participants. Yet it does not emerge when we examine CFS data. A closer examination of the space around this corner store and a qualitative analysis of narrative data reveal important clues as to why this store may indeed be a hot spot, but not generate disproportionate calls to the police. In short, our results suggest that researchers who rely solely on official data to study crime hot spots may risk missing some of the most dangerous places.

Keywords: crime, narrative, video, neighborhood

Procedia PDF Downloads 218
32375 Space Tourism Pricing Model Revolution from Time Independent Model to Time-Space Model

Authors: Kang Lin Peng

Abstract:

Space tourism emerged in 2001 and became famous in 2021, following the development of space technology. The space market is twisted because of the excess demand. Space tourism is currently rare and extremely expensive, with biased luxury product pricing, which is the seller’s market that consumers can not bargain with. Spaceship companies such as Virgin Galactic, Blue Origin, and Space X have been charged space tourism prices from 200 thousand to 55 million depending on various heights in space. There should be a reasonable price based on a fair basis. This study aims to derive a spacetime pricing model, which is different from the general pricing model on the earth’s surface. We apply general relativity theory to deduct the mathematical formula for the space tourism pricing model, which covers the traditional time-independent model. In the future, the price of space travel will be different from current flight travel when space travel is measured in lightyear units. The pricing of general commodities mainly considers the general equilibrium of supply and demand. The pricing model considers risks and returns with the dependent time variable as acceptable when commodities are on the earth’s surface, called flat spacetime. Current economic theories based on the independent time scale in the flat spacetime do not consider the curvature of spacetime. Current flight services flying the height of 6, 12, and 19 kilometers are charging with a pricing model that measures time coordinate independently. However, the emergence of space tourism is flying heights above 100 to 550 kilometers that have enlarged the spacetime curvature, which means tourists will escape from a zero curvature on the earth’s surface to the large curvature of space. Different spacetime spans should be considered in the pricing model of space travel to echo general relativity theory. Intuitively, this spacetime commodity needs to consider changing the spacetime curvature from the earth to space. We can assume the value of each spacetime curvature unit corresponding to the gradient change of each Ricci or energy-momentum tensor. Then we know how much to spend by integrating the spacetime from the earth to space. The concept is adding a price p component corresponding to the general relativity theory. The space travel pricing model degenerates into a time-independent model, which becomes a model of traditional commodity pricing. The contribution is that the deriving of the space tourism pricing model will be a breakthrough in philosophical and practical issues for space travel. The results of the space tourism pricing model extend the traditional time-independent flat spacetime mode. The pricing model embedded spacetime as the general relativity theory can better reflect the rationality and accuracy of space travel on the universal scale. The universal scale from independent-time scale to spacetime scale will bring a brand-new pricing concept for space traveling commodities. Fair and efficient spacetime economics will also bring to humans’ travel when we can travel in lightyear units in the future.

Keywords: space tourism, spacetime pricing model, general relativity theory, spacetime curvature

Procedia PDF Downloads 99
32374 Analysis of Fault Tolerance on Grid Computing in Real Time Approach

Authors: Parampal Kaur, Deepak Aggarwal

Abstract:

In the computational Grid, fault tolerance is an imperative issue to be considered during job scheduling. Due to the widespread use of resources, systems are highly prone to errors and failures. Hence, fault tolerance plays a key role in the grid to avoid the problem of unreliability. Scheduling the task to the appropriate resource is a vital requirement in computational Grid. The fittest resource scheduling algorithm searches for the appropriate resource based on the job requirements, in contrary to the general scheduling algorithms where jobs are scheduled to the resources with best performance factor. The proposed method is to improve the fault tolerance of the fittest resource scheduling algorithm by scheduling the job in coordination with job replication when the resource has low reliability. Based on the reliability index of the resource, the resource is identified as critical. The tasks are scheduled based on the criticality of the resources. Results show that the execution time of the tasks is comparatively reduced with the proposed algorithm using real-time approach rather than a simulator.

Keywords: computational grid, fault tolerance, task replication, job scheduling

Procedia PDF Downloads 417
32373 Power Iteration Clustering Based on Deflation Technique on Large Scale Graphs

Authors: Taysir Soliman

Abstract:

One of the current popular clustering techniques is Spectral Clustering (SC) because of its advantages over conventional approaches such as hierarchical clustering, k-means, etc. and other techniques as well. However, one of the disadvantages of SC is the time consuming process because it requires computing the eigenvectors. In the past to overcome this disadvantage, a number of attempts have been proposed such as the Power Iteration Clustering (PIC) technique, which is one of versions from SC; some of PIC advantages are: 1) its scalability and efficiency, 2) finding one pseudo-eigenvectors instead of computing eigenvectors, and 3) linear combination of the eigenvectors in linear time. However, its worst disadvantage is an inter-class collision problem because it used only one pseudo-eigenvectors which is not enough. Previous researchers developed Deflation-based Power Iteration Clustering (DPIC) to overcome problems of PIC technique on inter-class collision with the same efficiency of PIC. In this paper, we developed Parallel DPIC (PDPIC) to improve the time and memory complexity which is run on apache spark framework using sparse matrix. To test the performance of PDPIC, we compared it to SC, ESCG, ESCALG algorithms on four small graph benchmark datasets and nine large graph benchmark datasets, where PDPIC proved higher accuracy and better time consuming than other compared algorithms.

Keywords: spectral clustering, power iteration clustering, deflation-based power iteration clustering, Apache spark, large graph

Procedia PDF Downloads 171
32372 Performance Evaluation of Sand Casting Manufacturing Plant with WITNESS

Authors: Aniruddha Joshi

Abstract:

This paper discusses a simulation study of automated sand casting production system. Therefore, the first aims of this study is development of automated sand casting process model and analyze this model with a simulation software Witness. Production methodology aims to improve overall productivity through elimination of wastes and that leads to improve quality. Integration of automation with Simulation is beneficial to identify the obstacles in implementation and to take appropriate options to implement successfully. For this integration, there are different Simulation Software’s. To study this integration, with the help of “WITNESS” Simulation Software the model is created. This model is based on literature review. The input parameters are Setup Time, Number of machines, cycle time and output parameter is number of castings, avg, and time and percentage usage of machines. Obtained results are used for Statistical Analysis. This analysis concludes the optimal solution to get maximum output.

Keywords: automated sand casting production system, simulation, WITNESS software, performance evaluation

Procedia PDF Downloads 772
32371 Robson System Analysis in Kyiv Perinatal Centre

Authors: Victoria Bila, Iryna Ventskivska, Oleksandra Zahorodnia

Abstract:

The goal of the study: To study the distribution of patients of the Kiyv Perinatal Center according to the Robson system and compare it with world data. Materials and methods: a comparison of the distribution of patients of Kiyv Perinatal center according to the Robson system for 2 periods - the first quarter of 2019 and 2020. For each group, 3 indicators were analyzed - the share of this group in the overall structure of patients of the Perinatal Center for the reporting period, the frequency of abdominal delivery in this group, as well as the contribution of this group to the total number of abdominal delivery. Obtained data were compared with those of the WHO in the guidelines for the implementation of the Robson system in 2017. Results and its discussion: The distribution of patients of the Perinatal Center into groups in the Robson classification is not much different from that recommended by the author. So, among all women, patients of group 1 dominate; this indicator does not change in dynamics. A slight increase in the share of group 2 (6.7% in 2019 and 9.3% - 2020) was due to an increase in the number of labor induction. At the same time, the number of patients of groups 1 and 2 in the Perinatal Center is greater than in the world population, which is determined by the hospitalization of primiparous women with reproductive losses in the past. The Perinatal Center is distinguished from the world population and the proportion of women of group 5 - it was 5.4%, in the world - 7.6%. The frequency of caesarean section in the Perinatal Center is within limits typical for most countries (20.5-20.8%). Moreover, the dominant groups in the structure of caesarean sections are group 5 (21-23.3%) and group 2 (21.9-22.9%), which are the reserve for reducing the number of abdominal delivery. In group 2, certain results have already been achieved in this matter - the frequency of cesarean section in 2019 here amounted to 67.8%, in the first quarter of 2020 - 51.6%. This happened due to a change in the leading method of induction of labor. Thus, the Robson system is a convenient and affordable tool for assessing the structure of caesarean sections. The analysis showed that, in general, the structure of caesarean sections in the Perinatal Center is close to world data, and the identified deviations have explanations related to the specialization of the Center.

Keywords: cesarian section, Robson system, Kyiv Perinatal Center, labor induction

Procedia PDF Downloads 110
32370 Compressive Strength Evaluation of Underwater Concrete Structures Integrating the Combination of Rebound Hardness and Ultrasonic Pulse Velocity Methods with Artificial Neural Networks

Authors: Seunghee Park, Junkyeong Kim, Eun-Seok Shin, Sang-Hun Han

Abstract:

In this study, two kinds of nondestructive evaluation (NDE) techniques (rebound hardness and ultrasonic pulse velocity methods) are investigated for the effective maintenance of underwater concrete structures. A new methodology to estimate the underwater concrete strengths more effectively, named “artificial neural network (ANN) – based concrete strength estimation with the combination of rebound hardness and ultrasonic pulse velocity methods” is proposed and verified throughout a series of experimental works.

Keywords: underwater concrete, rebound hardness, Schmidt hammer, ultrasonic pulse velocity, ultrasonic sensor, artificial neural networks, ANN

Procedia PDF Downloads 510
32369 Evaluation of Clinical Decision Support System in Electronic Medical Record System: A Case of Malawi National Art Electronic Medical Record System

Authors: Pachawo Bisani, Goodall Nyirenda

Abstract:

The Malawi National Antiretroviral Therapy (NART) Electronic Medical Record (EMR) system was designed and developed with guidance from the Ministry of Health through the Department of HIV and AIDS (DHA) with the aim of supporting the management of HIV patient data and reporting in high prevalence ART clinics. As of 2021, the system has been scaled up to over 206 facilities across the country. The system is integrated with the clinical decision support system (CDSS) to assist healthcare providers in making a decision about an individual patient at a particular point in time. Despite NART EMR undergoing several evaluations and assessments, little has been done to evaluate the clinical decision support system in the NART EMR system. Hence, the study aimed to evaluate the use of CDSS in the NART EMR system in Malawi. The study adopted a mixed-method approach, and data was collected through interviews, observations, and questionnaires. The study has revealed that the CDSS tools were integrated into the ART clinic workflow, making it easy for the user to use it. The study has also revealed challenges in system reliability and information accuracy. Despite the challenges, the study further revealed that the system is effective and efficient, and overall, users are satisfied with the system. The study recommends that the implementers focus more on the logic behind the clinical decision-support intervention in order to address some of the concerns and enhance the accuracy of the information supplied. The study further suggests consulting the system's actual users throughout implementation.

Keywords: clinical decision support system, electronic medical record system, usability, antiretroviral therapy

Procedia PDF Downloads 68
32368 Development of an Intelligent Decision Support System for Smart Viticulture

Authors: C. M. Balaceanu, G. Suciu, C. S. Bosoc, O. Orza, C. Fernandez, Z. Viniczay

Abstract:

The Internet of Things (IoT) represents the best option for smart vineyard applications, even if it is necessary to integrate the technologies required for the development. This article is based on the research and the results obtained in the DISAVIT project. For Smart Agriculture, the project aims to provide a trustworthy, intelligent, integrated vineyard management solution that is based on the IoT. To have interoperability through the use of a multiprotocol technology (being the future connected wireless IoT) it is necessary to adopt an agnostic approach, providing a reliable environment to address cyber security, IoT-based threats and traceability through blockchain-based design, but also creating a concept for long-term implementations (modular, scalable). The ones described above represent the main innovative technical aspects of this project. The DISAVIT project studies and promotes the incorporation of better management tools based on objective data-based decisions, which are necessary for agriculture adapted and more resistant to climate change. It also exploits the opportunities generated by the digital services market for smart agriculture management stakeholders. The project's final result aims to improve decision-making, performance, and viticulturally infrastructure and increase real-time data accuracy and interoperability. Innovative aspects such as end-to-end solutions, adaptability, scalability, security and traceability, place our product in a favorable situation over competitors. None of the solutions in the market meet every one of these requirements by a unique product being innovative.

Keywords: blockchain, IoT, smart agriculture, vineyard

Procedia PDF Downloads 180
32367 The Relationship between Depression, HIV Stigma and Adherence to Antiretroviral Therapy among Adult Patients Living with HIV at a Tertiary Hospital in Durban, South Africa: The Mediating Roles of Self-Efficacy and Social Support

Authors: Muziwandile Luthuli

Abstract:

Although numerous factors predicting adherence to antiretroviral therapy (ART) among people living with HIV/AIDS (PLWHA) have been broadly studied on both regional and global level, up-to-date adherence of patients to ART remains an overarching, dynamic and multifaceted problem that needs to be investigated over time and across various contexts. There is a rarity of empirical data in the literature on interactive mechanisms by which psychosocial factors influence adherence to ART among PLWHA within the South African context. Therefore, this study was designed to investigate the relationship between depression, HIV stigma, and adherence to ART among adult patients living with HIV at a tertiary hospital in Durban, South Africa, and the mediating roles of self-efficacy and social support. The health locus of control theory and the social support theory were the underlying theoretical frameworks for this study. Using a cross-sectional research design, a total of 201 male and female adult patients aged between 18-75 years receiving ART at a tertiary hospital in Durban, KwaZulu-Natal were sampled, using time location sampling (TLS). A self-administered questionnaire was employed to collect the data in this study. Data were analysed through SPSS version 27. Several statistical analyses were conducted in this study, namely univariate statistical analysis, correlational analysis, Pearson’s chi-square analysis, cross-tabulation analysis, binary logistic regression analysis, and mediational analysis. Univariate analysis indicated that the sample mean age was 39.28 years (SD=12.115), while most participants were females 71.0% (n=142), never married 74.2% (n=147), and most were also secondary school educated 48.3% (n=97), as well as unemployed 65.7% (n=132). The prevalence rate of participants who had high adherence to ART was 53.7% (n=108), and 46.3% (n=93) of participants had low adherence to ART. Chi-square analysis revealed that employment status was the only statistically significant socio-demographic influence of adherence to ART in this study (χ2 (3) = 8.745; p < .033). Chi-square analysis showed that there was a statistically significant difference found between depression and adherence to ART (χ2 (4) = 16.140; p < .003), while between HIV stigma and adherence to ART, no statistically significant difference was found (χ2 (1) = .323; p >.570). Binary logistic regression indicated that depression was statistically associated with adherence to ART (OR= .853; 95% CI, .789–.922, P < 001), while the association between self-efficacy and adherence to ART was statistically significant (OR= 1.04; 95% CI, 1.001– 1.078, P < .045) after controlling for the effect of depression. However, the findings showed that the effect of depression on adherence to ART was not significantly mediated by self-efficacy (Sobel test for indirect effect, Z= 1.01, P > 0.31). Binary logistic regression showed that the effect of HIV stigma on adherence to ART was not statistically significant (OR= .980; 95% CI, .937– 1.025, P > .374), but the effect of social support on adherence to ART was statistically significant, only after the effect of HIV stigma was controlled for (OR= 1.017; 95% CI, 1.000– 1.035, P < .046). This study promotes behavioral and social change effected through evidence-based interventions by emphasizing the need for additional research that investigates the interactive mechanisms by which psychosocial factors influence adherence to ART. Depression is a significant predictor of adherence to ART. Thus, to alleviate the psychosocial impact of depression on adherence to ART, effective interventions must be devised, along with special consideration of self-efficacy and social support. Therefore, this study is helpful in informing and effecting change in health policy and healthcare services through its findings

Keywords: ART adherence, depression, HIV/AIDS, PLWHA

Procedia PDF Downloads 164
32366 Microarray Data Visualization and Preprocessing Using R and Bioconductor

Authors: Ruchi Yadav, Shivani Pandey, Prachi Srivastava

Abstract:

Microarrays provide a rich source of data on the molecular working of cells. Each microarray reports on the abundance of tens of thousands of mRNAs. Virtually every human disease is being studied using microarrays with the hope of finding the molecular mechanisms of disease. Bioinformatics analysis plays an important part of processing the information embedded in large-scale expression profiling studies and for laying the foundation for biological interpretation. A basic, yet challenging task in the analysis of microarray gene expression data is the identification of changes in gene expression that are associated with particular biological conditions. Careful statistical design and analysis are essential to improve the efficiency and reliability of microarray experiments throughout the data acquisition and analysis process. One of the most popular platforms for microarray analysis is Bioconductor, an open source and open development software project based on the R programming language. This paper describes specific procedures for conducting quality assessment, visualization and preprocessing of Affymetrix Gene Chip and also details the different bioconductor packages used to analyze affymetrix microarray data and describe the analysis and outcome of each plots.

Keywords: microarray analysis, R language, affymetrix visualization, bioconductor

Procedia PDF Downloads 458
32365 Water Body Detection and Estimation from Landsat Satellite Images Using Deep Learning

Authors: M. Devaki, K. B. Jayanthi

Abstract:

The identification of water bodies from satellite images has recently received a great deal of attention. Different methods have been developed to distinguish water bodies from various satellite images that vary in terms of time and space. Urban water identification issues body manifests in numerous applications with a great deal of certainty. There has been a sharp rise in the usage of satellite images to map natural resources, including urban water bodies and forests, during the past several years. This is because water and forest resources depend on each other so heavily that ongoing monitoring of both is essential to their sustainable management. The relevant elements from satellite pictures have been chosen using a variety of techniques, including machine learning. Then, a convolution neural network (CNN) architecture is created that can identify a superpixel as either one of two classes, one that includes water or doesn't from input data in a complex metropolitan scene. The deep learning technique, CNN, has advanced tremendously in a variety of visual-related tasks. CNN can improve classification performance by reducing the spectral-spatial regularities of the input data and extracting deep features hierarchically from raw pictures. Calculate the water body using the satellite image's resolution. Experimental results demonstrate that the suggested method outperformed conventional approaches in terms of water extraction accuracy from remote-sensing images, with an average overall accuracy of 97%.

Keywords: water body, Deep learning, satellite images, convolution neural network

Procedia PDF Downloads 68
32364 A Study on the Accelerated Life Cycle Test Method of the Motor for Home Appliances by Using Acceleration Factor

Authors: Youn-Sung Kim, Mi-Sung Kim, Jae-Kun Lee

Abstract:

This paper deals with the accelerated life cycle test method of the motor for home appliances that demand high reliability. Life Cycle of parts in home appliances also should be 10 years because life cycle of the home appliances such as washing machine, refrigerator, TV is at least 10 years. In case of washing machine, the life cycle test method of motor is advanced for 3000 cycle test (1cycle = 2hours). However, 3000 cycle test incurs loss for the time and cost. Objectives of this study are to reduce the life cycle test time and the number of test samples, which could be realized by using acceleration factor for the test time and reduction factor for the number of sample.

Keywords: accelerated life cycle test, motor reliability test, motor for washing machine, BLDC motor

Procedia PDF Downloads 612
32363 Thermal Fatigue Behavior of Austenitic Stainless Steels

Authors: Jung-Ho Moon, Tae Kwon Ha

Abstract:

Continually increasing working temperature and growing need for greater efficiency and reliability of automotive exhaust require systematic investigation into the thermal fatigue properties especially of high temperature stainless steels. In this study, thermal fatigue properties of 300 series austenitic stainless steels have been evaluated in the temperature ranges of 200-800°C and 200-900°C. Systematic methods for control of temperatures within the predetermined range and measurement of load applied to specimens as a function of temperature during thermal cycles have been established. Thermal fatigue tests were conducted under fully constrained condition, where both ends of specimens were completely fixed. Load relaxation behavior at the temperatures of thermal cycle was closely related with the thermal fatigue property.

Keywords: austenitic stainless steel, automotive exhaust, thermal fatigue, microstructure, load relaxation

Procedia PDF Downloads 357
32362 Bayesian Analysis of Topp-Leone Generalized Exponential Distribution

Authors: Najrullah Khan, Athar Ali Khan

Abstract:

The Topp-Leone distribution was introduced by Topp- Leone in 1955. In this paper, an attempt has been made to fit Topp-Leone Generalized exponential (TPGE) distribution. A real survival data set is used for illustrations. Implementation is done using R and JAGS and appropriate illustrations are made. R and JAGS codes have been provided to implement censoring mechanism using both optimization and simulation tools. The main aim of this paper is to describe and illustrate the Bayesian modelling approach to the analysis of survival data. Emphasis is placed on the modeling of data and the interpretation of the results. Crucial to this is an understanding of the nature of the incomplete or 'censored' data encountered. Analytic approximation and simulation tools are covered here, but most of the emphasis is on Markov chain based Monte Carlo method including independent Metropolis algorithm, which is currently the most popular technique. For analytic approximation, among various optimization algorithms and trust region method is found to be the best. In this paper, TPGE model is also used to analyze the lifetime data in Bayesian paradigm. Results are evaluated from the above mentioned real survival data set. The analytic approximation and simulation methods are implemented using some software packages. It is clear from our findings that simulation tools provide better results as compared to those obtained by asymptotic approximation.

Keywords: Bayesian Inference, JAGS, Laplace Approximation, LaplacesDemon, posterior, R Software, simulation

Procedia PDF Downloads 507
32361 Adsorption of Lead (II) and Copper (II) Ions onto Marula Nuts Activated Carbon

Authors: Lucky Malise, Hilary Rutto, Tumisang Seodigeng

Abstract:

Heavy metal contamination in waste water is a very serious issue affecting a lot of industrialized countries due to the health and environmental impact of these heavy metals on human life and the ecosystem. Adsorption using activated carbon is the most promising method for the removal of heavy metals from waste water but commercial activated carbon is expensive which gives rise to the need for alternatively activated carbon derived from cheap precursors, agricultural wastes, or byproducts from other processes. In this study activated bio-carbon derived from the carbonaceous material obtained from the pyrolysis of Marula nut shells was chemically activated and used as an adsorbent for the removal of lead (II) and copper (II) ions from aqueous solution. The surface morphology and chemistry of the adsorbent before and after chemical activation with zinc chloride impregnation were studied using SEM and FTIR analysis respectively and the results obtained indicate that chemical activation with zinc chloride improves the surface morphology of the adsorbent and enhances the intensity of the surface oxygen complexes on the surface of the adsorbent. The effect of process parameters such as adsorbent dosage, pH value of the solution, initial metal concentration, contact time, and temperature on the adsorption of lead (II) and copper (II) ions onto Marula nut activated carbon were investigated, and their optimum operating conditions were also determined. The experimental data was fitted to both the Langmuir and Freundlich isotherm models, and the data fitted best on the Freundlich isotherm model for both metal ions. The adsorption kinetics were also evaluated, and the experimental data fitted the pseudo-first order kinetic model better than the pseudo second-order kinetic model. The adsorption thermodynamics were also studied and the results indicate that the adsorption of lead and copper ions is spontaneous and exothermic in nature, feasible, and also involves a dissociative mechanism in the temperature range of 25-45 °C.

Keywords: adsorption, isotherms, kinetics, marula nut shells activated carbon, thermodynamics

Procedia PDF Downloads 253
32360 A Unified Fitting Method for the Set of Unified Constitutive Equations for Modelling Microstructure Evolution in Hot Deformation

Authors: Chi Zhang, Jun Jiang

Abstract:

Constitutive equations are very important in finite element (FE) modeling, and the accuracy of the material constants in the equations have significant effects on the accuracy of the FE models. A wide range of constitutive equations are available; however, fitting the material constants in the constitutive equations could be complex and time-consuming due to the strong non-linearity and relationship between the constants. This work will focus on the development of a set of unified MATLAB programs for fitting the material constants in the constitutive equations efficiently. Users will only need to supply experimental data in the required format and run the program without modifying functions or precisely guessing the initial values, or finding the parameters in previous works and will be able to fit the material constants efficiently.

Keywords: constitutive equations, FE modelling, MATLAB program, non-linear curve fitting

Procedia PDF Downloads 80
32359 Deep-Learning Coupled with Pragmatic Categorization Method to Classify the Urban Environment of the Developing World

Authors: Qianwei Cheng, A. K. M. Mahbubur Rahman, Anis Sarker, Abu Bakar Siddik Nayem, Ovi Paul, Amin Ahsan Ali, M. Ashraful Amin, Ryosuke Shibasaki, Moinul Zaber

Abstract:

Thomas Friedman, in his famous book, argued that the world in this 21st century is flat and will continue to be flatter. This is attributed to rapid globalization and the interdependence of humanity that engendered tremendous in-flow of human migration towards the urban spaces. In order to keep the urban environment sustainable, policy makers need to plan based on extensive analysis of the urban environment. With the advent of high definition satellite images, high resolution data, computational methods such as deep neural network analysis, and hardware capable of high-speed analysis; urban planning is seeing a paradigm shift. Legacy data on urban environments are now being complemented with high-volume, high-frequency data. However, the first step of understanding urban space lies in useful categorization of the space that is usable for data collection, analysis, and visualization. In this paper, we propose a pragmatic categorization method that is readily usable for machine analysis and show applicability of the methodology on a developing world setting. Categorization to plan sustainable urban spaces should encompass the buildings and their surroundings. However, the state-of-the-art is mostly dominated by classification of building structures, building types, etc. and largely represents the developed world. Hence, these methods and models are not sufficient for developing countries such as Bangladesh, where the surrounding environment is crucial for the categorization. Moreover, these categorizations propose small-scale classifications, which give limited information, have poor scalability and are slow to compute in real time. Our proposed method is divided into two steps-categorization and automation. We categorize the urban area in terms of informal and formal spaces and take the surrounding environment into account. 50 km × 50 km Google Earth image of Dhaka, Bangladesh was visually annotated and categorized by an expert and consequently a map was drawn. The categorization is based broadly on two dimensions-the state of urbanization and the architectural form of urban environment. Consequently, the urban space is divided into four categories: 1) highly informal area; 2) moderately informal area; 3) moderately formal area; and 4) highly formal area. In total, sixteen sub-categories were identified. For semantic segmentation and automatic categorization, Google’s DeeplabV3+ model was used. The model uses Atrous convolution operation to analyze different layers of texture and shape. This allows us to enlarge the field of view of the filters to incorporate larger context. Image encompassing 70% of the urban space was used to train the model, and the remaining 30% was used for testing and validation. The model is able to segment with 75% accuracy and 60% Mean Intersection over Union (mIoU). In this paper, we propose a pragmatic categorization method that is readily applicable for automatic use in both developing and developed world context. The method can be augmented for real-time socio-economic comparative analysis among cities. It can be an essential tool for the policy makers to plan future sustainable urban spaces.

Keywords: semantic segmentation, urban environment, deep learning, urban building, classification

Procedia PDF Downloads 161
32358 Corporate Social Responsibility: A Paradigm Shift in the New Indian Companies Act, 2013

Authors: Suvankar Chakraborty

Abstract:

Introduction: Corporate Social Responsibility means the obligations of business to act in a manner which will serve the best interests of the Society. The Companies Act , 2013 for the first time has emphasized on the fact that every company having net worth of rupees five hundred crore or more, or turnover of rupees one thousand crore or more or a net profit of rupees five crore or more during any financial year shall constitute a Corporate Social Responsibility Committee of the Board consisting of three or more directors, out of which at least one director shall be an independent director. In the previous Companies Act, 1956 there was no such compulsion for constituting a Corporate Social Responsibility Committee. Objective: This study examines the changes in the perception of corporate sectors so far as social responsibility is concerned. Methodology: The study is based on secondary data obtained from various websites of different corporate sectors and the Gazette of India related to Companies Act, 1956 and the new Companies Act, 2013. For capturing the perception of the corporate world regarding the provisions of CSR in the new Companies Act, 2013, primary data has been collected through structured questionnaire. Findings: Corporate Social Responsibility can put a company on a strong base of sustainable development and in facing the business risk of foreclosure or winding up. Shouldering social responsibility on a long-term basis can help a company not only in increasing its reputation in the business world but also helps in minimizing Government intervention. . But, there can hardly be any universal rule that the area of social responsibility being wholly and solely dependent on the ethical aspect of the corporate sectors. But having said that it may be asserted that business ethics may be a key driver of CSR activities rather than rule based CSR activities in the years to come.

Keywords: business ethics, corporate social responsibility, companies act, 2013, CSR committee

Procedia PDF Downloads 284