Search results for: hardy cross networks accuracy
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9703

Search results for: hardy cross networks accuracy

5113 Assessment of Smart Mechatronics Application in Agriculture

Authors: Sairoel Amertet, Girma Gebresenbet

Abstract:

Smart mechatronics systems in agriculture can be traced back to the mid-1980s, when research into automated fruit harvesting systems began in Japan, Europe, and the United States. Since then, impressive advances have been made in smart mechatronics systems. Furthermore, smart mechatronics systems are promising areas, and as a result, we were intrigued to learn more about them. Consequently, the purpose of this study was to examine the smart mechatronic systems that have been applied to agricultural areas so far, with inspiration from the smart mechatronic system in other sectors. To get an overview of the current state of the art, benefits and drawbacks of smart mechatronics systems, various approaches were investigated. Moreover, smart mechatronic modules and various networks applied in agriculture processing were examined. Finally, we explored how the data retrieved using the one-way analysis of variance related to each other. The result showed that there were strongly related keywords for different journals. With the virtually limited use of sophisticated mechatronics in the agricultural industry and, at the same time, the low production rate, the demand for food security has fallen dramatically. Therefore, the application of smart mechatronics systems in agricultural sectors would be taken into consideration in order to overcome these issues.

Keywords: mechatronics, robotic, robotic system, automation, agriculture mechanism

Procedia PDF Downloads 75
5112 Deployed Confidence: The Testing in Production

Authors: Shreya Asthana

Abstract:

Testers know that the feature they tested on stage is working perfectly in production only after release went live. Sometimes something breaks in production and testers get to know through the end user’s bug raised. The panic mode starts when your staging test results do not reflect current production behavior. And you started doubting your testing skills when finally the user reported a bug to you. Testers can deploy their confidence on release day by testing on production. Once you start doing testing in production, you will see test result accuracy because it will be running on real time data and execution will be a little faster as compared to staging one due to elimination of bad data. Feature flagging, canary releases, and data cleanup can help to achieve this technique of testing. By this paper it will be easier to understand the steps to achieve production testing before making your feature live, and to modify IT company’s testing procedure, so testers can provide the bug free experience to the end users. This study is beneficial because too many people think that testing should be done in staging but not in production and now this is high time to pull out people from their old mindset of testing into a new testing world. At the end of the day, it all just matters if the features are working in production or not.

Keywords: bug free production, new testing mindset, testing strategy, testing approach

Procedia PDF Downloads 69
5111 Evaluation of Alternative Approaches for Additional Damping in Dynamic Calculations of Railway Bridges under High-Speed Traffic

Authors: Lara Bettinelli, Bernhard Glatz, Josef Fink

Abstract:

Planning engineers and researchers use various calculation models with different levels of complexity, calculation efficiency and accuracy in dynamic calculations of railway bridges under high-speed traffic. When choosing a vehicle model to depict the dynamic loading on the bridge structure caused by passing high-speed trains, different goals are pursued: On the one hand, the selected vehicle models should allow the calculation of a bridge’s vibrations as realistic as possible. On the other hand, the computational efficiency and manageability of the models should be preferably high to enable a wide range of applications. The commonly adopted and straightforward vehicle model is the moving load model (MLM), which simplifies the train to a sequence of static axle loads moving at a constant speed over the structure. However, the MLM can significantly overestimate the structure vibrations, especially when resonance events occur. More complex vehicle models, which depict the train as a system of oscillating and coupled masses, can reproduce the interaction dynamics between the vehicle and the bridge superstructure to some extent and enable the calculation of more realistic bridge accelerations. At the same time, such multi-body models require significantly greater processing capacities and precise knowledge of various vehicle properties. The European standards allow for applying the so-called additional damping method when simple load models, such as the MLM, are used in dynamic calculations. An additional damping factor depending on the bridge span, which should take into account the vibration-reducing benefits of the vehicle-bridge interaction, is assigned to the supporting structure in the calculations. However, numerous studies show that when the current standard specifications are applied, the calculation results for the bridge accelerations are in many cases still too high compared to the measured bridge accelerations, while in other cases, they are not on the safe side. A proposal to calculate the additional damping based on extensive dynamic calculations for a parametric field of simply supported bridges with a ballasted track was developed to address this issue. In this contribution, several different approaches to determine the additional damping of the supporting structure considering the vehicle-bridge interaction when using the MLM are compared with one another. Besides the standard specifications, this includes the approach mentioned above and two additional recently published alternative formulations derived from analytical approaches. For a bridge catalogue of 65 existing bridges in Austria in steel, concrete or composite construction, calculations are carried out with the MLM for two different high-speed trains and the different approaches for additional damping. The results are compared with the calculation results obtained by applying a more sophisticated multi-body model of the trains used. The evaluation and comparison of the results allow assessing the benefits of different calculation concepts for the additional damping regarding their accuracy and possible applications. The evaluation shows that by applying one of the recently published redesigned additional damping methods, the calculation results can reflect the influence of the vehicle-bridge interaction on the design-relevant structural accelerations considerably more reliable than by using normative specifications.

Keywords: Additional Damping Method, Bridge Dynamics, High-Speed Railway Traffic, Vehicle-Bridge-Interaction

Procedia PDF Downloads 159
5110 Modeling Food Popularity Dependencies Using Social Media Data

Authors: DEVASHISH KHULBE, MANU PATHAK

Abstract:

The rise in popularity of major social media platforms have enabled people to share photos and textual information about their daily life. One of the popular topics about which information is shared is food. Since a lot of media about food are attributed to particular locations and restaurants, information like spatio-temporal popularity of various cuisines can be analyzed. Tracking the popularity of food types and retail locations across space and time can also be useful for business owners and restaurant investors. In this work, we present an approach using off-the shelf machine learning techniques to identify trends and popularity of cuisine types in an area using geo-tagged data from social media, Google images and Yelp. After adjusting for time, we use the Kernel Density Estimation to get hot spots across the location and model the dependencies among food cuisines popularity using Bayesian Networks. We consider the Manhattan borough of New York City as the location for our analyses but the approach can be used for any area with social media data and information about retail businesses.

Keywords: Web Mining, Geographic Information Systems, Business popularity, Spatial Data Analyses

Procedia PDF Downloads 111
5109 Adaptive Decision Feedback Equalizer Utilizing Fixed-Step Error Signal for Multi-Gbps Serial Links

Authors: Alaa Abdullah Altaee

Abstract:

This paper presents an adaptive decision feedback equalizer (ADFE) for multi-Gbps serial links utilizing a fix-step error signal extracted from cross-points of received data symbols. The extracted signal is generated based on violation of received data symbols with minimum detection requirements at the clock and data recovery (CDR) stage. The iterations of the adaptation process search for the optimum feedback tap coefficients to maximize the data eye-opening and minimize the adaptation convergence time. The effectiveness of the proposed architecture is validated using the simulation results of a serial link designed in an IBM 130 nm 1.2V CMOS technology. The data link with variable channel lengths is analyzed using Spectre from Cadence Design Systems with BSIM4 device models.

Keywords: adaptive DFE, CMOS equalizer, error detection, serial links, timing jitter, wire-line communication

Procedia PDF Downloads 114
5108 The Influence of Disturbances Generated by Arc Furnaces on the Power Quality

Authors: Z. Olczykowski

Abstract:

The paper presents the impact of work on the electric arc furnace. Arc equipment is one of the largest receivers powered by the power system. Electric arc disturbances arising during melting process occurring in these furnaces are the cause of an abrupt change of the passive power of furnaces. Currents drawn by these devices undergo an abrupt change, which in turn cause voltage fluctuations and light flicker. The quantitative evaluation of the voltage fluctuations is now the basic criterion of assessment of an influence of unquiet receiver on the supplying net. The paper presents the method of determination of range of voltage fluctuations and light flicker at parallel operation of arc devices. The results of measurements of voltage fluctuations and light flicker indicators recorded in power supply networks of steelworks were presented, with different number of parallel arc devices. Measurements of energy quality parameters were aimed at verifying the proposed method in practice. It was also analyzed changes in other parameters of electricity: the content of higher harmonics, asymmetry, voltage dips.

Keywords: power quality, arc furnaces, propagation of voltage fluctuations, disturbances

Procedia PDF Downloads 132
5107 Prediction of the Thermal Parameters of a High-Temperature Metallurgical Reactor Using Inverse Heat Transfer

Authors: Mohamed Hafid, Marcel Lacroix

Abstract:

This study presents an inverse analysis for predicting the thermal conductivities and the heat flux of a high-temperature metallurgical reactor simultaneously. Once these thermal parameters are predicted, the time-varying thickness of the protective phase-change bank that covers the inside surface of the brick walls of a metallurgical reactor can be calculated. The enthalpy method is used to solve the melting/solidification process of the protective bank. The inverse model rests on the Levenberg-Marquardt Method (LMM) combined with the Broyden method (BM). A statistical analysis for the thermal parameter estimation is carried out. The effect of the position of the temperature sensors, total number of measurements and measurement noise on the accuracy of inverse predictions is investigated. Recommendations are made concerning the location of temperature sensors.

Keywords: inverse heat transfer, phase change, metallurgical reactor, Levenberg–Marquardt method, Broyden method, bank thickness

Procedia PDF Downloads 329
5106 Bitplanes Image Encryption/Decryption Using Edge Map (SSPCE Method) and Arnold Transform

Authors: Ali A. Ukasha

Abstract:

Data security needed in data transmission, storage, and communication to ensure the security. The single step parallel contour extraction (SSPCE) method is used to create the edge map as a key image from the different Gray level/Binary image. Performing the X-OR operation between the key image and each bit plane of the original image for image pixel values change purpose. The Arnold transform used to changes the locations of image pixels as image scrambling process. Experiments have demonstrated that proposed algorithm can fully encrypt 2D Gary level image and completely reconstructed without any distortion. Also shown that the analyzed algorithm have extremely large security against some attacks like salt & pepper and JPEG compression. Its proof that the Gray level image can be protected with a higher security level. The presented method has easy hardware implementation and suitable for multimedia protection in real time applications such as wireless networks and mobile phone services.

Keywords: SSPCE method, image compression, salt and peppers attacks, bitplanes decomposition, Arnold transform, lossless image encryption

Procedia PDF Downloads 490
5105 A Comprehensive Study of Camouflaged Object Detection Using Deep Learning

Authors: Khalak Bin Khair, Saqib Jahir, Mohammed Ibrahim, Fahad Bin, Debajyoti Karmaker

Abstract:

Object detection is a computer technology that deals with searching through digital images and videos for occurrences of semantic elements of a particular class. It is associated with image processing and computer vision. On top of object detection, we detect camouflage objects within an image using Deep Learning techniques. Deep learning may be a subset of machine learning that's essentially a three-layer neural network Over 6500 images that possess camouflage properties are gathered from various internet sources and divided into 4 categories to compare the result. Those images are labeled and then trained and tested using vgg16 architecture on the jupyter notebook using the TensorFlow platform. The architecture is further customized using Transfer Learning. Methods for transferring information from one or more of these source tasks to increase learning in a related target task are created through transfer learning. The purpose of this transfer of learning methodologies is to aid in the evolution of machine learning to the point where it is as efficient as human learning.

Keywords: deep learning, transfer learning, TensorFlow, camouflage, object detection, architecture, accuracy, model, VGG16

Procedia PDF Downloads 138
5104 Reminiscence Therapy for Alzheimer’s Disease Restrained on Logistic Regression Based Linear Bootstrap Aggregating

Authors: P. S. Jagadeesh Kumar, Mingmin Pan, Xianpei Li, Yanmin Yuan, Tracy Lin Huan

Abstract:

Researchers are doing enchanting research into the inherited features of Alzheimer’s disease and probable consistent therapies. In Alzheimer’s, memories are extinct in reverse order; memories formed lately are more transitory than those from formerly. Reminiscence therapy includes the conversation of past actions, trials and knowledges with another individual or set of people, frequently with the help of perceptible reminders such as photos, household and other acquainted matters from the past, music and collection of tapes. In this manuscript, the competence of reminiscence therapy for Alzheimer’s disease is measured using logistic regression based linear bootstrap aggregating. Logistic regression is used to envisage the experiential features of the patient’s memory through various therapies. Linear bootstrap aggregating shows better stability and accuracy of reminiscence therapy used in statistical classification and regression of memories related to validation therapy, supportive psychotherapy, sensory integration and simulated presence therapy.

Keywords: Alzheimer’s disease, linear bootstrap aggregating, logistic regression, reminiscence therapy

Procedia PDF Downloads 302
5103 Hydrogel Based on Cellulose Acetate Used as Scaffold for Cell Growth

Authors: A. Maria G. Melero, A. M. Senna, J. A. Domingues, M. A. Hausen, E. Aparecida R. Duek, V. R. Botaro

Abstract:

A hydrogel from cellulose acetate cross linked with ethylenediaminetetraacetic dianhydride (HAC-EDTA) was synthesized by our research group, and submitted to characterization and biological tests. Cytocompatibility analysis was performed by confocal microscopy using human adipocyte derived stem cells (ASCs). The FTIR analysis showed characteristic bands of cellulose acetate and hydroxyl groups and the tensile tests evidence that HAC-EDTA present a Young’s modulus of 643.7 MPa. The confocal analysis revealed that there was cell growth at the surface of HAC-EDTA. After one day of culture the cells presented spherical morphology, which may be caused by stress of the sequestration of Ca2+ and Mg2+ ions at the cell medium by HAC-EDTA, as demonstrated by ICP-MS. However, after seven days and 14 days of culture, the cells present fibroblastoid morphology, phenotype expected by this cellular type. The results give efforts to indicate this new material as a potential biomaterial for tissue engineering, in the future in vivo approach.

Keywords: cellulose acetate, hydrogel, biomaterial, cellular growth

Procedia PDF Downloads 190
5102 Annotation Ontology for Semantic Web Development

Authors: Hadeel Al Obaidy, Amani Al Heela

Abstract:

The main purpose of this paper is to examine the concept of semantic web and the role that ontology and semantic annotation plays in the development of semantic web services. The paper focuses on semantic web infrastructure illustrating how ontology and annotation work to provide the learning capabilities for building content semantically. To improve productivity and quality of software, the paper applies approaches, notations and techniques offered by software engineering. It proposes a conceptual model to develop semantic web services for the infrastructure of web information retrieval system of digital libraries. The developed system uses ontology and annotation to build a knowledge based system to define and link the meaning of a web content to retrieve information for users’ queries. The results are more relevant through keywords and ontology rule expansion that will be more accurate to satisfy the requested information. The level of results accuracy would be enhanced since the query semantically analyzed work with the conceptual architecture of the proposed system.

Keywords: semantic web services, software engineering, semantic library, knowledge representation, ontology

Procedia PDF Downloads 168
5101 Violence and Unintentional Injuries among Secondary School Students in Jordan

Authors: Malakeh Zuhdi Malak, Mahmoud Taher Kalaldeh

Abstract:

In Jordan, no available data exists regarding violence and unintentional injuries among secondary school students aged 15-19 years. The purpose of this study was to assess the violence and unintentional injuries among those students, and to compare these two behaviors between male and female students. A descriptive cross-sectional design was carried out on randomly selected eight comprehensive secondary schools (four schools for females and four schools for males) from the public school educational directorate located in Amman. A modified Arabic version of the General School Health Survey questionnaire was used to measure violence and unintentional injuries. A sample of 750 secondary school students was studied. The findings showed that 26.8 % of students had been physically attacked. Overall, 43.3 % of students had been involved in a physical fight and 20.1% of them had been bullied. Overall, 45.3% of students were seriously injured. There was a difference between male and female students regarding to physical attack, physical fight, and serious injuries. In conclusion, it is necessary to develop effective training program in life skills for students that functions to reduce risk-taking behaviors that often leading to violence and unintentional injuries.

Keywords: secondary school students, violence, unintentional injuries, bullying

Procedia PDF Downloads 366
5100 DocPro: A Framework for Processing Semantic and Layout Information in Business Documents

Authors: Ming-Jen Huang, Chun-Fang Huang, Chiching Wei

Abstract:

With the recent advance of the deep neural network, we observe new applications of NLP (natural language processing) and CV (computer vision) powered by deep neural networks for processing business documents. However, creating a real-world document processing system needs to integrate several NLP and CV tasks, rather than treating them separately. There is a need to have a unified approach for processing documents containing textual and graphical elements with rich formats, diverse layout arrangement, and distinct semantics. In this paper, a framework that fulfills this unified approach is presented. The framework includes a representation model definition for holding the information generated by various tasks and specifications defining the coordination between these tasks. The framework is a blueprint for building a system that can process documents with rich formats, styles, and multiple types of elements. The flexible and lightweight design of the framework can help build a system for diverse business scenarios, such as contract monitoring and reviewing.

Keywords: document processing, framework, formal definition, machine learning

Procedia PDF Downloads 208
5099 Exploring Acceptance of Artificial Intelligence Software Solution Amongst Healthcare Personnel: A Case in a Private Medical Centre

Authors: Sandra So, Mohd Roslan Ismail, Safurah Jaafar

Abstract:

With the rapid proliferation of data in healthcare has provided an opportune platform creation of Artificial Intelligence (AI). AI has brought a paradigm shift for healthcare professionals, promising improvement in delivery and quality. This study aims to determine the perception of healthcare personnel on perceived ease of use, perceived usefulness, and subjective norm toward attitude for artificial intelligence acceptance. A cross-sectional single institutional study of employees’ perception of adopting AI in the hospital was conducted. The survey was conducted using a questionnaire adapted from Technology Acceptance Model and a four-point Likert scale was used. There were 96 or 75.5% of the total population responded. This study has shown the significant relationship and the importance of ease of use, perceived usefulness, and subjective norm to the acceptance of AI. In the study results, it concluded that the determining factor to the strong acceptance of AI in their practices is mostly those respondents with the most interaction with the patients and clinical management.

Keywords: artificial intelligence, machine learning, perceived ease of use, perceived usefulness, subjective norm

Procedia PDF Downloads 220
5098 The Incompressible Preference of Turbulence

Authors: Samuel David Dunstan

Abstract:

An elementary observation of a laminar cylindrical Poiseulle-Couette flow profile reveals no distinction in the parabolic streamwise profile from one without a cross-stream flow in whatever reference frame the observation is made. This is because the laminar flow is in solid-body rotation, and there is no intrinsic fluid rotation. Hence the main streamwise Poiseuille flow is unaffected. However, in turbulent (unsteady) cylindrical Poiseuille-Couette flow, the rotational reference frame must be considered, and any observation from an external inertial reference frame can give outright incorrect results. A common misconception in the study of fluid mechanics is the position of the observer does not matter. In this DNS (direct numerical simulation) study, firstly, turbulent flow in a pipe with axial rotation is established. Then in turbulent flow in the concentric pipe, with inner wall rotation, it is shown how the wall streak direction is oriented by the rotational reference frame. The Coriolis force here is not so fictitious after all!

Keywords: concentric pipe, rotational and inertial frames, frame invariance, wall streaks, flow orientation

Procedia PDF Downloads 85
5097 Perceptual Learning with Hand-Eye Coordination as an Effective Tool for Managing Amblyopia: A Prospective Study

Authors: Anandkumar S. Purohit

Abstract:

Introduction: Amblyopia is a serious condition resulting in monocular impairment of vision. Although traditional treatment improves vision, we attempted the results of perceptual learning in this study. Methods: The prospective cohort study included all patients with amblyopia who were subjected to perceptual learning. The presenting data on vision, stereopsis, and contrast sensitivity were documented in a pretested online format, and the pre‑ and post‑treatment information was compared using descriptive, cross‑tabulation, and comparative methods on SPSS 22. Results: The cohort consisted of 47 patients (23 females and 24 males) with a mean age of 14.11 ± 7.13 years. A significant improvement was detected in visual acuity after the PL sessions, and the median follow‑up period was 17 days. Stereopsis improved significantly in all age groups. Conclusion: PL with hand-eye coordination is an effective method for managing amblyopia. This approach can improve vision in all age groups.

Keywords: amblyopia, perceptual learning, hand-eye coordination, visual acuity, stereopsis, contrast sensitivity, ophthalmology

Procedia PDF Downloads 18
5096 Synthesis and Characterization of Biodegradable Elastomeric Polyester Amide for Tissue Engineering Applications

Authors: Abdulrahman T. Essa, Ahmed Aied, Omar Hamid, Felicity R. A. J. Rose, Kevin M. Shakesheff

Abstract:

Biodegradable poly(ester amide)s are promising polymers for biomedical applications such as drug delivery and tissue engineering because of their optimized chemical and physical properties. In this study, we developed a biodegradable polyester amide elastomer poly(serinol sebacate) (PSS) composed of crosslinked networks based on serinol and sebacic acid. The synthesized polymers were characterized to evaluate their chemical structures, mechanical properties, degradation behaviors and in vitro cytocompatibility. Analysis of proton nuclear magnetic resonance and Fourier transform infrared spectroscopy revealed the structure of the polymer. The PSS exhibit excellent solubility in a variety of solvents such as methanol, dimethyl sulfoxide and dimethylformamide. More importantly, the mechanical properties of PSS could be tuned by changing the curing conditions. In addition, the 3T3 fibroblast cells cultured on the PSS demonstrated good cell attachment and high viability.

Keywords: biodegradable, biomaterial, elastomer, mechanical properties, poly(serinol sebacate)

Procedia PDF Downloads 349
5095 Key Factors Influencing Individual Knowledge Capability in KIFs

Authors: Salman Iqbal

Abstract:

Knowledge management (KM) literature has mainly focused on the antecedents of KM. The purpose of this study is to investigate the effect of specific human resource management (HRM) practices on employee knowledge sharing and its outcome as individual knowledge capability. Based on previous literature, a model is proposed for the study and hypotheses are formulated. The cross-sectional dataset comes from a sample of 19 knowledge intensive firms (KIFs). This study has run an item parceling technique followed by Confirmatory Factor Analysis (CFA) on the latent constructs of the research model. Employees’ collaboration and their interpersonal trust can help to improve their knowledge sharing behaviour and knowledge capability within organisations. This study suggests that in future, by using a larger sample, better statistical insight is possible. The findings of this study are beneficial for scholars, policy makers and practitioners. The empirical results of this study are entirely based on employees’ perceptions and make a significant research contribution, given there is a dearth of empirical research focusing on the subcontinent.

Keywords: employees’ collaboration, individual knowledge capability, knowledge sharing, monetary rewards, structural equation modelling

Procedia PDF Downloads 271
5094 Rejuvenate: Face and Body Retouching Using Image Inpainting

Authors: Hossam Abdelrahman, Sama Rostom, Reem Yassein, Yara Mohamed, Salma Salah, Nour Awny

Abstract:

In today’s environment, people are becoming increasingly interested in their appearance. However, they are afraid of their unknown appearance after a plastic surgery or treatment. Accidents, burns and genetic problems such as bowing of body parts of people have a negative impact on their mental health with their appearance and this makes them feel uncomfortable and underestimated. The approach presents a revolutionary deep learning-based image inpainting method that analyses the various picture structures and corrects damaged images. In this study, A model is proposed based on the in-painting of medical images with Stable Diffusion Inpainting method. Reconstructing missing and damaged sections of an image is known as image inpainting is a key progress facilitated by deep neural networks. The system uses the input of the user of an image to indicate a problem, the system will then modify the image and output the fixed image, facilitating for the patient to see the final result.

Keywords: generative adversarial network, large mask inpainting, stable diffusion inpainting, plastic surgery

Procedia PDF Downloads 69
5093 Effects of Dispersion on Peristaltic Flow of a Micropolar Fluid Through a Porous Medium with Wall Effects in the Presence of Slip

Authors: G. Ravi Kiran, G. Radhakrishnamacharya

Abstract:

This paper investigates the effects of slip boundary condition and wall properties on the dispersion of a solute matter in peristaltic flow of an incompressible micropolar fluid through a porous medium. Long wavelength approximation, Taylor's limiting condition and dynamic boundary conditions at the flexible walls are used to obtain the average effective dispersion coefficient in the presence of combined homogeneous and heterogeneous chemical reactions. The effects of various pertinent parameters on the effective dispersion coefficient are discussed. It is observed that peristalsis enhances dispersion. It also increases with micropolar parameter, cross viscosity coefficient, Darcy number, slip parameter and wall parameters. Further, dispersion decreases with homogenous chemical reaction rate and heterogeneous chemical reaction rate.

Keywords: chemical reaction, dispersion, peristalsis, slip condition, wall properties

Procedia PDF Downloads 461
5092 Dynamic Response of Nano Spherical Shell Subjected to Termo-Mechanical Shock Using Nonlocal Elasticity Theory

Authors: J. Ranjbarn, A. Alibeigloo

Abstract:

In this paper, we present an analytical method for analysis of nano-scale spherical shell subjected to thermo-mechanical shocks based on nonlocal elasticity theory. Thermo-mechanical properties of nano shpere is assumed to be temperature dependent. Governing partial differential equation of motion is solved analytically by using Laplace transform for time domain and power series for spacial domain. The results in Laplace domain is transferred to time domain by employing the fast inverse Laplace transform (FLIT) method. Accuracy of present approach is assessed by comparing the the numerical results with the results of published work in literature. Furtheremore, the effects of non-local parameter and wall thickness on the dynamic characteristics of the nano-sphere are studied.

Keywords: nano-scale spherical shell, nonlocal elasticity theory, thermomechanical shock, dynamic response

Procedia PDF Downloads 369
5091 Implementation of IWA-ASM1 Model for Simulating the Wastewater Treatment Plant of Beja by GPS-X 5.1

Authors: Fezzani Boubaker

Abstract:

The modified activated sludge model (ASM1 or Mantis) is a generic structured model and a common platform for dynamic simulation of varieties of aerobic processes for optimization and upgrading of existing plants and for new facilities design. In this study, the modified ASM1 included in the GPS-X software was used to simulate the wastewater treatment plant (WWTP) of Beja treating domestic sewage mixed with baker‘s yeast factory effluent. The results of daily measurements and operating records were used to calibrate the model. A sensitivity and an automatic optimization analysis were conducted to determine the most sensitive and optimal parameters. The results indicated that the ASM1 model could simulate with good accuracy: the COD concentration of effluents from the WWTP of Beja for all months of the year 2012. In addition, it prevents the disruption observed at the output of the plant by injecting the baker‘s yeast factory effluent at high concentrations varied between 20 and 80 g/l.

Keywords: ASM1, activated sludge, baker’s yeast effluent, modelling, simulation, GPS-X 5.1 software

Procedia PDF Downloads 338
5090 Sparse Coding Based Classification of Electrocardiography Signals Using Data-Driven Complete Dictionary Learning

Authors: Fuad Noman, Sh-Hussain Salleh, Chee-Ming Ting, Hadri Hussain, Syed Rasul

Abstract:

In this paper, a data-driven dictionary approach is proposed for the automatic detection and classification of cardiovascular abnormalities. Electrocardiography (ECG) signal is represented by the trained complete dictionaries that contain prototypes or atoms to avoid the limitations of pre-defined dictionaries. The data-driven trained dictionaries simply take the ECG signal as input rather than extracting features to study the set of parameters that yield the most descriptive dictionary. The approach inherently learns the complicated morphological changes in ECG waveform, which is then used to improve the classification. The classification performance was evaluated with ECG data under two different preprocessing environments. In the first category, QT-database is baseline drift corrected with notch filter and it filters the 60 Hz power line noise. In the second category, the data are further filtered using fast moving average smoother. The experimental results on QT database confirm that our proposed algorithm shows a classification accuracy of 92%.

Keywords: electrocardiogram, dictionary learning, sparse coding, classification

Procedia PDF Downloads 377
5089 Deep Learning-Based Classification of 3D CT Scans with Real Clinical Data; Impact of Image format

Authors: Maryam Fallahpoor, Biswajeet Pradhan

Abstract:

Background: Artificial intelligence (AI) serves as a valuable tool in mitigating the scarcity of human resources required for the evaluation and categorization of vast quantities of medical imaging data. When AI operates with optimal precision, it minimizes the demand for human interpretations and, thereby, reduces the burden on radiologists. Among various AI approaches, deep learning (DL) stands out as it obviates the need for feature extraction, a process that can impede classification, especially with intricate datasets. The advent of DL models has ushered in a new era in medical imaging, particularly in the context of COVID-19 detection. Traditional 2D imaging techniques exhibit limitations when applied to volumetric data, such as Computed Tomography (CT) scans. Medical images predominantly exist in one of two formats: neuroimaging informatics technology initiative (NIfTI) and digital imaging and communications in medicine (DICOM). Purpose: This study aims to employ DL for the classification of COVID-19-infected pulmonary patients and normal cases based on 3D CT scans while investigating the impact of image format. Material and Methods: The dataset used for model training and testing consisted of 1245 patients from IranMehr Hospital. All scans shared a matrix size of 512 × 512, although they exhibited varying slice numbers. Consequently, after loading the DICOM CT scans, image resampling and interpolation were performed to standardize the slice count. All images underwent cropping and resampling, resulting in uniform dimensions of 128 × 128 × 60. Resolution uniformity was achieved through resampling to 1 mm × 1 mm × 1 mm, and image intensities were confined to the range of (−1000, 400) Hounsfield units (HU). For classification purposes, positive pulmonary COVID-19 involvement was designated as 1, while normal images were assigned a value of 0. Subsequently, a U-net-based lung segmentation module was applied to obtain 3D segmented lung regions. The pre-processing stage included normalization, zero-centering, and shuffling. Four distinct 3D CNN models (ResNet152, ResNet50, DensNet169, and DensNet201) were employed in this study. Results: The findings revealed that the segmentation technique yielded superior results for DICOM images, which could be attributed to the potential loss of information during the conversion of original DICOM images to NIFTI format. Notably, ResNet152 and ResNet50 exhibited the highest accuracy at 90.0%, and the same models achieved the best F1 score at 87%. ResNet152 also secured the highest Area under the Curve (AUC) at 0.932. Regarding sensitivity and specificity, DensNet201 achieved the highest values at 93% and 96%, respectively. Conclusion: This study underscores the capacity of deep learning to classify COVID-19 pulmonary involvement using real 3D hospital data. The results underscore the significance of employing DICOM format 3D CT images alongside appropriate pre-processing techniques when training DL models for COVID-19 detection. This approach enhances the accuracy and reliability of diagnostic systems for COVID-19 detection.

Keywords: deep learning, COVID-19 detection, NIFTI format, DICOM format

Procedia PDF Downloads 80
5088 Predicting Financial Distress in South Africa

Authors: Nikki Berrange, Gizelle Willows

Abstract:

Business rescue has become increasingly popular since its inclusion in the Companies Act of South Africa in May 2011. The Alternate Exchange (AltX) of the Johannesburg Stock Exchange has experienced a marked increase in the number of companies entering business rescue. This study sampled twenty companies listed on the AltX to determine whether Altman’s Z-score model for emerging markets (ZEM) or Taffler’s Z-score model is a more accurate model in predicting financial distress for small to medium size companies in South Africa. The study was performed over three different time horizons; one, two and three years prior to the event of financial distress, in order to determine how many companies each model predicted would be unlikely to succeed as well as the predictive ability and accuracy of the respective models. The study found that Taffler’s Z-score model had a greater ability at predicting financial distress from all three-time horizons.

Keywords: Altman’s ZEM-score, Altman’s Z-score, AltX, business rescue, Taffler’s Z-score

Procedia PDF Downloads 361
5087 A Review of Encryption Algorithms Used in Cloud Computing

Authors: Derick M. Rakgoale, Topside E. Mathonsi, Vusumuzi Malele

Abstract:

Cloud computing offers distributed online and on-demand computational services from anywhere in the world. Cloud computing services have grown immensely over the past years, especially in the past year due to the Coronavirus pandemic. Cloud computing has changed the working environment and introduced work from work phenomenon, which enabled the adoption of technologies to fulfill the new workings, including cloud services offerings. The increased cloud computing adoption has come with new challenges regarding data privacy and its integrity in the cloud environment. Previously advanced encryption algorithms failed to reduce the memory space required for cloud computing performance, thus increasing the computational cost. This paper reviews the existing encryption algorithms used in cloud computing. In the future, artificial neural networks (ANN) algorithm design will be presented as a security solution to ensure data integrity, confidentiality, privacy, and availability of user data in cloud computing. Moreover, MATLAB will be used to evaluate the proposed solution, and simulation results will be presented.

Keywords: cloud computing, data integrity, confidentiality, privacy, availability

Procedia PDF Downloads 126
5086 Integrated Target Tracking and Control for Automated Car-Following of Truck Platforms

Authors: Fadwa Alaskar, Fang-Chieh Chou, Carlos Flores, Xiao-Yun Lu, Alexandre M. Bayen

Abstract:

This article proposes a perception model for enhancing the accuracy and stability of car-following control of a longitudinally automated truck. We applied a fusion-based tracking algorithm on measurements of a single preceding vehicle needed for car-following control. This algorithm fuses two types of data, radar and LiDAR data, to obtain more accurate and robust longitudinal perception of the subject vehicle in various weather conditions. The filter’s resulting signals are fed to the gap control algorithm at every tracking loop composed by a high-level gap control and lower acceleration tracking system. Several highway tests have been performed with two trucks. The tests show accurate and fast tracking of the target, which impacts on the gap control loop positively. The experiments also show the fulfilment of control design requirements, such as fast speed variations tracking and robust time gap following.

Keywords: object tracking, perception, sensor fusion, adaptive cruise control, cooperative adaptive cruise control

Procedia PDF Downloads 226
5085 A Deletion-Cost Based Fast Compression Algorithm for Linear Vector Data

Authors: Qiuxiao Chen, Yan Hou, Ning Wu

Abstract:

As there are deficiencies of the classic Douglas-Peucker Algorithm (DPA), such as high risks of deleting key nodes by mistake, high complexity, time consumption and relatively slow execution speed, a new Deletion-Cost Based Compression Algorithm (DCA) for linear vector data was proposed. For each curve — the basic element of linear vector data, all the deletion costs of its middle nodes were calculated, and the minimum deletion cost was compared with the pre-defined threshold. If the former was greater than or equal to the latter, all remaining nodes were reserved and the curve’s compression process was finished. Otherwise, the node with the minimal deletion cost was deleted, its two neighbors' deletion costs were updated, and the same loop on the compressed curve was repeated till the termination. By several comparative experiments using different types of linear vector data, the comparison between DPA and DCA was performed from the aspects of compression quality and computing efficiency. Experiment results showed that DCA outperformed DPA in compression accuracy and execution efficiency as well.

Keywords: Douglas-Peucker algorithm, linear vector data, compression, deletion cost

Procedia PDF Downloads 247
5084 Investigation the Effect of Velocity Inlet and Carrying Fluid on the Flow inside Coronary Artery

Authors: Mohammadreza Nezamirad, Nasim Sabetpour, Azadeh Yazdi, Amirmasoud Hamedi

Abstract:

In this study OpenFOAM 4.4.2 was used to investigate flow inside the coronary artery of the heart. This step is the first step of our future project, which is to include conjugate heat transfer of the heart with three main coronary arteries. Three different velocities were used as inlet boundary conditions to see the effect of velocity increase on velocity, pressure, and wall shear of the coronary artery. Also, three different fluids, namely the University of Wisconsin solution, gelatin, and blood was used to investigate the effect of different fluids on flow inside the coronary artery. A code based on Reynolds Stress Navier Stokes (RANS) equations was written and implemented with the real boundary condition that was calculated based on MRI images. In order to improve the accuracy of the current numerical scheme, hex dominant mesh is utilized. When the inlet velocity increases to 0.5 m/s, velocity, wall shear stress, and pressure increase at the narrower parts.

Keywords: CFD, simulation, OpenFOAM, heart

Procedia PDF Downloads 143