Search results for: multiphase machine
1013 Simulation of Pedestrian Service Time at Different Delay Times
Authors: Imran Badshah
Abstract:
Pedestrian service time reflects the performance of the facility, and it’s a key parameter to analyze the capability of facilities provided to serve pedestrians. The level of service of pedestrians (LOS) mainly depends on pedestrian time and safety. The pedestrian time utilized by taking a service is mainly influenced by the number of available services and the time utilized by each pedestrian in receiving a service; that is called a delay time. In this paper, we analyzed the simulated pedestrian service time with different delay times. A simulation is performed in AnyLogic by developing a model that reflects the real scenario of pedestrian services such as ticket machine gates at rail stations, airports, shopping malls, and cinema halls. The simulated pedestrian time is determined for various delay values. The simulated result shows how pedestrian time changes with the delay pattern. The histogram and time plot graph of a model gives the mean, maximum and minimum values of the pedestrian time. This study helps us to check the behavior of pedestrian time at various services such as subway stations, airports, shopping malls, and cinema halls.Keywords: agent-based simulation, anylogic model, pedestrian behavior, time delay
Procedia PDF Downloads 2101012 Multimodal Convolutional Neural Network for Musical Instrument Recognition
Authors: Yagya Raj Pandeya, Joonwhoan Lee
Abstract:
The dynamic behavior of music and video makes it difficult to evaluate musical instrument playing in a video by computer system. Any television or film video clip with music information are rich sources for analyzing musical instruments using modern machine learning technologies. In this research, we integrate the audio and video information sources using convolutional neural network (CNN) and pass network learned features through recurrent neural network (RNN) to preserve the dynamic behaviors of audio and video. We use different pre-trained CNN for music and video feature extraction and then fine tune each model. The music network use 2D convolutional network and video network use 3D convolution (C3D). Finally, we concatenate each music and video feature by preserving the time varying features. The long short term memory (LSTM) network is used for long-term dynamic feature characterization and then use late fusion with generalized mean. The proposed network performs better performance to recognize the musical instrument using audio-video multimodal neural network.Keywords: multimodal, 3D convolution, music-video feature extraction, generalized mean
Procedia PDF Downloads 2151011 Knowledge Reactor: A Contextual Computing Work in Progress for Eldercare
Authors: Scott N. Gerard, Aliza Heching, Susann M. Keohane, Samuel S. Adams
Abstract:
The world-wide population of people over 60 years of age is growing rapidly. The explosion is placing increasingly onerous demands on individual families, multiple industries and entire countries. Current, human-intensive approaches to eldercare are not sustainable, but IoT and AI technologies can help. The Knowledge Reactor (KR) is a contextual, data fusion engine built to address this and other similar problems. It fuses and centralizes IoT and System of Record/Engagement data into a reactive knowledge graph. Cognitive applications and services are constructed with its multiagent architecture. The KR can scale-up and scaledown, because it exploits container-based, horizontally scalable services for graph store (JanusGraph) and pub-sub (Kafka) technologies. While the KR can be applied to many domains that require IoT and AI technologies, this paper describes how the KR specifically supports the challenging domain of cognitive eldercare. Rule- and machine learning-based analytics infer activities of daily living from IoT sensor readings. KR scalability, adaptability, flexibility and usability are demonstrated.Keywords: ambient sensing, AI, artificial intelligence, eldercare, IoT, internet of things, knowledge graph
Procedia PDF Downloads 1751010 Bag of Words Representation Based on Fusing Two Color Local Descriptors and Building Multiple Dictionaries
Authors: Fatma Abdedayem
Abstract:
We propose an extension to the famous method called Bag of words (BOW) which proved a successful role in the field of image categorization. Practically, this method based on representing image with visual words. In this work, firstly, we extract features from images using Spatial Pyramid Representation (SPR) and two dissimilar color descriptors which are opponent-SIFT and transformed-color-SIFT. Secondly, we fuse color local features by joining the two histograms coming from these descriptors. Thirdly, after collecting of all features, we generate multi-dictionaries coming from n random feature subsets that obtained by dividing all features into n random groups. Then, by using these dictionaries separately each image can be represented by n histograms which are lately concatenated horizontally and form the final histogram, that allows to combine Multiple Dictionaries (MDBoW). In the final step, in order to classify image we have applied Support Vector Machine (SVM) on the generated histograms. Experimentally, we have used two dissimilar image datasets in order to test our proposition: Caltech 256 and PASCAL VOC 2007.Keywords: bag of words (BOW), color descriptors, multi-dictionaries, MDBoW
Procedia PDF Downloads 2971009 On the Utility of Bidirectional Transformers in Gene Expression-Based Classification
Authors: Babak Forouraghi
Abstract:
A genetic circuit is a collection of interacting genes and proteins that enable individual cells to implement and perform vital biological functions such as cell division, growth, death, and signaling. In cell engineering, synthetic gene circuits are engineered networks of genes specifically designed to implement functionalities that are not evolved by nature. These engineered networks enable scientists to tackle complex problems such as engineering cells to produce therapeutics within the patient's body, altering T cells to target cancer-related antigens for treatment, improving antibody production using engineered cells, tissue engineering, and production of genetically modified plants and livestock. Construction of computational models to realize genetic circuits is an especially challenging task since it requires the discovery of the flow of genetic information in complex biological systems. Building synthetic biological models is also a time-consuming process with relatively low prediction accuracy for highly complex genetic circuits. The primary goal of this study was to investigate the utility of a pre-trained bidirectional encoder transformer that can accurately predict gene expressions in genetic circuit designs. The main reason behind using transformers is their innate ability (attention mechanism) to take account of the semantic context present in long DNA chains that are heavily dependent on the spatial representation of their constituent genes. Previous approaches to gene circuit design, such as CNN and RNN architectures, are unable to capture semantic dependencies in long contexts, as required in most real-world applications of synthetic biology. For instance, RNN models (LSTM, GRU), although able to learn long-term dependencies, greatly suffer from vanishing gradient and low-efficiency problem when they sequentially process past states and compresses contextual information into a bottleneck with long input sequences. In other words, these architectures are not equipped with the necessary attention mechanisms to follow a long chain of genes with thousands of tokens. To address the above-mentioned limitations, a transformer model was built in this work as a variation to the existing DNA Bidirectional Encoder Representations from Transformers (DNABERT) model. It is shown that the proposed transformer is capable of capturing contextual information from long input sequences with an attention mechanism. In previous works on genetic circuit design, the traditional approaches to classification and regression, such as Random Forrest, Support Vector Machine, and Artificial Neural Networks, were able to achieve reasonably high R2 accuracy levels of 0.95 to 0.97. However, the transformer model utilized in this work, with its attention-based mechanism, was able to achieve a perfect accuracy level of 100%. Further, it is demonstrated that the efficiency of the transformer-based gene expression classifier is not dependent on the presence of large amounts of training examples, which may be difficult to compile in many real-world gene circuit designs.Keywords: machine learning, classification and regression, gene circuit design, bidirectional transformers
Procedia PDF Downloads 631008 Spatiotemporal Neural Network for Video-Based Pose Estimation
Authors: Bin Ji, Kai Xu, Shunyu Yao, Jingjing Liu, Ye Pan
Abstract:
Human pose estimation is a popular research area in computer vision for its important application in human-machine interface. In recent years, 2D human pose estimation based on convolution neural network has got great progress and development. However, in more and more practical applications, people often need to deal with tasks based on video. It’s not far-fetched for us to consider how to combine the spatial and temporal information together to achieve a balance between computing cost and accuracy. To address this issue, this study proposes a new spatiotemporal model, namely Spatiotemporal Net (STNet) to combine both temporal and spatial information more rationally. As a result, the predicted keypoints heatmap is potentially more accurate and spatially more precise. Under the condition of ensuring the recognition accuracy, the algorithm deal with spatiotemporal series in a decoupled way, which greatly reduces the computation of the model, thus reducing the resource consumption. This study demonstrate the effectiveness of our network over the Penn Action Dataset, and the results indicate superior performance of our network over the existing methods.Keywords: convolutional long short-term memory, deep learning, human pose estimation, spatiotemporal series
Procedia PDF Downloads 1501007 Review for Mechanical Tests of Corner Joints on Wooden Windows and Effects to the Stiffness
Authors: Milan Podlena, Stepan Hysek, Jiri Prochazka, Martin Bohm, Jan Bomba
Abstract:
Corner joints are the weakest part of windows, where the members are connected together. Since the dimensions of the windows started become bigger, the strength requirements for corner joints started to increase as well. Therefore, the aim of this study was to test the samples of corner joints of wooden windows. Moisture content of test specimens was stabilized in the climate chamber. After conditioning, test specimens were loaded in the laboratory conditions onto an universal testing machine and the failure load was measured. Data was recalculated by using goniometric, bending moment and stiffness equation to the stiffness coefficients and the bending moments were investigated. The results showed difference that was observed for the mortise with tenon joint and the dowel joint. This difference was explained by a varied adhesive bond area, which is related to the dimensions of dowels (diameter and length) as well. The bending moments and stiffness ware (except of type of corner joint) also affected by type of used adhesive, type of dowels and wood species.Keywords: corner joint, wooden window, bending moment, stiffness
Procedia PDF Downloads 2201006 Investigation of a Hybrid Process: Multipoint Incremental Forming
Authors: Safa Boudhaouia, Mohamed Amen Gahbiche, Eliane Giraud, Wacef Ben Salem, Philippe Dal Santo
Abstract:
Multi-point forming (MPF) and asymmetric incremental forming (ISF) are two flexible processes for sheet metal manufacturing. To take advantages of these two techniques, a hybrid process has been developed: The Multipoint Incremental Forming (MPIF). This process accumulates at once the advantages of each of these last mentioned forming techniques, which makes it a very interesting and particularly an efficient process for single, small, and medium series production. In this paper, an experimental and a numerical investigation of this technique are presented. To highlight the flexibility of this process and its capacity to manufacture standard and complex shapes, several pieces were produced by using MPIF. The forming experiments are performed on a 3-axis CNC machine. Moreover, a numerical model of the MPIF process has been implemented in ABAQUS and the analysis showed a good agreement with experimental results in terms of deformed shape. Furthermore, the use of an elastomeric interpolator allows avoiding classical local defaults like dimples, which are generally caused by the asymmetric contact and also improves the distribution of residual strain. Future works will apply this approach to other alloys used in aeronautic or automotive applications.Keywords: incremental forming, numerical simulation, MPIF, multipoint forming
Procedia PDF Downloads 3581005 Operating System Based Virtualization Models in Cloud Computing
Authors: Dev Ras Pandey, Bharat Mishra, S. K. Tripathi
Abstract:
Cloud computing is ready to transform the structure of businesses and learning through supplying the real-time applications and provide an immediate help for small to medium sized businesses. The ability to run a hypervisor inside a virtual machine is important feature of virtualization and it is called nested virtualization. In today’s growing field of information technology, many of the virtualization models are available, that provide a convenient approach to implement, but decision for a single model selection is difficult. This paper explains the applications of operating system based virtualization in cloud computing with an appropriate/suitable model with their different specifications and user’s requirements. In the present paper, most popular models are selected, and the selection was based on container and hypervisor based virtualization. Selected models were compared with a wide range of user’s requirements as number of CPUs, memory size, nested virtualization supports, live migration and commercial supports, etc. and we identified a most suitable model of virtualization.Keywords: virtualization, OS based virtualization, container based virtualization, hypervisor based virtualization
Procedia PDF Downloads 3311004 Digital Twin Technology: A Solution for Remote Operation and Productivity Improvement During Covid-19 Era and Future
Authors: Muhamad Sahir Bin Ahmad Shatiry, Wan Normeza Wan Zakaria, Mohamad Zaki Hassan
Abstract:
The pandemic Covid19 has significantly impacted the world; the spreading of the Covid19 virus initially from China has dramatically impacted the world's economy. Therefore, the world reacts with establishing the new way or norm in daily life. The rapid rise of the latest technology has been seen by introducing many technologies to ease human life to have a minor contract between humans and avoid spreading the virus Covid19. Digital twin technologies are one of the technologies created before the pandemic Covid19 but slow adoption in the industry. Throughout the Covid19, most of the companies in the world started to explore to use it. The digital twin technology provides the virtual platform to replicate the existing condition or setup for anything such as office, manufacturing line, factories' machine, building, and many more. This study investigates the effect on the economic perspective after the companies use the Digital Twin technology in the industry. To minimize the contact between humans and to have the ability to operate the system digitally remotely. In this study, the explanation of the digital twin technology impacts the world's microeconomic and macroeconomic.Keywords: productivity, artificially intelligence, IoT, digital twin
Procedia PDF Downloads 2051003 A Data Science Pipeline for Algorithmic Trading: A Comparative Study in Applications to Finance and Cryptoeconomics
Authors: Luyao Zhang, Tianyu Wu, Jiayi Li, Carlos-Gustavo Salas-Flores, Saad Lahrichi
Abstract:
Recent advances in AI have made algorithmic trading a central role in finance. However, current research and applications are disconnected information islands. We propose a generally applicable pipeline for designing, programming, and evaluating algorithmic trading of stock and crypto tokens. Moreover, we provide comparative case studies for four conventional algorithms, including moving average crossover, volume-weighted average price, sentiment analysis, and statistical arbitrage. Our study offers a systematic way to program and compare different trading strategies. Moreover, we implement our algorithms by object-oriented programming in Python3, which serves as open-source software for future academic research and applications.Keywords: algorithmic trading, AI for finance, fintech, machine learning, moving average crossover, volume weighted average price, sentiment analysis, statistical arbitrage, pair trading, object-oriented programming, python3
Procedia PDF Downloads 1491002 A Systematic Approach for Identifying Turning Center Capabilities with Vertical Machining Center in Milling Operation
Authors: Joseph Chen, N. Hundal
Abstract:
Conventional machining is a form of subtractive manufacturing, in which a collection of material-working processes utilizing power-driven machine tools are used to remove undesired material to achieve a desired geometry. This paper presents an approach for comparison between turning center and vertical machining center by optimization of cutting parameters at cylindrical workpieces leading to minimum surface roughness by using taguchi methodology. Aluminum alloy was taken to conduct experiments due to its unique high strength-weight ratio that is maintained at elevated temperatures and their exceptional corrosion resistance. During testing, the effects of the cutting parameters on the surface roughness were investigated. Additionally, by using taguchi methodology for each of the cutting parameters (spindle speed, depth of cut, insert diameter, and feed rate) minimum surface roughness for the process of turn-milling was determined according to the cutting parameters. A confirmation experiment demonstrates the effectiveness of taguchi method.Keywords: surface roughness, Taguchi parameter design, turning center, turn-milling operations, vertical machining center
Procedia PDF Downloads 3301001 A Dynamic Neural Network Model for Accurate Detection of Masked Faces
Authors: Oladapo Tolulope Ibitoye
Abstract:
Neural networks have become prominent and widely engaged in algorithmic-based machine learning networks. They are perfect in solving day-to-day issues to a certain extent. Neural networks are computing systems with several interconnected nodes. One of the numerous areas of application of neural networks is object detection. This is a prominent area due to the coronavirus disease pandemic and the post-pandemic phases. Wearing a face mask in public slows the spread of the virus, according to experts’ submission. This calls for the development of a reliable and effective model for detecting face masks on people's faces during compliance checks. The existing neural network models for facemask detection are characterized by their black-box nature and large dataset requirement. The highlighted challenges have compromised the performance of the existing models. The proposed model utilized Faster R-CNN Model on Inception V3 backbone to reduce system complexity and dataset requirement. The model was trained and validated with very few datasets and evaluation results shows an overall accuracy of 96% regardless of skin tone.Keywords: convolutional neural network, face detection, face mask, masked faces
Procedia PDF Downloads 701000 Investigations of Protein Aggregation Using Sequence and Structure Based Features
Authors: M. Michael Gromiha, A. Mary Thangakani, Sandeep Kumar, D. Velmurugan
Abstract:
The main cause of several neurodegenerative diseases such as Alzhemier, Parkinson, and spongiform encephalopathies is formation of amyloid fibrils and plaques in proteins. We have analyzed different sets of proteins and peptides to understand the influence of sequence-based features on protein aggregation process. The comparison of 373 pairs of homologous mesophilic and thermophilic proteins showed that aggregation-prone regions (APRs) are present in both. But, the thermophilic protein monomers show greater ability to ‘stow away’ the APRs in their hydrophobic cores and protect them from solvent exposure. The comparison of amyloid forming and amorphous b-aggregating hexapeptides suggested distinct preferences for specific residues at the six positions as well as all possible combinations of nine residue pairs. The compositions of residues at different positions and residue pairs have been converted into energy potentials and utilized for distinguishing between amyloid forming and amorphous b-aggregating peptides. Our method could correctly identify the amyloid forming peptides at an accuracy of 95-100% in different datasets of peptides.Keywords: aggregation, amyloids, thermophilic proteins, amino acid residues, machine learning techniques
Procedia PDF Downloads 615999 Numerical Investigation of Cavitation on Different Venturi Shapes by Computational Fluid Dynamics
Authors: Sedat Yayla, Mehmet Oruc, Shakhwan Yaseen
Abstract:
Cavitation phenomena might rigorously impair machine parts such as pumps, propellers and impellers or devices as the pressure in the fluid declines under the liquid's saturation pressure. To evaluate the influence of cavitation, in this research two-dimensional computational fluid dynamics (CFD) venturi models with variety of inlet pressure values, throat lengths and vapor fluid contents were applied. In this research three different vapor contents (0%, 5% 10%), four inlet pressures (2, 4, 6, 8 and 10 atm) and two venturi models were employed at different throat lengths ( 5, 10, 15 and 20 mm) for discovering the impact of each parameter on the cavitation number. It is uncovered that there is a positive correlation between pressure inlet and vapor fluid content and cavitation number. Furthermore, it is unveiled that velocity remains almost constant at the inlet pressures of 6, 8,10atm, nevertheless increasing the length of throat results in the substantial escalation in the velocity of the throat at inlet pressures of 2 and 4 atm. Furthermore, velocity and cavitation number were negatively correlated. The results of the cavitation number varied between 0.092 and 0.495 depending upon the velocity values of the throat.Keywords: cavitation number, computational fluid dynamics, mixture of fluid, two-phase flow, velocity of throat
Procedia PDF Downloads 402998 Parametric Study of Ball and Socket Joint for Bio-Mimicking Exoskeleton
Authors: Mukesh Roy, Basant Singh Sikarwar, Ravi Prakash, Priya Ranjan, Ayush Goyal
Abstract:
More than 11% of people suffer from weakness in the bone resulting in inability in walking or climbing stairs or from limited upper body and limb immobility. This motivates a fresh bio-mimicking solution to the design of an exo-skeleton to support human movement in the case of partial or total immobility either due to congenital or genetic factors or due to some accident or due to geratological factors. A deeper insight and detailed understanding is required into the workings of the ball and socket joints. Our research is to mimic ball and socket joints to design snugly fitting exoskeletons. Our objective is to design an exoskeleton which is comfortable and the presence of which is not felt if not in use. Towards this goal, a parametric study is conducted to provide detailed design parameters to fabricate an exoskeleton. This work builds up on real data of the design of the exoskeleton, so that the designed exo-skeleton will be able to provide required strength and support to the subject.Keywords: bio-mimicking, exoskeleton, ball joint, socket joint, artificial limb, patient rehabilitation, joints, human-machine interface, wearable robotics
Procedia PDF Downloads 296997 Unpowered Knee Exoskeleton with Compliant Joints for Stair Descent Assistance
Authors: Pengfan Wu, Xiaoan Chen, Ye He, Tianchi Chen
Abstract:
This paper introduces the design of an unpowered knee exoskeleton to assist human walking by redistributing the moment of the knee joint during stair descent (SD). Considering the knee moment varying with the knee joint angle and the work of the knee joint is all negative, the custom-built spring was used to convert negative work into the potential energy of the spring during flexion, and the obtained energy work as assistance during extension to reduce the consumption of lower limb muscles. The human-machine adaptability problem was left by traditional rigid wearable due to the knee involves sliding and rotating without a fixed-axis rotation, and this paper designed the two-direction grooves to follow the human-knee kinematics, and the wire spring provides a certain resistance to the pin in the groove to prevent extra degrees of freedom. The experiment was performed on a normal stair by healthy young wearing the device on both legs with the surface electromyography recorded. The results show that the quadriceps (knee extensor) were reduced significantly.Keywords: unpowered exoskeleton, stair descent, knee compliant joint, energy redistribution
Procedia PDF Downloads 125996 Cross Project Software Fault Prediction at Design Phase
Authors: Pradeep Singh, Shrish Verma
Abstract:
Software fault prediction models are created by using the source code, processed metrics from the same or previous version of code and related fault data. Some company do not store and keep track of all artifacts which are required for software fault prediction. To construct fault prediction model for such company, the training data from the other projects can be one potential solution. The earlier we predict the fault the less cost it requires to correct. The training data consists of metrics data and related fault data at function/module level. This paper investigates fault predictions at early stage using the cross-project data focusing on the design metrics. In this study, empirical analysis is carried out to validate design metrics for cross project fault prediction. The machine learning techniques used for evaluation is Naïve Bayes. The design phase metrics of other projects can be used as initial guideline for the projects where no previous fault data is available. We analyze seven data sets from NASA Metrics Data Program which offer design as well as code metrics. Overall, the results of cross project is comparable to the within company data learning.Keywords: software metrics, fault prediction, cross project, within project.
Procedia PDF Downloads 344995 Numerical Analysis on the Effect of Abrasive Parameters on Wall Shear Stress and Jet Exit Kinetic Energy
Authors: D. Deepak, N. Yagnesh Sharma
Abstract:
Abrasive Water Jet (AWJ) machining is a relatively new nontraditional machine tool used in machining of fiber reinforced composite. The quality of machined surface depends on jet exit kinetic energy which depends on various operating and material parameters. In the present work the effect abrasive parameters such as its size, concentration and type on jet kinetic energy is investigated using computational fluid dynamics (CFD). In addition, the effect of these parameters on wall shear stress developed inside the nozzle is also investigated. It is found that for the same operating parameters, increase in the abrasive volume fraction (concentration) results in significant decrease in the wall shear stress as well as the jet exit kinetic energy. Increase in the abrasive particle size results in marginal decrease in the jet exit kinetic energy. Numerical simulation also indicates that garnet abrasives produce better jet exit kinetic energy than aluminium oxide and silicon carbide.Keywords: abrasive water jet machining, jet kinetic energy, operating pressure, wall shear stress, Garnet abrasive
Procedia PDF Downloads 379994 Online Learning for Modern Business Models: Theoretical Considerations and Algorithms
Authors: Marian Sorin Ionescu, Olivia Negoita, Cosmin Dobrin
Abstract:
This scientific communication reports and discusses learning models adaptable to modern business problems and models specific to digital concepts and paradigms. In the PAC (probably approximately correct) learning model approach, in which the learning process begins by receiving a batch of learning examples, the set of learning processes is used to acquire a hypothesis, and when the learning process is fully used, this hypothesis is used in the prediction of new operational examples. For complex business models, a lot of models should be introduced and evaluated to estimate the induced results so that the totality of the results are used to develop a predictive rule, which anticipates the choice of new models. In opposition, for online learning-type processes, there is no separation between the learning (training) and predictive phase. Every time a business model is approached, a test example is considered from the beginning until the prediction of the appearance of a model considered correct from the point of view of the business decision. After choosing choice a part of the business model, the label with the logical value "true" is known. Some of the business models are used as examples of learning (training), which helps to improve the prediction mechanisms for future business models.Keywords: machine learning, business models, convex analysis, online learning
Procedia PDF Downloads 142993 Use of Improved Genetic Algorithm in Cloud Computing to Reduce Energy Consumption in Migration of Virtual Machines
Authors: Marziyeh Bahrami, Hamed Pahlevan Hsseini, Behnam Ghamami, Arman Alvanpour, Hamed Ezzati, Amir Salar Sadeghi
Abstract:
One of the ways to increase the efficiency of services in the system of agents and, of course, in the world of cloud computing, is to use virtualization techniques. The aim of this research is to create changes in cloud computing services that will reduce as much as possible the energy consumption related to the migration of virtual machines and, in some way, the energy related to the allocation of resources and reduce the amount of pollution. So far, several methods have been proposed to increase the efficiency of cloud computing services in order to save energy in the cloud environment. The method presented in this article tries to prevent energy consumption by data centers and the subsequent production of carbon and biological pollutants as much as possible by increasing the efficiency of cloud computing services. The results show that the proposed algorithm, using the improvement in virtualization techniques and with the help of a genetic algorithm, improves the efficiency of cloud services in the matter of migrating virtual machines and finally saves consumption. becomes energy.Keywords: consumption reduction, cloud computing, genetic algorithm, live migration, virtual Machine
Procedia PDF Downloads 61992 Using Speech Emotion Recognition as a Longitudinal Biomarker for Alzheimer’s Diseases
Authors: Yishu Gong, Liangliang Yang, Jianyu Zhang, Zhengyu Chen, Sihong He, Xusheng Zhang, Wei Zhang
Abstract:
Alzheimer’s disease (AD) is a progressive neurodegenerative disorder that affects millions of people worldwide and is characterized by cognitive decline and behavioral changes. People living with Alzheimer’s disease often find it hard to complete routine tasks. However, there are limited objective assessments that aim to quantify the difficulty of certain tasks for AD patients compared to non-AD people. In this study, we propose to use speech emotion recognition (SER), especially the frustration level, as a potential biomarker for quantifying the difficulty patients experience when describing a picture. We build an SER model using data from the IEMOCAP dataset and apply the model to the DementiaBank data to detect the AD/non-AD group difference and perform longitudinal analysis to track the AD disease progression. Our results show that the frustration level detected from the SER model can possibly be used as a cost-effective tool for objective tracking of AD progression in addition to the Mini-Mental State Examination (MMSE) score.Keywords: Alzheimer’s disease, speech emotion recognition, longitudinal biomarker, machine learning
Procedia PDF Downloads 114991 Removal of Copper from Wastewaters by Nano-Micro Bubble Ion Flotation
Authors: R. Ahmadi, A. Khodadadi, M. Abdollahi
Abstract:
The removal of copper from a dilute synthetic wastewater (10 mg/L) was studied by ion flotation at laboratory scale. Anionic sodium dodecyl sulfate (SDS) was used as a collector and ethanol as a frother. Different parameters such as pH, collector and frother concentrations, foam height and bubble size distribution (multi bubble ion flotation) were tested to determine the optimum flotation conditions in a Denver type flotation machine. To see into the effect of bubbles size distribution in this paper, a nano-micro bubble generator was designed. The nano and microbubbles that are generated in this way were combined with normal size bubbles generated mechanically. Under the optimum conditions (concentration of SDS: 192mg/l, ethanol: 0.5%v/v, pH value: 4 and froth height=12.5 cm) the best removal obtained for the system Cu/SDS with a dry foam (water recovery: 15.5%) was 85.6%. Coalescence of nano-microbubbles with bubbles of normal size belonging to mechanical flotation cell improved the removal of Cu to a maximum floatability of 92.8% and reduced the water recovery to a 13.1%.The flotation time decreased considerably at 37.5% when the multi bubble ion flotation was used.Keywords: froth flotation, copper, water treatment, optimization, recycling
Procedia PDF Downloads 503990 1D Convolutional Networks to Compute Mel-Spectrogram, Chromagram, and Cochleogram for Audio Networks
Authors: Elias Nemer, Greg Vines
Abstract:
Time-frequency transformation and spectral representations of audio signals are commonly used in various machine learning applications. Training networks on frequency features such as the Mel-Spectrogram or Cochleogram have been proven more effective and convenient than training on-time samples. In practical realizations, these features are created on a different processor and/or pre-computed and stored on disk, requiring additional efforts and making it difficult to experiment with different features. In this paper, we provide a PyTorch framework for creating various spectral features as well as time-frequency transformation and time-domain filter-banks using the built-in trainable conv1d() layer. This allows computing these features on the fly as part of a larger network and enabling easier experimentation with various combinations and parameters. Our work extends the work in the literature developed for that end: First, by adding more of these features and also by allowing the possibility of either starting from initialized kernels or training them from random values. The code is written as a template of classes and scripts that users may integrate into their own PyTorch classes or simply use as is and add more layers for various applications.Keywords: neural networks Mel-Spectrogram, chromagram, cochleogram, discrete Fourrier transform, PyTorch conv1d()
Procedia PDF Downloads 236989 Improving Security in Healthcare Applications Using Federated Learning System With Blockchain Technology
Authors: Aofan Liu, Qianqian Tan, Burra Venkata Durga Kumar
Abstract:
Data security is of the utmost importance in the healthcare area, as sensitive patient information is constantly sent around and analyzed by many different parties. The use of federated learning, which enables data to be evaluated locally on devices rather than being transferred to a central server, has emerged as a potential solution for protecting the privacy of user information. To protect against data breaches and unauthorized access, federated learning alone might not be adequate. In this context, the application of blockchain technology could provide the system extra protection. This study proposes a distributed federated learning system that is built on blockchain technology in order to enhance security in healthcare. This makes it possible for a wide variety of healthcare providers to work together on data analysis without raising concerns about the confidentiality of the data. The technical aspects of the system, including as the design and implementation of distributed learning algorithms, consensus mechanisms, and smart contracts, are also investigated as part of this process. The technique that was offered is a workable alternative that addresses concerns about the safety of healthcare while also fostering collaborative research and the interchange of data.Keywords: data privacy, distributed system, federated learning, machine learning
Procedia PDF Downloads 136988 Towards an Intelligent Ontology Construction Cost Estimation System: Using BIM and New Rules of Measurement Techniques
Authors: F. H. Abanda, B. Kamsu-Foguem, J. H. M. Tah
Abstract:
Construction cost estimation is one of the most important aspects of construction project design. For generations, the process of cost estimating has been manual, time-consuming and error-prone. This has partly led to most cost estimates to be unclear and riddled with inaccuracies that at times lead to over- or under-estimation of construction cost. The development of standard set of measurement rules that are understandable by all those involved in a construction project, have not totally solved the challenges. Emerging Building Information Modelling (BIM) technologies can exploit standard measurement methods to automate cost estimation process and improves accuracies. This requires standard measurement methods to be structured in ontologically and machine readable format; so that BIM software packages can easily read them. Most standard measurement methods are still text-based in textbooks and require manual editing into tables or Spreadsheet during cost estimation. The aim of this study is to explore the development of an ontology based on New Rules of Measurement (NRM) commonly used in the UK for cost estimation. The methodology adopted is Methontology, one of the most widely used ontology engineering methodologies. The challenges in this exploratory study are also reported and recommendations for future studies proposed.Keywords: BIM, construction projects, cost estimation, NRM, ontology
Procedia PDF Downloads 551987 Stress Analysis of Vertebra Using Photoelastic and Finite Element Methods
Authors: Jamal A. Hassan, Ali Q. Abdulrazzaq, Sadiq J. Abass
Abstract:
In this study, both the photoelastic, as well as the finite element methods, are used to study the stress distribution within human vertebra (L4) under forces similar to those that occur during normal life. Two & three dimensional models of vertebra were created by the software AutoCAD. The coordinates obtained were fed into a computer numerical control (CNC) tensile machine to fabricate the models from photoelastic sheets. Completed models were placed in a transmission polariscope and loaded with static force (up to 1500N). Stresses can be quantified and localized by counting the number of fringes. In both methods the Principle stresses were calculated at different regions. The results noticed that the maximum von-mises stress on the area of the extreme superior vertebral body surface and the facet surface with high normal stress (σ) and shear stress (τ). The facets and other posterior elements have a load-bearing function to help support the weight of the upper body and anything that it carries, and are also acted upon by spinal muscle forces. The numerical FE results have been compared with the experimental method using photoelasticity which shows good agreement between experimental and simulation results.Keywords: photoelasticity, stress, load, finite element
Procedia PDF Downloads 286986 Machine Learning-Driven Prediction of Cardiovascular Diseases: A Supervised Approach
Authors: Thota Sai Prakash, B. Yaswanth, Jhade Bhuvaneswar, Marreddy Divakar Reddy, Shyam Ji Gupta
Abstract:
Across the globe, there are a lot of chronic diseases, and heart disease stands out as one of the most perilous. Sadly, many lives are lost to this condition, even though early intervention could prevent such tragedies. However, identifying heart disease in its initial stages is not easy. To address this challenge, we propose an automated system aimed at predicting the presence of heart disease using advanced techniques. By doing so, we hope to empower individuals with the knowledge needed to take proactive measures against this potentially fatal illness. Our approach towards this problem involves meticulous data preprocessing and the development of predictive models utilizing classification algorithms such as Support Vector Machines (SVM), Decision Tree, and Random Forest. We assess the efficiency of every model based on metrics like accuracy, ensuring that we select the most reliable option. Additionally, we conduct thorough data analysis to reveal the importance of different attributes. Among the models considered, Random Forest emerges as the standout performer with an accuracy rate of 96.04% in our study.Keywords: support vector machines, decision tree, random forest
Procedia PDF Downloads 42985 Time-Frequency Feature Extraction Method Based on Micro-Doppler Signature of Ground Moving Targets
Authors: Ke Ren, Huiruo Shi, Linsen Li, Baoshuai Wang, Yu Zhou
Abstract:
Since some discriminative features are required for ground moving targets classification, we propose a new feature extraction method based on micro-Doppler signature. Firstly, the time-frequency analysis of measured data indicates that the time-frequency spectrograms of the three kinds of ground moving targets, i.e., single walking person, two people walking and a moving wheeled vehicle, are discriminative. Then, a three-dimensional time-frequency feature vector is extracted from the time-frequency spectrograms to depict these differences. At last, a Support Vector Machine (SVM) classifier is trained with the proposed three-dimensional feature vector. The classification accuracy to categorize ground moving targets into the three kinds of the measured data is found to be over 96%, which demonstrates the good discriminative ability of the proposed micro-Doppler feature.Keywords: micro-doppler, time-frequency analysis, feature extraction, radar target classification
Procedia PDF Downloads 406984 An Efficient Algorithm for Solving the Transmission Network Expansion Planning Problem Integrating Machine Learning with Mathematical Decomposition
Authors: Pablo Oteiza, Ricardo Alvarez, Mehrdad Pirnia, Fuat Can
Abstract:
To effectively combat climate change, many countries around the world have committed to a decarbonisation of their electricity, along with promoting a large-scale integration of renewable energy sources (RES). While this trend represents a unique opportunity to effectively combat climate change, achieving a sound and cost-efficient energy transition towards low-carbon power systems poses significant challenges for the multi-year Transmission Network Expansion Planning (TNEP) problem. The objective of the multi-year TNEP is to determine the necessary network infrastructure to supply the projected demand in a cost-efficient way, considering the evolution of the new generation mix, including the integration of RES. The rapid integration of large-scale RES increases the variability and uncertainty in the power system operation, which in turn increases short-term flexibility requirements. To meet these requirements, flexible generating technologies such as energy storage systems must be considered within the TNEP as well, along with proper models for capturing the operational challenges of future power systems. As a consequence, TNEP formulations are becoming more complex and difficult to solve, especially for its application in realistic-sized power system models. To meet these challenges, there is an increasing need for developing efficient algorithms capable of solving the TNEP problem with reasonable computational time and resources. In this regard, a promising research area is the use of artificial intelligence (AI) techniques for solving large-scale mixed-integer optimization problems, such as the TNEP. In particular, the use of AI along with mathematical optimization strategies based on decomposition has shown great potential. In this context, this paper presents an efficient algorithm for solving the multi-year TNEP problem. The algorithm combines AI techniques with Column Generation, a traditional decomposition-based mathematical optimization method. One of the challenges of using Column Generation for solving the TNEP problem is that the subproblems are of mixed-integer nature, and therefore solving them requires significant amounts of time and resources. Hence, in this proposal we solve a linearly relaxed version of the subproblems, and trained a binary classifier that determines the value of the binary variables, based on the results obtained from the linearized version. A key feature of the proposal is that we integrate the binary classifier into the optimization algorithm in such a way that the optimality of the solution can be guaranteed. The results of a study case based on the HRP 38-bus test system shows that the binary classifier has an accuracy above 97% for estimating the value of the binary variables. Since the linearly relaxed version of the subproblems can be solved with significantly less time than the integer programming counterpart, the integration of the binary classifier into the Column Generation algorithm allowed us to reduce the computational time required for solving the problem by 50%. The final version of this paper will contain a detailed description of the proposed algorithm, the AI-based binary classifier technique and its integration into the CG algorithm. To demonstrate the capabilities of the proposal, we evaluate the algorithm in case studies with different scenarios, as well as in other power system models.Keywords: integer optimization, machine learning, mathematical decomposition, transmission planning
Procedia PDF Downloads 86