Search results for: slice thickness accuracy
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5250

Search results for: slice thickness accuracy

4170 An Investigation of a Three-Dimensional Constitutive Model of Gas Diffusion Layers in Polymer Electrolyte Membrane Fuel Cells

Authors: Yanqin Chen, Chao Jiang, Chongdu Cho

Abstract:

This research presents the three-dimensional mechanical characteristics of a commercial gas diffusion layer by experiment and simulation results. Although the mechanical performance of gas diffusion layers has attracted much attention, its reliability and accuracy are still a major challenge. With the help of simulation analysis methods, it is beneficial to the gas diffusion layer’s extensive commercial development and the overall stress analysis of proton electrolyte membrane fuel cells during its pre-production design period. Therefore, in this paper, a three-dimensional constitutive model of a commercial gas diffusion layer, including its material stiffness matrix parameters, is developed and coded, in the user-defined material model of a commercial finite element method software for simulation. Then, the model is validated by comparing experimental results as well as simulation outcomes. As a result, both the experimental data and simulation results show a good agreement with each other, with high accuracy.

Keywords: gas diffusion layer, proton electrolyte membrane fuel cell, stiffness matrix, three-dimensional mechanical characteristics, user-defined material model

Procedia PDF Downloads 162
4169 Analyzing Current Transformers Saturation Characteristics for Different Connected Burden Using LabVIEW Data Acquisition Tool

Authors: D. Subedi, S. Pradhan

Abstract:

Current transformers are an integral part of power system because it provides a proportional safe amount of current for protection and measurement applications. However when the power system experiences an abnormal situation leading to huge current flow, then this huge current is proportionally injected to the protection and metering circuit. Since the protection and metering equipment’s are designed to withstand only certain amount of current with respect to time, these high currents pose a risk to man and equipment. Therefore during such instances, the CT saturation characteristics have a huge influence on the safety of both man and equipment and also on the reliability of the protection and metering system. This paper shows the effect of burden on the Accuracy Limiting factor/ Instrument security factor of current transformers and also the change in saturation characteristics of the CT’s. The response of the CT to varying levels of overcurrent at different connected burden will be captured using the data acquisition software LabVIEW. Analysis is done on the real time data gathered using LabVIEW. Variation of current transformer saturation characteristics with changes in burden will be discussed.

Keywords: accuracy limiting factor, burden, current transformer, instrument security factor, saturation characteristics

Procedia PDF Downloads 419
4168 Polarimetric Synthetic Aperture Radar Data Classification Using Support Vector Machine and Mahalanobis Distance

Authors: Najoua El Hajjaji El Idrissi, Necip Gokhan Kasapoglu

Abstract:

Polarimetric Synthetic Aperture Radar-based imaging is a powerful technique used for earth observation and classification of surfaces. Forest evolution has been one of the vital areas of attention for the remote sensing experts. The information about forest areas can be achieved by remote sensing, whether by using active radars or optical instruments. However, due to several weather constraints, such as cloud cover, limited information can be recovered using optical data and for that reason, Polarimetric Synthetic Aperture Radar (PolSAR) is used as a powerful tool for forestry inventory. In this [14paper, we applied support vector machine (SVM) and Mahalanobis distance to the fully polarimetric AIRSAR P, L, C-bands data from the Nezer forest areas, the classification is based in the separation of different tree ages. The classification results were evaluated and the results show that the SVM performs better than the Mahalanobis distance and SVM achieves approximately 75% accuracy. This result proves that SVM classification can be used as a useful method to evaluate fully polarimetric SAR data with sufficient value of accuracy.

Keywords: classification, synthetic aperture radar, SAR polarimetry, support vector machine, mahalanobis distance

Procedia PDF Downloads 135
4167 Modeling the Saltatory Conduction in Myelinated Axons by Order Reduction

Authors: Ruxandra Barbulescu, Daniel Ioan, Gabriela Ciuprina

Abstract:

The saltatory conduction is the way the action potential is transmitted along a myelinated axon. The potential diffuses along the myelinated compartments and it is regenerated in the Ranvier nodes due to the ion channels allowing the flow across the membrane. For an efficient simulation of populations of neurons, it is important to use reduced order models both for myelinated compartments and for Ranvier nodes and to have control over their accuracy and inner parameters. The paper presents a reduced order model of this neural system which allows an efficient simulation method for the saltatory conduction in myelinated axons. This model is obtained by concatenating reduced order linear models of 1D myelinated compartments and nonlinear 0D models of Ranvier nodes. The models for the myelinated compartments are selected from a series of spatially distributed models developed and hierarchized according to their modeling errors. The extracted model described by a nonlinear PDE of hyperbolic type is able to reproduce the saltatory conduction with acceptable accuracy and takes into account the finite propagation speed of potential. Finally, this model is again reduced in order to make it suitable for the inclusion in large-scale neural circuits.

Keywords: action potential, myelinated segments, nonlinear models, Ranvier nodes, reduced order models, saltatory conduction

Procedia PDF Downloads 163
4166 Screen Method of Distributed Cooperative Navigation Factors for Unmanned Aerial Vehicle Swarm

Authors: Can Zhang, Qun Li, Yonglin Lei, Zhi Zhu, Dong Guo

Abstract:

Aiming at the problem of factor screen in distributed collaborative navigation of dense UAV swarm, an efficient distributed collaborative navigation factor screen method is proposed. The method considered the balance between computing load and positioning accuracy. The proposed algorithm utilized the factor graph model to implement a distributed collaborative navigation algorithm. The GNSS information of the UAV itself and the ranging information between the UAVs are used as the positioning factors. In this distributed scheme, a local factor graph is established for each UAV. The positioning factors of nodes with good geometric position distribution and small variance are selected to participate in the navigation calculation. To demonstrate and verify the proposed methods, the simulation and experiments in different scenarios are performed in this research. Simulation results show that the proposed scheme achieves a good balance between the computing load and positioning accuracy in the distributed cooperative navigation calculation of UAV swarm. This proposed algorithm has important theoretical and practical value for both industry and academic areas.

Keywords: screen method, cooperative positioning system, UAV swarm, factor graph, cooperative navigation

Procedia PDF Downloads 83
4165 Trions in Semiconductor Quantum Dot System

Authors: Jayden Leonard, Nguyen Que Huong

Abstract:

In this work, we study the Trion state in a spherical quantum dot of a direct band gap semiconductor with a shell of organic material. The electronic structure of the Trion due to degenerate valence band will be considered. The coupling between the wannier exciton inside the dot and the Frenkel exciton in the shell will make the Trion state become hybrid. The competition between “semiconductor” and “organic” phases of the Trion and the transitions between them depend on Parameters of the system such as the materials, the size of the dot and the thickness of the shell, etc… and could be manipulated using those parameters.

Keywords: trion, exciton, quantum dot, heterostructure

Procedia PDF Downloads 178
4164 Semi-Autonomous Surgical Robot for Pedicle Screw Insertion on ex vivo Bovine Bone: Improved Workflow and Real-Time Process Monitoring

Authors: Robnier Reyes, Andrew J. P. Marques, Joel Ramjist, Chris R. Pasarikovski, Victor X. D. Yang

Abstract:

Over the past three decades, surgical robotic systems have demonstrated their ability to improve surgical outcomes. The LBR Med is a collaborative robotic arm that is meant to work with a surgeon to streamline surgical workflow. It has 7 degrees of freedom and thus can be easily oriented. Position and torque sensors at each joint allow it to maintain a position accuracy of 150 µm with real-time force and torque feedback, making it ideal for complex surgical procedures. Spinal fusion procedures involve the placement of as many as 20 pedicle screws, requiring a great deal of accuracy due to proximity to the spinal canal and surrounding vessels. Any deviation from intended path can lead to major surgical complications. Assistive surgical robotic systems are meant to serve as collaborative devices easing the workload of the surgeon, thereby improving pedicle screw placement by mitigating fatigue related inaccuracies. Moreover, robotic spinal systems have shown marked improvements over conventional freehanded techniques in both screw placement accuracy and fusion quality and have greatly reduced the need for screw revision, intraoperatively and post-operatively. However, current assistive spinal fusion robots, such as the ROSA Spine, are limited in functionality to positioning surgical instruments. While they offer a small degree of improvement in pedicle screw placement accuracy, they do not alleviate surgeon fatigue, nor do they provide real-time force and torque feedback during screw insertion. We propose a semi-autonomous surgical robot workflow for spinal fusion where the surgeon guides the robot to its initial position and orientation, and the robot drives the pedicle screw accurately into the vertebra. Here, we demonstrate feasibility by inserting pedicle screws into ex-vivo bovine rib bone. The robot monitors position, force and torque with respect to predefined values selected by the surgeon to ensure the highest possible spinal fusion quality. The workflow alleviates the strain on the surgeon by having the robot perform the screw placement while the ability to monitor the process in real-time keeps the surgeon in the system loop. The approach we have taken in terms of level autonomy for the robot reflects its ability to safely collaborate with the surgeon in the operating room without external navigation systems.

Keywords: ex vivo bovine bone, pedicle screw, surgical robot, surgical workflow

Procedia PDF Downloads 171
4163 Thermal and Mechanical Finite Element Analysis of a Mineral Casting Machine Frame

Authors: H. Zou, B. Wang

Abstract:

Thermal distortion of the machine tool plays a critical role in its machining accuracy. This study investigates the thermal performance of a high-precision machine frame with future-oriented mineral casting components. A thermo-mechanical finite element model (FEM) was established to evaluate the thermal behavior of the frame under environmental thermal fluctuations. The validity of the presented FEM model was confirmed experimentally by a series of laser interferometer tests. Good agreement between numerical and experimental results demonstrates that the proposed model can accurately predict the thermal deformation of the frame with thermo-mechanical coupling effect. The results also show that keeping the workshop in thermally stable conditions is crucial for improving the machine accuracy of the system with large scale components. The goal of this paper is to investigate the feasibility of innovative mineral casting material applied in high-precision drilling machine and to provide a strategy for machine tool industry seeking a perfect substitute for classic frame materials such as cast iron and granite.

Keywords: thermo-mechanical model, finite element method, laser interferometer, mineral casting frame

Procedia PDF Downloads 303
4162 Petroleum Play Fairway Analysis of the Middle Paleocene Lower Beda Formation, Concession 71, South-Central Sirt Basin, Libya

Authors: Hatem K. Hamed, Mohamed S. Hrouda

Abstract:

The Middle Paleocene Lower Beda Formation was deposited in a ramp system with local shoaling. The main constituent is limestone, with subordinate dolomites and Shales. Reservoir quality is largely influenced by depositional environments and diagenesis processes. Generally the reservoir quality of Lower Beda Formation is low risk on the Inferred Horst and in the Southern Shelf where the Lower Beda formation comprises mainly of calcarenties. In the vicinity of the well GG1 the Lower Beda comprise mainly of argillaceous calcilutites and shale. The reservoir quality gradually improves from high risk to moderate risk towards KK1, LL1 and NN1 wells. The average gross thickness of Lower Beda Formation is about 300 ft. The net thickness varies from about 270 ft. in the E1-71 well to about 30 ft. in the vicinity of GG1-71 well. The net thickest of Lower Beda form a NNW-SSW trend with an average of 250 ft. the change in facies is due to change in the depositional environment, from lagoonal to shoal barrier to open marine affected the reservoir quality. The Upper Cretaceous Sirte Shale is the main source rock. It is developed within the three troughs surrounding the study area. S-Marada Trough to the N- E, Gerad Trough to the N N-W, and Abu Tummym Sub-basin to the S-W of the Inferred Horst. Sirte shale reaches 1000ft, of organically rich section. It has good organic contents over large area 2% to 3%. Hydrocarbon shows were encountered in several wells in Beda Formation this is an indication of vertical and lateral migration of hydrocarbon. The overlying Upper Paleocene Khalifa Formation is a transgressive shale, it is an effective regional top seal. Lithofacies variations in Khalifa Shale, from shales to limestones in the southern shelf in R1-71 well approximately 50-75% of the secession is limestone. About 47 million barrel of hydrocarbon recoverable reserves is expected to be trapped in structural and stratigraphic traps in Beda Formation in the study area.

Keywords: Sirte basin, Beda formation, concession 71, petroleum play fairway analysis

Procedia PDF Downloads 91
4161 Method Validation for Determining Platinum and Palladium in Catalysts Using Inductively Coupled Plasma Optical Emission Spectrometry

Authors: Marin Senila, Oana Cadar, Thorsten Janisch, Patrick Lacroix-Desmazes

Abstract:

The study presents the analytical capability and validation of a method based on microwave-assisted acid digestion for quantitative determination of platinum and palladium in catalysts using inductively coupled plasma optical emission spectrometry (ICP-OES). In order to validate the method, the main figures of merit such as limit of detection and limit of quantification, precision and accuracy were considered and the measurement uncertainty was estimated based on the bottom-up approach according to the international guidelines of ISO/IEC 17025. Limit of detections, estimated from blank signal using 3 s criterion, were 3.0 mg/kg for Pt and respectively 3.6 mg/kg for Pd, while limits of quantification were 9.0 mg/kg for Pt and respectively 10.8 mg/kg for Pd. Precisions, evaluated as standard deviations of repeatability (n=5 parallel samples), were less than 10% for both precious metals. Accuracies of the method, verified by recovery estimation certified reference material NIST SRM 2557 - pulverized recycled monolith, were 99.4 % for Pt and 101% for Pd. The obtained limit of quantifications and accuracy were satisfactory for the intended purpose. The paper offers all the steps necessary to validate the determination method for Pt and Pd in catalysts using inductively coupled plasma optical emission spectrometry.

Keywords: catalyst analysis, ICP-OES, method validation, platinum, palladium

Procedia PDF Downloads 171
4160 Channel Estimation for Orthogonal Frequency Division Multiplexing Systems over Doubly Selective Channels Base on DCS-DCSOMP Algorithm

Authors: Linyu Wang, Furui Huo, Jianhong Xiang

Abstract:

The Doppler shift generated by high-speed movement and multipath effects in the channel are the main reasons for the generation of a time-frequency doubly-selective (DS) channel. There is severe inter-carrier interference (ICI) in the DS channel. Channel estimation for an orthogonal frequency division multiplexing (OFDM) system over a DS channel is very difficult. The simultaneous orthogonal matching pursuit algorithm under distributed compressive sensing theory (DCS-SOMP) has been used in channel estimation for OFDM systems over DS channels. However, the reconstruction accuracy of the DCS-SOMP algorithm is not high enough in the low SNR stage. To solve this problem, in this paper, we propose an improved DCS-SOMP algorithm based on the inner product difference comparison operation (DCS-DCSOMP). The reconstruction accuracy is improved by increasing the number of candidate indexes and designing the comparison conditions of inner product difference. We combine the DCS-DCSOMP algorithm with the basis expansion model (BEM) to reduce the complexity of channel estimation. Simulation results show the effectiveness of the proposed algorithm and its advantages over other algorithms.

Keywords: OFDM, doubly selective, channel estimation, compressed sensing

Procedia PDF Downloads 99
4159 Rapid Monitoring of Earthquake Damages Using Optical and SAR Data

Authors: Saeid Gharechelou, Ryutaro Tateishi

Abstract:

Earthquake is an inevitable catastrophic natural disaster. The damages of buildings and man-made structures, where most of the human activities occur are the major cause of casualties from earthquakes. A comparison of optical and SAR data is presented in the case of Kathmandu valley which was hardly shaken by 2015-Nepal Earthquake. Though many existing researchers have conducted optical data based estimated or suggested combined use of optical and SAR data for improved accuracy, however finding cloud-free optical images when urgently needed are not assured. Therefore, this research is specializd in developing SAR based technique with the target of rapid and accurate geospatial reporting. Should considers that limited time available in post-disaster situation offering quick computation exclusively based on two pairs of pre-seismic and co-seismic single look complex (SLC) images. The InSAR coherence pre-seismic, co-seismic and post-seismic was used to detect the change in damaged area. In addition, the ground truth data from field applied to optical data by random forest classification for detection of damaged area. The ground truth data collected in the field were used to assess the accuracy of supervised classification approach. Though a higher accuracy obtained from the optical data then integration by optical-SAR data. Limitation of cloud-free images when urgently needed for earthquak evevent are and is not assured, thus further research on improving the SAR based damage detection is suggested. Availability of very accurate damage information is expected for channelling the rescue and emergency operations. It is expected that the quick reporting of the post-disaster damage situation quantified by the rapid earthquake assessment should assist in channeling the rescue and emergency operations, and in informing the public about the scale of damage.

Keywords: Sentinel-1A data, Landsat-8, earthquake damage, InSAR, rapid damage monitoring, 2015-Nepal earthquake

Procedia PDF Downloads 173
4158 Mott Transition in the VO2/LSCO Heterojunction

Authors: Yi Hu, Chun-Chi Lin, Shau-En Yeh, Shin Lee

Abstract:

In this study, p–n heterojunctions with La0.5Sr0.5CoO3 (LSCO) and W-doped VO2 thin films were fabricated by the radio frequency (r.f.) magnetron sputtering technique and sol-gel process, respectively. The thickness of VO2 and LSCO thin films are about 40 nm and 400 nm, respectively. Good crystalline match between LSCO and VO2 films was observed from the SEM. The built-in voltages for the junction are about 1.1 V and 2.3 V for the sample in the metallic and insulating state, respectively. The sample can undergo the current induced MIT during applying field when the sample was heated at 40 and 50ºC. This is in agreement with the value obtained from the difference in the work functions of LSCO and VO2. The band structure of the heterojunction was proposed based on the results of analysis.

Keywords: hetrojection, Mott transition, switching , VO2

Procedia PDF Downloads 592
4157 Gradient Boosted Trees on Spark Platform for Supervised Learning in Health Care Big Data

Authors: Gayathri Nagarajan, L. D. Dhinesh Babu

Abstract:

Health care is one of the prominent industries that generate voluminous data thereby finding the need of machine learning techniques with big data solutions for efficient processing and prediction. Missing data, incomplete data, real time streaming data, sensitive data, privacy, heterogeneity are few of the common challenges to be addressed for efficient processing and mining of health care data. In comparison with other applications, accuracy and fast processing are of higher importance for health care applications as they are related to the human life directly. Though there are many machine learning techniques and big data solutions used for efficient processing and prediction in health care data, different techniques and different frameworks are proved to be effective for different applications largely depending on the characteristics of the datasets. In this paper, we present a framework that uses ensemble machine learning technique gradient boosted trees for data classification in health care big data. The framework is built on Spark platform which is fast in comparison with other traditional frameworks. Unlike other works that focus on a single technique, our work presents a comparison of six different machine learning techniques along with gradient boosted trees on datasets of different characteristics. Five benchmark health care datasets are considered for experimentation, and the results of different machine learning techniques are discussed in comparison with gradient boosted trees. The metric chosen for comparison is misclassification error rate and the run time of the algorithms. The goal of this paper is to i) Compare the performance of gradient boosted trees with other machine learning techniques in Spark platform specifically for health care big data and ii) Discuss the results from the experiments conducted on datasets of different characteristics thereby drawing inference and conclusion. The experimental results show that the accuracy is largely dependent on the characteristics of the datasets for other machine learning techniques whereas gradient boosting trees yields reasonably stable results in terms of accuracy without largely depending on the dataset characteristics.

Keywords: big data analytics, ensemble machine learning, gradient boosted trees, Spark platform

Procedia PDF Downloads 243
4156 Computational Study of Flow and Heat Transfer Characteristics of an Incompressible Fluid in a Channel Using Lattice Boltzmann Method

Authors: Imdat Taymaz, Erman Aslan, Kemal Cakir

Abstract:

The Lattice Boltzmann Method (LBM) is performed to computationally investigate the laminar flow and heat transfer of an incompressible fluid with constant material properties in a 2D channel with a built-in triangular prism. Both momentum and energy transport is modelled by the LBM. A uniform lattice structure with a single time relaxation rule is used. Interpolation methods are applied for obtaining a higher flexibility on the computational grid, where the information is transferred from the lattice structure to the computational grid by Lagrange interpolation. The flow is researched on for different Reynolds number, while Prandtl number is keeping constant as a 0.7. The results show how the presence of a triangular prism effects the flow and heat transfer patterns for the steady-state and unsteady-periodic flow regimes. As an evaluation of the accuracy of the developed LBM code, the results are compared with those obtained by a commercial CFD code. It is observed that the present LBM code produces results that have similar accuracy with the well-established CFD code, as an additionally, LBM needs much smaller CPU time for the prediction of the unsteady phonema.

Keywords: laminar forced convection, lbm, triangular prism

Procedia PDF Downloads 378
4155 Walmart Sales Forecasting using Machine Learning in Python

Authors: Niyati Sharma, Om Anand, Sanjeev Kumar Prasad

Abstract:

Assuming future sale value for any of the organizations is one of the major essential characteristics of tactical development. Walmart Sales Forecasting is the finest illustration to work with as a beginner; subsequently, it has the major retail data set. Walmart uses this sales estimate problem for hiring purposes also. We would like to analyzing how the internal and external effects of one of the largest companies in the US can walk out their Weekly Sales in the future. Demand forecasting is the planned prerequisite of products or services in the imminent on the basis of present and previous data and different stages of the market. Since all associations is facing the anonymous future and we do not distinguish in the future good demand. Hence, through exploring former statistics and recent market statistics, we envisage the forthcoming claim and building of individual goods, which are extra challenging in the near future. As a result of this, we are producing the required products in pursuance of the petition of the souk in advance. We will be using several machine learning models to test the exactness and then lastly, train the whole data by Using linear regression and fitting the training data into it. Accuracy is 8.88%. The extra trees regression model gives the best accuracy of 97.15%.

Keywords: random forest algorithm, linear regression algorithm, extra trees classifier, mean absolute error

Procedia PDF Downloads 151
4154 Efficient High Fidelity Signal Reconstruction Based on Level Crossing Sampling

Authors: Negar Riazifar, Nigel G. Stocks

Abstract:

This paper proposes strategies in level crossing (LC) sampling and reconstruction that provide high fidelity signal reconstruction for speech signals; these strategies circumvent the problem of exponentially increasing number of samples as the bit-depth is increased and hence are highly efficient. Specifically, the results indicate that the distribution of the intervals between samples is one of the key factors in the quality of signal reconstruction; including samples with short intervals do not improve the accuracy of the signal reconstruction, whilst samples with large intervals lead to numerical instability. The proposed sampling method, termed reduced conventional level crossing (RCLC) sampling, exploits redundancy between samples to improve the efficiency of the sampling without compromising performance. A reconstruction technique is also proposed that enhances the numerical stability through linear interpolation of samples separated by large intervals. Interpolation is demonstrated to improve the accuracy of the signal reconstruction in addition to the numerical stability. We further demonstrate that the RCLC and interpolation methods can give useful levels of signal recovery even if the average sampling rate is less than the Nyquist rate.

Keywords: level crossing sampling, numerical stability, speech processing, trigonometric polynomial

Procedia PDF Downloads 148
4153 Enhancing Athlete Training using Real Time Pose Estimation with Neural Networks

Authors: Jeh Patel, Chandrahas Paidi, Ahmed Hambaba

Abstract:

Traditional methods for analyzing athlete movement often lack the detail and immediacy required for optimal training. This project aims to address this limitation by developing a Real-time human pose estimation system specifically designed to enhance athlete training across various sports. This system leverages the power of convolutional neural networks (CNNs) to provide a comprehensive and immediate analysis of an athlete’s movement patterns during training sessions. The core architecture utilizes dilated convolutions to capture crucial long-range dependencies within video frames. Combining this with the robust encoder-decoder architecture to further refine pose estimation accuracy. This capability is essential for precise joint localization across the diverse range of athletic poses encountered in different sports. Furthermore, by quantifying movement efficiency, power output, and range of motion, the system provides data-driven insights that can be used to optimize training programs. Pose estimation data analysis can also be used to develop personalized training plans that target specific weaknesses identified in an athlete’s movement patterns. To overcome the limitations posed by outdoor environments, the project employs strategies such as multi-camera configurations or depth sensing techniques. These approaches can enhance pose estimation accuracy in challenging lighting and occlusion scenarios, where pose estimation accuracy in challenging lighting and occlusion scenarios. A dataset is collected From the labs of Martin Luther King at San Jose State University. The system is evaluated through a series of tests that measure its efficiency and accuracy in real-world scenarios. Results indicate a high level of precision in recognizing different poses, substantiating the potential of this technology in practical applications. Challenges such as enhancing the system’s ability to operate in varied environmental conditions and further expanding the dataset for training were identified and discussed. Future work will refine the model’s adaptability and incorporate haptic feedback to enhance the interactivity and richness of the user experience. This project demonstrates the feasibility of an advanced pose detection model and lays the groundwork for future innovations in assistive enhancement technologies.

Keywords: computer vision, deep learning, human pose estimation, U-NET, CNN

Procedia PDF Downloads 60
4152 Spontaneous and Posed Smile Detection: Deep Learning, Traditional Machine Learning, and Human Performance

Authors: Liang Wang, Beste F. Yuksel, David Guy Brizan

Abstract:

A computational model of affect that can distinguish between spontaneous and posed smiles with no errors on a large, popular data set using deep learning techniques is presented in this paper. A Long Short-Term Memory (LSTM) classifier, a type of Recurrent Neural Network, is utilized and compared to human classification. Results showed that while human classification (mean of 0.7133) was above chance, the LSTM model was more accurate than human classification and other comparable state-of-the-art systems. Additionally, a high accuracy rate was maintained with small amounts of training videos (70 instances). The derivation of important features to further understand the success of our computational model were analyzed, and it was inferred that thousands of pairs of points within the eyes and mouth are important throughout all time segments in a smile. This suggests that distinguishing between a posed and spontaneous smile is a complex task, one which may account for the difficulty and lower accuracy of human classification compared to machine learning models.

Keywords: affective computing, affect detection, computer vision, deep learning, human-computer interaction, machine learning, posed smile detection, spontaneous smile detection

Procedia PDF Downloads 128
4151 Attention-Based ResNet for Breast Cancer Classification

Authors: Abebe Mulugojam Negash, Yongbin Yu, Ekong Favour, Bekalu Nigus Dawit, Molla Woretaw Teshome, Aynalem Birtukan Yirga

Abstract:

Breast cancer remains a significant health concern, necessitating advancements in diagnostic methodologies. Addressing this, our paper confronts the notable challenges in breast cancer classification, particularly the imbalance in datasets and the constraints in the accuracy and interpretability of prevailing deep learning approaches. We proposed an attention-based residual neural network (ResNet), which effectively combines the robust features of ResNet with an advanced attention mechanism. Enhanced through strategic data augmentation and positive weight adjustments, this approach specifically targets the issue of data imbalance. The proposed model is tested on the BreakHis dataset and achieved accuracies of 99.00%, 99.04%, 98.67%, and 98.08% in different magnifications (40X, 100X, 200X, and 400X), respectively. We evaluated the performance by using different evaluation metrics such as precision, recall, and F1-Score and made comparisons with other state-of-the-art methods. Our experiments demonstrate that the proposed model outperforms existing approaches, achieving higher accuracy in breast cancer classification.

Keywords: residual neural network, attention mechanism, positive weight, data augmentation

Procedia PDF Downloads 109
4150 Developing High-Definition Flood Inundation Maps (HD-Fims) Using Raster Adjustment with Scenario Profiles (RASPTM)

Authors: Robert Jacobsen

Abstract:

Flood inundation maps (FIMs) are an essential tool in communicating flood threat scenarios to the public as well as in floodplain governance. With an increasing demand for online raster FIMs, the FIM State-of-the-Practice (SOP) is rapidly advancing to meet the dual requirements for high-resolution and high-accuracy—or High-Definition. Importantly, today’s technology also enables the resolution of problems of local—neighborhood-scale—bias errors that often occur in FIMs, even with the use of SOP two-dimensional flood modeling. To facilitate the development of HD-FIMs, a new GIS method--Raster Adjustment with Scenario Profiles, RASPTM—is described for adjusting kernel raster FIMs to match refined scenario profiles. With RASPTM, flood professionals can prepare HD-FIMs for a wide range of scenarios with available kernel rasters, including kernel rasters prepared from vector FIMs. The paper provides detailed procedures for RASPTM, along with an example of applying RASPTM to prepare an HD-FIM for the August 2016 Flood in Louisiana using both an SOP kernel raster and a kernel raster derived from an older vector-based flood insurance rate map. The accuracy of the HD-FIMs achieved with the application of RASPTM to the two kernel rasters is evaluated.

Keywords: hydrology, mapping, high-definition, inundation

Procedia PDF Downloads 83
4149 The Truism of the True and Fair View of Auditor’s Report

Authors: Ofuan James Ilaboya, Okhae J. Ibhadode

Abstract:

The objective of this paper is to theoretically examine the truism of the “true and fair view” in the context of financial reporting. The paper examines the concepts such as true, fair, true and fair view, problems of true and fair view, the origin/history of true and fair view, review of attributes and key issues relating to true and fair view. The methodological approach adopted in this paper is library-based research, focusing on the review of relevant and related extant literature. The findings based on the review of relevant and related literature is suggestive of the fact that the true and fair concept in financial reporting environment is contentious. The study concludes that given the circumstances as chronicled on this paper, it is evident that the truism of the true and fair view of the auditor’s opinion is under serious threat. The way forward may be for the auditor to certify the accuracy and the correctness of the financial statement. While this position being canvassed here may help to substantially bridge the age-long expectation gap, it may as well require an upward review of the current audit fee structure in order to be able to operationalize the onerous task of certifying the accuracy and correctness of the financial statement. This position is contentious and will require a robust consideration which is not within the purview of the present review.

Keywords: fiduciary duty, financial statement, true and correct, true and fair

Procedia PDF Downloads 137
4148 Comparing Deep Architectures for Selecting Optimal Machine Translation

Authors: Despoina Mouratidis, Katia Lida Kermanidis

Abstract:

Machine translation (MT) is a very important task in Natural Language Processing (NLP). MT evaluation is crucial in MT development, as it constitutes the means to assess the success of an MT system, and also helps improve its performance. Several methods have been proposed for the evaluation of (MT) systems. Some of the most popular ones in automatic MT evaluation are score-based, such as the BLEU score, and others are based on lexical similarity or syntactic similarity between the MT outputs and the reference involving higher-level information like part of speech tagging (POS). This paper presents a language-independent machine learning framework for classifying pairwise translations. This framework uses vector representations of two machine-produced translations, one from a statistical machine translation model (SMT) and one from a neural machine translation model (NMT). The vector representations consist of automatically extracted word embeddings and string-like language-independent features. These vector representations used as an input to a multi-layer neural network (NN) that models the similarity between each MT output and the reference, as well as between the two MT outputs. To evaluate the proposed approach, a professional translation and a "ground-truth" annotation are used. The parallel corpora used are English-Greek (EN-GR) and English-Italian (EN-IT), in the educational domain and of informal genres (video lecture subtitles, course forum text, etc.) that are difficult to be reliably translated. They have tested three basic deep learning (DL) architectures to this schema: (i) fully-connected dense, (ii) Convolutional Neural Network (CNN), and (iii) Long Short-Term Memory (LSTM). Experiments show that all tested architectures achieved better results when compared against those of some of the well-known basic approaches, such as Random Forest (RF) and Support Vector Machine (SVM). Better accuracy results are obtained when LSTM layers are used in our schema. In terms of a balance between the results, better accuracy results are obtained when dense layers are used. The reason for this is that the model correctly classifies more sentences of the minority class (SMT). For a more integrated analysis of the accuracy results, a qualitative linguistic analysis is carried out. In this context, problems have been identified about some figures of speech, as the metaphors, or about certain linguistic phenomena, such as per etymology: paronyms. It is quite interesting to find out why all the classifiers led to worse accuracy results in Italian as compared to Greek, taking into account that the linguistic features employed are language independent.

Keywords: machine learning, machine translation evaluation, neural network architecture, pairwise classification

Procedia PDF Downloads 134
4147 Study on The Model of Microscopic Contact Parameters for Grinding M300 Using Elastic Abrasive Tool

Authors: Wu Xiaojun, Liu Ruiping, Yu Xingzhan, Wu Qian

Abstract:

In precision grinding, utilizing the elastic matrix ball has higher processing efficiency and better superficial quality than traditional grinding. The diversity of characteristics which elastic abrasive tool contact with bend surface results in irregular wear abrasion,and abrasive tool machining status get complicated. There is no theoretical interpretation that parameters affect the grinding accuracy.Aiming at corrosion resistance, wear resistance and other characteristics of M 300 material, it is often used as a material on aerospace precision components. The paper carried out grinding and polishing experiments by using material of M 300,to theoretically show the relationship between stress magnitude and grinding efficiency,and predict the optimal combination of grinding parameter for effective grinding, just for the high abrasion resistance features of M 300, analyzing the micro-contact of elastic ball abrasive tool (Whetstone), using mathematical methods deduce the functional relationship between residual peak removal rate and the main parameters which impact the grinding accuracy on the plane case.Thus laying the foundation for the study of elastic abrasive prediction and compensation.

Keywords: flexible abrasive tool, polishing parameters, Hertz theory, removal rate

Procedia PDF Downloads 547
4146 Comparative Study of Radiation Protection in a Hospital Environment

Authors: Lahoucine Zaama, Sanae Douama

Abstract:

In this work, we present the results of a dosimetry study in a Moroccan radiology department . The results are compared with those of a similar study in France. Furthermore, it determines the coefficient of transmission of the lead sheets of different thicknesses depending on the voltage (KV) in a direct exposure. The objective of this study is to choose the thickness of the radiation means to determine the leaf sample sealed with the smallest percentage value radiation transmission, and that in the context of optimization. Thus the comparison among the studies is essential to consider conduct studies and research in this framework to achieve the goal of optimization.

Keywords: radiology, dosimetry, radiation, dose, transmission

Procedia PDF Downloads 498
4145 Mg AZ31B Alloy Processed through ECASD

Authors: P. Fernández-Morales, D. Peláez, C. Isaza, J. M. Meza, E. Mendoza

Abstract:

Mg AZ31B alloy sheets were processed through equal-channel angular sheet drawing (ECASD) process, following the route A and C at room temperature and varying the processing speed. SEM was used to analyze the microstructure. The grain size was refined and presence of twins was observed. Vickers microhardness and tensile testing were carried out to evaluate the mechanical properties, showing in general; a remarkable increase in the first pass and slight increases during subsequent passes and, that the route C produces better uniform properties distribution through the thickness of the samples.

Keywords: ECASD, Mg Alloy, mechanical properties, microstructure

Procedia PDF Downloads 367
4144 Clustering of Association Rules of ISIS & Al-Qaeda Based on Similarity Measures

Authors: Tamanna Goyal, Divya Bansal, Sanjeev Sofat

Abstract:

In world-threatening terrorist attacks, where early detection, distinction, and prediction are effective diagnosis techniques and for functionally accurate and precise analysis of terrorism data, there are so many data mining & statistical approaches to assure accuracy. The computational extraction of derived patterns is a non-trivial task which comprises specific domain discovery by means of sophisticated algorithm design and analysis. This paper proposes an approach for similarity extraction by obtaining the useful attributes from the available datasets of terrorist attacks and then applying feature selection technique based on the statistical impurity measures followed by clustering techniques on the basis of similarity measures. On the basis of degree of participation of attributes in the rules, the associative dependencies between the attacks are analyzed. Consequently, to compute the similarity among the discovered rules, we applied a weighted similarity measure. Finally, the rules are grouped by applying using hierarchical clustering. We have applied it to an open source dataset to determine the usability and efficiency of our technique, and a literature search is also accomplished to support the efficiency and accuracy of our results.

Keywords: association rules, clustering, similarity measure, statistical approaches

Procedia PDF Downloads 323
4143 Monocrystalline Silicon Surface Passivation by Porous Silicon

Authors: Mohamed Ben Rabha

Abstract:

In this paper, we report on the effect of porous silicon (PS) treatment on the surface passivation of monocrystalline silicon (c-Si). PS film with a thickness of 80 nm was deposited by stain etching. It was demonstrated that PS coating is a very interesting solution for surface passivation. The level of surface passivation is determined by techniques based on photoconductance and FTIR. As a results, the effective minority carrier lifetime increase from 2 µs to 7 µs at ∆n=1015 cm-3 and the reflectivity reduce from 28 % to about 7 % after PS coating.

Keywords: porous silicon, effective minority carrier lifetime, reflectivity

Procedia PDF Downloads 448
4142 Stability Indicating Method Development and Validation for Estimation of Antiasthmatic Drug in Combined Dosages Formed by RP-HPLC

Authors: Laxman H. Surwase, Lalit V. Sonawane, Bhagwat N. Poul

Abstract:

A simple stability indicating high performance liquid chromatographic method has been developed for the simultaneous determination of Levosalbutamol Sulphate and Ipratropium Bromide in bulk and pharmaceutical dosage form using reverse phase Zorbax Eclipse Plus C8 column (250mm×4.6mm), with mobile phase phosphate buffer (0.05M KH2PO4): acetonitrile (55:45v/v) pH 3.5 adjusted with ortho-phosphoric acid, the flow rate was 1.0 mL/min and the detection was carried at 212 nm. The retention times of Levosalbutamol Sulphate and Ipratropium Bromide were 2.2007 and 2.6611 min respectively. The correlation coefficient of Levosalbutamol Sulphate and Ipratropium Bromide was found to be 0.997 and 0.998.Calibration plots were linear over the concentration ranges 10-100µg/mL for both Levosalbutamol Sulphate and Ipratropium Bromide. The LOD and LOQ of Levosalbutamol Sulphate were 2.520µg/mL and 7.638µg/mL while for Ipratropium Bromide was 1.201µg/mL and 3.640 µg/mL. The accuracy of the proposed method was determined by recovery studies and found to be 100.15% for Levosalbutamol Sulphate and 100.19% for Ipratropium Bromide respectively. The method was validated for accuracy, linearity, sensitivity, precision, robustness, system suitability. The proposed method could be utilized for routine analysis of Levosalbutamol Sulphate and Ipratropium Bromide in bulk and pharmaceutical capsule dosage form.

Keywords: levosalbutamol sulphate, ipratropium bromide, RP-HPLC, phosphate buffer, acetonitrile

Procedia PDF Downloads 352
4141 Prediction of Live Birth in a Matched Cohort of Elective Single Embryo Transfers

Authors: Mohsen Bahrami, Banafsheh Nikmehr, Yueqiang Song, Anuradha Koduru, Ayse K. Vuruskan, Hongkun Lu, Tamer M. Yalcinkaya

Abstract:

In recent years, we have witnessed an explosion of studies aimed at using a combination of artificial intelligence (AI) and time-lapse imaging data on embryos to improve IVF outcomes. However, despite promising results, no study has used a matched cohort of transferred embryos which only differ in pregnancy outcome, i.e., embryos from a single clinic which are similar in parameters, such as: morphokinetic condition, patient age, and overall clinic and lab performance. Here, we used time-lapse data on embryos with known pregnancy outcomes to see if the rich spatiotemporal information embedded in this data would allow the prediction of the pregnancy outcome regardless of such critical parameters. Methodology—We did a retrospective analysis of time-lapse data from our IVF clinic utilizing Embryoscope 100% of the time for embryo culture to blastocyst stage with known clinical outcomes, including live birth vs nonpregnant (embryos with spontaneous abortion outcomes were excluded). We used time-lapse data from 200 elective single transfer embryos randomly selected from January 2019 to June 2021. Our sample included 100 embryos in each group with no significant difference in patient age (P=0.9550) and morphokinetic scores (P=0.4032). Data from all patients were combined to make a 4th order tensor, and feature extraction were subsequently carried out by a tensor decomposition methodology. The features were then used in a machine learning classifier to classify the two groups. Major Findings—The performance of the model was evaluated using 100 random subsampling cross validation (train (80%) - test (20%)). The prediction accuracy, averaged across 100 permutations, exceeded 80%. We also did a random grouping analysis, in which labels (live birth, nonpregnant) were randomly assigned to embryos, which yielded 50% accuracy. Conclusion—The high accuracy in the main analysis and the low accuracy in random grouping analysis suggest a consistent spatiotemporal pattern which is associated with pregnancy outcomes, regardless of patient age and embryo morphokinetic condition, and beyond already known parameters, such as: early cleavage or early blastulation. Despite small samples size, this ongoing analysis is the first to show the potential of AI methods in capturing the complex morphokinetic changes embedded in embryo time-lapse data, which contribute to successful pregnancy outcomes, regardless of already known parameters. The results on a larger sample size with complementary analysis on prediction of other key outcomes, such as: euploidy and aneuploidy of embryos will be presented at the meeting.

Keywords: IVF, embryo, machine learning, time-lapse imaging data

Procedia PDF Downloads 94