Search results for: verification and validation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1805

Search results for: verification and validation

1715 V0 Physics at LHCb. RIVET Analysis Module for Z Boson Decay to Di-Electron

Authors: A. E. Dumitriu

Abstract:

The LHCb experiment is situated at one of the four points around CERN’s Large Hadron Collider, being a single-arm forward spectrometer covering 10 mrad to 300 (250) mrad in the bending (non-bending) plane, designed primarily to study particles containing b and c quarks. Each one of LHCb’s sub-detectors specializes in measuring a different characteristic of the particles produced by colliding protons, its significant detection characteristics including a high precision tracking system and 2 ring-imaging Cherenkov detectors for particle identification. The major two topics that I am currently concerned in are: the RIVET project (Robust Independent Validation of Experiment and Theory) which is an efficient and portable tool kit of C++ class library useful for validation and tuning of Monte Carlo (MC) event generator models by providing a large collection of standard experimental analyses useful for High Energy Physics MC generator development, validation, tuning and regression testing and V0 analysis for 2013 LHCb NoBias type data (trigger on bunch + bunch crossing) at √s=2.76 TeV.

Keywords: LHCb physics, RIVET plug-in, RIVET, CERN

Procedia PDF Downloads 393
1714 Fiber Based Pushover Analysis of Reinforced Concrete Frame

Authors: Shewangizaw Tesfaye Wolde

Abstract:

The current engineering community has developed a method called performance based seismic design in which we design structures based on predefined performance levels set by the parties. Since we design our structures economically for the maximum actions expected in the life of structures they go beyond their elastic limit, in need of nonlinear analysis. In this paper conventional pushover analysis (nonlinear static analysis) is used for the performance assessment of the case study Reinforced Concrete (RC) Frame building located in Addis Ababa City, Ethiopia where proposed peak ground acceleration value by RADIUS 1999 project and others is more than twice as of EBCS-8:1995 (RADIUS 1999 project) by taking critical planar frame. Fiber beam-column model is used to control material nonlinearity with tension stiffening effect. The reliability of the fiber model and validation of software outputs are checked under verification chapter. Therefore, the aim of this paper is to propose a way for structural performance assessment of existing reinforced concrete frame buildings as well as design check.

Keywords: seismic, performance, fiber model, tension stiffening, reinforced concrete

Procedia PDF Downloads 36
1713 Enhanced Automated Teller Machine Using Short Message Service Authentication Verification

Authors: Rasheed Gbenga Jimoh, Akinbowale Nathaniel Babatunde

Abstract:

The use of Automated Teller Machine (ATM) has become an important tool among commercial banks, customers of banks have come to depend on and trust the ATM conveniently meet their banking needs. Although the overwhelming advantages of ATM cannot be over-emphasized, its alarming fraud rate has become a bottleneck in it’s full adoption in Nigeria. This study examined the menace of ATM in the society another cost of running ATM services by banks in the country. The researcher developed a prototype of an enhanced Automated Teller Machine Authentication using Short Message Service (SMS) Verification. The developed prototype was tested by Ten (10) respondents who are users of ATM cards in the country and the data collected was analyzed using Statistical Package for Social Science (SPSS). Based on the results of the analysis, it is being envisaged that the developed prototype will go a long way in reducing the alarming rate of ATM fraud in Nigeria.

Keywords: ATM, ATM fraud, e-banking, prototyping

Procedia PDF Downloads 279
1712 Zero-Knowledge Proof-of-Reserve: A Confidential Approach to Cryptocurrency Asset Verification

Authors: Sam Ng, Lewis Leighton, Sam Atkinson, Carson Yan, Landan Hu, Leslie Cheung, Brian Yap, Kent Lung, Ketat Sarakune

Abstract:

This paper introduces a method for verifying cryptocurrency reserves that balances the need for both transparency and data confidentiality. Our methodology employs cryptographic techniques, including Merkle Trees, Bulletproof, and zkSnark, to verify that total assets equal or exceed total liabilities, represented by customer funds. Importantly, this verification is achieved without disclosing sensitive information such as the total asset value, customer count, or cold wallet addresses. We delve into the construction and implementation of this methodology. While the system is robust and scalable, we also identify areas for potential enhancements to improve its efficiency and versatility. As the digital asset landscape continues to evolve, our approach provides a solid foundation for ensuring continued trust and security in digital asset platforms.

Keywords: cryptocurrency, crypto-currency, proof-of-reserve, por, zero-knowledge, ZKP

Procedia PDF Downloads 42
1711 Development of Multi-Leaf Collimator-Based Isocenter Verification Tool Using Electrical Portal Imaging Device for Stereotactic Radiosurgery

Authors: Panatda Intanin, Sangutid Thongsawad, Chirapha Tannanonta, Todsaporn Fuangrod

Abstract:

Stereotactic radiosurgery (SRS) is a highly precision delivery technique that requires comprehensive quality assurance (QA) tests prior to treatment delivery. An isocenter of delivery beam plays a critical role that affect the treatment accuracy. The uncertainty of isocenter is traditionally accessed using circular cone equipment, Winston-Lutz (WL) phantom and film. This technique is considered time consuming and highly dependent on the observer. In this work, the development of multileaf collimator (MLC)-based isocenter verification tool using electronic portal imaging device (EPID) was proposed and evaluated. A mechanical isocenter alignment with ball bearing diameter 5 mm and circular cone diameter 10 mm fixed to gantry head defines the radiation field was set as the conventional WL test method. The conventional setup was to compare to the proposed setup; using MLC (10 x 10 mm) to define the radiation filed instead of cone. This represents more realistic delivery field than using circular cone equipment. The acquisition from electronic portal imaging device (EPID) and radiographic film were performed in both experiments. The gantry angles were set as following: 0°, 90°, 180° and 270°. A software tool was in-house developed using MATLAB/SIMULINK programming to determine the centroid of radiation field and shadow of WL phantom automatically. This presents higher accuracy than manual measurement. The deviation between centroid of both cone-based and MLC-based WL tests were quantified. To compare between film and EPID image, the deviation for all gantry angle was 0.26±0.19mm and 0.43±0.30 for cone-based and MLC-based WL tests. For the absolute deviation calculation on EPID images between cone and MLC-based WL test was 0.59±0.28 mm and the absolute deviation on film images was 0.14±0.13 mm. Therefore, the MLC-based isocenter verification using EPID present high sensitivity tool for SRS QA.

Keywords: isocenter verification, quality assurance, EPID, SRS

Procedia PDF Downloads 117
1710 Basic Study of Mammographic Image Magnification System with Eye-Detector and Simple EEG Scanner

Authors: Aika Umemuro, Mitsuru Sato, Mizuki Narita, Saya Hori, Saya Sakurai, Tomomi Nakayama, Ayano Nakazawa, Toshihiro Ogura

Abstract:

Mammography requires the detection of very small calcifications, and physicians search for microcalcifications by magnifying the images as they read them. The mouse is necessary to zoom in on the images, but this can be tiring and distracting when many images are read in a single day. Therefore, an image magnification system combining an eye-detector and a simple electroencephalograph (EEG) scanner was devised, and its operability was evaluated. Two experiments were conducted in this study: the measurement of eye-detection error using an eye-detector and the measurement of the time required for image magnification using a simple EEG scanner. Eye-detector validation showed that the mean distance of eye-detection error ranged from 0.64 cm to 2.17 cm, with an overall mean of 1.24 ± 0.81 cm for the observers. The results showed that the eye detection error was small enough for the magnified area of the mammographic image. The average time required for point magnification in the verification of the simple EEG scanner ranged from 5.85 to 16.73 seconds, and individual differences were observed. The reason for this may be that the size of the simple EEG scanner used was not adjustable, so it did not fit well for some subjects. The use of a simple EEG scanner with size adjustment would solve this problem. Therefore, the image magnification system using the eye-detector and the simple EEG scanner is useful.

Keywords: EEG scanner, eye-detector, mammography, observers

Procedia PDF Downloads 195
1709 Model Predictive Controller for Pasteurization Process

Authors: Tesfaye Alamirew Dessie

Abstract:

Our study focuses on developing a Model Predictive Controller (MPC) and evaluating it against a traditional PID for a pasteurization process. Utilizing system identification from the experimental data, the dynamics of the pasteurization process were calculated. Using best fit with data validation, residual, and stability analysis, the quality of several model architectures was evaluated. The validation data fit the auto-regressive with exogenous input (ARX322) model of the pasteurization process by roughly 80.37 percent. The ARX322 model structure was used to create MPC and PID control techniques. After comparing controller performance based on settling time, overshoot percentage, and stability analysis, it was found that MPC controllers outperform PID for those parameters.

Keywords: MPC, PID, ARX, pasteurization

Procedia PDF Downloads 123
1708 Creation and Validation of a Measurement Scale of E-Management: An Exploratory and Confirmatory Study

Authors: Hamadi Khlif

Abstract:

This paper deals with the understanding of the concept of e-management and the development of a measuring instrument adapted to the new problems encountered during the application of this new practice within the modern enterprise. Two principal e-management factors have been isolated in an exploratory study carried out among 260 participants. A confirmatory study applied to a second sample of 270 participants has been established in a cross-validation of the scale of measurement. The study presents the literature review specifically dedicated to e-management and the results of the exploratory and confirmatory phase of the development of this scale, which demonstrates satisfactory psychometric qualities. The e-management has two dimensions: a managerial dimension and a technological dimension.

Keywords: e-management, management, ICT deployment, mode of management

Procedia PDF Downloads 288
1707 Cloud-Based Dynamic Routing with Feedback in Formal Methods

Authors: Jawid Ahmad Baktash, Mursal Dawodi, Tomokazu Nagata

Abstract:

With the rapid growth of Cloud Computing, Formal Methods became a good choice for the refinement of message specification and verification for Dynamic Routing in Cloud Computing. Cloud-based Dynamic Routing is becoming increasingly popular. We propose feedback in Formal Methods for Dynamic Routing and Cloud Computing; the model and topologies show how to send messages from index zero to all others formally. The responsibility of proper verification becomes crucial with Dynamic Routing in the cloud. Formal Methods can play an essential role in the routing and development of Networks, and the testing of distributed systems. Event-B is a formal technique that consists of describing the problem rigorously and introduces solutions or details in the refinement steps. Event-B is a variant of B, designed for developing distributed systems and message passing of the dynamic routing. In Event-B and formal methods, the events consist of guarded actions occurring spontaneously rather than being invoked.

Keywords: cloud, dynamic routing, formal method, Pro-B, event-B

Procedia PDF Downloads 385
1706 Experimental Verification and Finite Element Analysis of a Sliding Door System Used in Automotive Industry

Authors: C. Guven, M. Tufekci, E. Bayik, O. Gedik, M. Tas

Abstract:

A sliding door system is used in commercial vehicles and passenger cars to allow a larger unobstructed access to the interior for loading and unloading. The movement of a sliding door on vehicle body is ensured by mechanisms and tracks having special cross-section which is manufactured by roll forming and stretch bending process. There are three tracks and three mechanisms which are called upper, central and lower on a sliding door system. There are static requirements as strength on different directions, rigidity for mechanisms, and door drop off, door sag; dynamic requirements as high energy slam opening-closing and durability requirement to validate these products. In addition, there is a kinematic requirement to find out force values from door handle during manual operating. In this study, finite element analysis and physical test results which are realized for sliding door systems will be shared comparatively.

Keywords: finite element analysis, sliding door, experimental, verification, vehicle tests

Procedia PDF Downloads 304
1705 The Study on Blast Effect of Polymer Gel by Trazul Lead Block Test and Concrete Block Test

Authors: Young-Hun Ko, Seung-Jun Kim, Khaqan Baluch, Hyung- Sik Yang

Abstract:

In this study, the polymer gel was used as coupling material in a blasting hole and its comparison was made with other coupling materials like sand, water, and air. Trazul lead block test and AUTODYN numerical analysis were conducted to analyze the effects of the coupling materials on the intensity of the explosion, as well as the verification tests were conducted by using concrete block test. The emulsion explosives were used in decoupling conditions, sand, water, and polymer gel were used as the coupling materials. The lead block test and the numerical analysis showed that the expansion of the blast hole in the lead block was similar to that of the water and gelatin and followed by sand and air conditions. The validation of concrete block test result showed the similar result as Trazul lead block test and the explosion strength was measured at 0.8 for polymer gel, 0.7 for sand, and 0.6 for no coupling material, in comparison to the full charge (1.0) case.

Keywords: Trazul lead block test, AUTODYN numerical analysis, coupling material, polymer gel, soil covering concrete block explosion test

Procedia PDF Downloads 261
1704 Validation and Projections for Solar Radiation up to 2100: HadGEM2-AO Global Circulation Model

Authors: Elison Eduardo Jardim Bierhals, Claudineia Brazil, Deivid Pires, Rafael Haag, Elton Gimenez Rossini

Abstract:

The objective of this work is to evaluate the results of solar radiation projections between 2006 and 2013 for the state of Rio Grande do Sul, Brazil. The projections are provided by the General Circulation Models (MCGs) belonging to the Coupled Model Intercomparison Phase 5 (CMIP5). In all, the results of the simulation of six models are evaluated, compared to monthly data, measured by a network of thirteen meteorological stations of the National Meteorological Institute (INMET). The performance of the models is evaluated by the Nash coefficient and the Bias. The results are presented in the form of tables, graphs and spatialization maps. The ACCESS1-0 RCP 4.5 model presented the best results for the solar radiation simulations, for the most optimistic scenario, in much of the state. The efficiency coefficients (CEF) were between 0.95 and 0.98. In the most pessimistic scenario, HADGen2-AO RCP 8.5 had the best accuracy among the analyzed models, presenting coefficients of efficiency between 0.94 and 0.98. From this validation, solar radiation projection maps were elaborated, indicating a seasonal increase of this climatic variable in some regions of the Brazilian territory, mainly in the spring.

Keywords: climate change, projections, solar radiation, validation

Procedia PDF Downloads 156
1703 Identity Verification Based on Multimodal Machine Learning on Red Green Blue (RGB) Red Green Blue-Depth (RGB-D) Voice Data

Authors: LuoJiaoyang, Yu Hongyang

Abstract:

In this paper, we experimented with a new approach to multimodal identification using RGB, RGB-D and voice data. The multimodal combination of RGB and voice data has been applied in tasks such as emotion recognition and has shown good results and stability, and it is also the same in identity recognition tasks. We believe that the data of different modalities can enhance the effect of the model through mutual reinforcement. We try to increase the three modalities on the basis of the dual modalities and try to improve the effectiveness of the network by increasing the number of modalities. We also implemented the single-modal identification system separately, tested the data of these different modalities under clean and noisy conditions, and compared the performance with the multimodal model. In the process of designing the multimodal model, we tried a variety of different fusion strategies and finally chose the fusion method with the best performance. The experimental results show that the performance of the multimodal system is better than that of the single modality, especially in dealing with noise, and the multimodal system can achieve an average improvement of 5%.

Keywords: multimodal, three modalities, RGB-D, identity verification

Procedia PDF Downloads 46
1702 Optimization of Element Type for FE Model and Verification of Analyses with Physical Tests

Authors: Mustafa Tufekci, Caner Guven

Abstract:

In Automotive Industry, sliding door systems that are also used as body closures, are safety members. Extreme product tests are realized to prevent failures in a design process, but these tests realized experimentally result in high costs. Finite element analysis is an effective tool used for the design process. These analyses are used before production of a prototype for validation of design according to customer requirement. In result of this, the substantial amount of time and cost is saved. Finite element model is created for geometries that are designed in 3D CAD programs. Different element types as bar, shell and solid, can be used for creating mesh model. The cheaper model can be created by the selection of element type, but combination of element type that was used in model, number and geometry of element and degrees of freedom affects the analysis result. Sliding door system is a good example which used these methods for this study. Structural analysis was realized for sliding door mechanism by using FE models. As well, physical tests that have same boundary conditions with FE models were realized. Comparison study for these element types, were done regarding test and analyses results then the optimum combination was achieved.

Keywords: finite element analysis, sliding door mechanism, element type, structural analysis

Procedia PDF Downloads 295
1701 Landslide Susceptibility Mapping: A Comparison between Logistic Regression and Multivariate Adaptive Regression Spline Models in the Municipality of Oudka, Northern of Morocco

Authors: S. Benchelha, H. C. Aoudjehane, M. Hakdaoui, R. El Hamdouni, H. Mansouri, T. Benchelha, M. Layelmam, M. Alaoui

Abstract:

The logistic regression (LR) and multivariate adaptive regression spline (MarSpline) are applied and verified for analysis of landslide susceptibility map in Oudka, Morocco, using geographical information system. From spatial database containing data such as landslide mapping, topography, soil, hydrology and lithology, the eight factors related to landslides such as elevation, slope, aspect, distance to streams, distance to road, distance to faults, lithology map and Normalized Difference Vegetation Index (NDVI) were calculated or extracted. Using these factors, landslide susceptibility indexes were calculated by the two mentioned methods. Before the calculation, this database was divided into two parts, the first for the formation of the model and the second for the validation. The results of the landslide susceptibility analysis were verified using success and prediction rates to evaluate the quality of these probabilistic models. The result of this verification was that the MarSpline model is the best model with a success rate (AUC = 0.963) and a prediction rate (AUC = 0.951) higher than the LR model (success rate AUC = 0.918, rate prediction AUC = 0.901).

Keywords: landslide susceptibility mapping, regression logistic, multivariate adaptive regression spline, Oudka, Taounate

Procedia PDF Downloads 160
1700 Runtime Monitoring Using Policy-Based Approach to Control Information Flow for Mobile Apps

Authors: Mohamed Sarrab, Hadj Bourdoucen

Abstract:

Mobile applications are verified to check the correctness or evaluated to check the performance with respect to specific security properties such as availability, integrity, and confidentiality. Where they are made available to the end users of the mobile application is achievable only to a limited degree using software engineering static verification techniques. The more sensitive the information, such as credit card data, personal medical information or personal emails being processed by mobile application, the more important it is to ensure the confidentiality of this information. Monitoring non-trusted mobile application during execution in an environment where sensitive information is present is difficult and unnerving. The paper addresses the issue of monitoring and controlling the flow of confidential information during non-trusted mobile application execution. The approach concentrates on providing a dynamic and usable information security solution by interacting with the mobile users during the run-time of mobile application in response to information flow events.

Keywords: mobile application, run-time verification, usable security, direct information flow

Procedia PDF Downloads 353
1699 The Comparison and Optimization of the Analytic Method for Canthaxanthin, Food Colorants

Authors: Hee-Jae Suh, Kyung-Su Kim, Min-Ji Kim, Yeon-Seong Jeong, Ok-Hwan Lee, Jae-Wook Shin, Hyang-Sook Chun, Chan Lee

Abstract:

Canthaxanthin is keto-carotenoid produced from beta-carotene and it has been approved to be used in many countries as a food coloring agent. Canthaxanthin has been analyzed using High Performance Liquid Chromatography (HPLC) system with various ways of pretreatment methods. Four official methods for verification of canthaxanthin at FSA (UK), AOAC (US), EFSA (EU) and MHLW (Japan) were compared to improve its analytical and the pretreatment method. The Linearity, the limit of detection (LOD), the limit of quantification (LOQ), the accuracy, the precision and the recovery ratio were determined from each method with modification in pretreatment method. All HPLC methods exhibited correlation coefficients of calibration curves for canthaxanthin as 0.9999. The analysis methods from FSA, AOAC, and MLHW showed the LOD of 0.395 ppm, 0.105 ppm, and 0.084 ppm, and the LOQ of 1.196 ppm, 0.318 ppm, 0.254 ppm, respectively. Among tested methods, HPLC method of MHLW with modification in pretreatments was finally selected for the analysis of canthaxanthin in lab, because it exhibited the resolution factor of 4.0 and the selectivity of 1.30. This analysis method showed a correlation coefficients value of 0.9999 and the lowest LOD and LOQ. Furthermore, the precision ratio was lower than 1 and the accuracy was almost 100%. The method presented the recovery ratio of 90-110% with modification in pretreatment method. The cross-validation of coefficient variation was 5 or less among tested three institutions in Korea.

Keywords: analytic method, canthaxanthin, food colorants, pretreatment method

Procedia PDF Downloads 657
1698 Assessing Relationships between Glandularity and Gray Level by Using Breast Phantoms

Authors: Yun-Xuan Tang, Pei-Yuan Liu, Kun-Mu Lu, Min-Tsung Tseng, Liang-Kuang Chen, Yuh-Feng Tsai, Ching-Wen Lee, Jay Wu

Abstract:

Breast cancer is predominant of malignant tumors in females. The increase in the glandular density increases the risk of breast cancer. BI-RADS is a frequently used density indicator in mammography; however, it significantly overestimates the glandularity. Therefore, it is very important to accurately and quantitatively assess the glandularity by mammography. In this study, 20%, 30% and 50% glandularity phantoms were exposed using a mammography machine at 28, 30 and 31 kVp, and 30, 55, 80 and 105 mAs, respectively. The regions of interest (ROIs) were drawn to assess the gray level. The relationship between the glandularity and gray level under various compression thicknesses, kVp, and mAs was established by the multivariable linear regression. A phantom verification was performed with automatic exposure control (AEC). The regression equation was obtained with an R-square value of 0.928. The average gray levels of the verification phantom were 8708, 8660 and 8434 for 0.952, 0.963 and 0.985 g/cm3, respectively. The percent differences of glandularity to the regression equation were 3.24%, 2.75% and 13.7%. We concluded that the proposed method could be clinically applied in mammography to improve the glandularity estimation and further increase the importance of breast cancer screening.

Keywords: mammography, glandularity, gray value, BI-RADS

Procedia PDF Downloads 464
1697 Method Validation for Determining Platinum and Palladium in Catalysts Using Inductively Coupled Plasma Optical Emission Spectrometry

Authors: Marin Senila, Oana Cadar, Thorsten Janisch, Patrick Lacroix-Desmazes

Abstract:

The study presents the analytical capability and validation of a method based on microwave-assisted acid digestion for quantitative determination of platinum and palladium in catalysts using inductively coupled plasma optical emission spectrometry (ICP-OES). In order to validate the method, the main figures of merit such as limit of detection and limit of quantification, precision and accuracy were considered and the measurement uncertainty was estimated based on the bottom-up approach according to the international guidelines of ISO/IEC 17025. Limit of detections, estimated from blank signal using 3 s criterion, were 3.0 mg/kg for Pt and respectively 3.6 mg/kg for Pd, while limits of quantification were 9.0 mg/kg for Pt and respectively 10.8 mg/kg for Pd. Precisions, evaluated as standard deviations of repeatability (n=5 parallel samples), were less than 10% for both precious metals. Accuracies of the method, verified by recovery estimation certified reference material NIST SRM 2557 - pulverized recycled monolith, were 99.4 % for Pt and 101% for Pd. The obtained limit of quantifications and accuracy were satisfactory for the intended purpose. The paper offers all the steps necessary to validate the determination method for Pt and Pd in catalysts using inductively coupled plasma optical emission spectrometry.

Keywords: catalyst analysis, ICP-OES, method validation, platinum, palladium

Procedia PDF Downloads 135
1696 Ontology-Driven Knowledge Discovery and Validation from Admission Databases: A Structural Causal Model Approach for Polytechnic Education in Nigeria

Authors: Bernard Igoche Igoche, Olumuyiwa Matthew, Peter Bednar, Alexander Gegov

Abstract:

This study presents an ontology-driven approach for knowledge discovery and validation from admission databases in Nigerian polytechnic institutions. The research aims to address the challenges of extracting meaningful insights from vast amounts of admission data and utilizing them for decision-making and process improvement. The proposed methodology combines the knowledge discovery in databases (KDD) process with a structural causal model (SCM) ontological framework. The admission database of Benue State Polytechnic Ugbokolo (Benpoly) is used as a case study. The KDD process is employed to mine and distill knowledge from the database, while the SCM ontology is designed to identify and validate the important features of the admission process. The SCM validation is performed using the conditional independence test (CIT) criteria, and an algorithm is developed to implement the validation process. The identified features are then used for machine learning (ML) modeling and prediction of admission status. The results demonstrate the adequacy of the SCM ontological framework in representing the admission process and the high predictive accuracies achieved by the ML models, with k-nearest neighbors (KNN) and support vector machine (SVM) achieving 92% accuracy. The study concludes that the proposed ontology-driven approach contributes to the advancement of educational data mining and provides a foundation for future research in this domain.

Keywords: admission databases, educational data mining, machine learning, ontology-driven knowledge discovery, polytechnic education, structural causal model

Procedia PDF Downloads 22
1695 Optimization of Cutting Forces in Drilling of Polimer Composites via Taguchi Methodology

Authors: Eser Yarar, Fahri Vatansever, A. Tamer Erturk, Sedat Karabay

Abstract:

In this study, drilling behavior of multi-layer orthotropic polyester composites reinforced with woven polyester fiber and PTFE particle was investigated. Conventional drilling methods have low cost and ease of use. Therefore, it is one of the most preferred machining methods. The increasing range of use of composite materials in many areas has led to the investigation of the machinability performance of these materials. The drilling capability of the synthetic polymer composite material was investigated by measuring the cutting forces using different tool diameters, feed rate and high cutting speed parameters. Cutting forces were measured using a dynamometer in the experiments. In order to evaluate the results of the experiment, the Taguchi experimental design method was used. According to the results, the optimum cutting parameters were obtained for 0.1 mm/rev, 1070 rpm and 2 mm diameter drill bit. Verification tests were performed for the optimum cutting parameters obtained according to the model. Verification experiments showed the success of the established model.

Keywords: cutting force, drilling, polimer composite, Taguchi

Procedia PDF Downloads 137
1694 Camera Model Identification for Mi Pad 4, Oppo A37f, Samsung M20, and Oppo f9

Authors: Ulrich Wake, Eniman Syamsuddin

Abstract:

The model for camera model identificaiton is trained using pretrained model ResNet43 and ResNet50. The dataset consists of 500 photos of each phone. Dataset is divided into 1280 photos for training, 320 photos for validation and 400 photos for testing. The model is trained using One Cycle Policy Method and tested using Test-Time Augmentation. Furthermore, the model is trained for 50 epoch using regularization such as drop out and early stopping. The result is 90% accuracy for validation set and above 85% for Test-Time Augmentation using ResNet50. Every model is also trained by slightly updating the pretrained model’s weights

Keywords: ​ One Cycle Policy, ResNet34, ResNet50, Test-Time Agumentation

Procedia PDF Downloads 169
1693 Personalized Infectious Disease Risk Prediction System: A Knowledge Model

Authors: Retno A. Vinarti, Lucy M. Hederman

Abstract:

This research describes a knowledge model for a system which give personalized alert to users about infectious disease risks in the context of weather, location and time. The knowledge model is based on established epidemiological concepts augmented by information gleaned from infection-related data repositories. The existing disease risk prediction research has more focuses on utilizing raw historical data and yield seasonal patterns of infectious disease risk emergence. This research incorporates both data and epidemiological concepts gathered from Atlas of Human Infectious Disease (AHID) and Centre of Disease Control (CDC) as basic reasoning of infectious disease risk prediction. Using CommonKADS methodology, the disease risk prediction task is an assignment synthetic task, starting from knowledge identification through specification, refinement to implementation. First, knowledge is gathered from AHID primarily from the epidemiology and risk group chapters for each infectious disease. The result of this stage is five major elements (Person, Infectious Disease, Weather, Location and Time) and their properties. At the knowledge specification stage, the initial tree model of each element and detailed relationships are produced. This research also includes a validation step as part of knowledge refinement: on the basis that the best model is formed using the most common features, Frequency-based Selection (FBS) is applied. The portion of the Infectious Disease risk model relating to Person comes out strongest, with Location next, and Weather weaker. For Person attribute, Age is the strongest, Activity and Habits are moderate, and Blood type is weakest. At the Location attribute, General category (e.g. continents, region, country, and island) results much stronger than Specific category (i.e. terrain feature). For Weather attribute, Less Precise category (i.e. season) comes out stronger than Precise category (i.e. exact temperature or humidity interval). However, given that some infectious diseases are significantly more serious than others, a frequency based metric may not be appropriate. Future work will incorporate epidemiological measurements of disease seriousness (e.g. odds ratio, hazard ratio and fatality rate) into the validation metrics. This research is limited to modelling existing knowledge about epidemiology and chain of infection concepts. Further step, verification in knowledge refinement stage, might cause some minor changes on the shape of tree.

Keywords: epidemiology, knowledge modelling, infectious disease, prediction, risk

Procedia PDF Downloads 207
1692 Modeling and Validation of Microspheres Generation in the Modified T-Junction Device

Authors: Lei Lei, Hongbo Zhang, Donald J. Bergstrom, Bing Zhang, K. Y. Song, W. J. Zhang

Abstract:

This paper presents a model for a modified T-junction device for microspheres generation. The numerical model is developed using a commercial software package: COMSOL Multiphysics. In order to test the accuracy of the numerical model, multiple variables, such as the flow rate of cross-flow, fluid properties, structure, and geometry of the microdevice are applied. The results from the model are compared with the experimental results in the diameter of the microsphere generated. The comparison shows a good agreement. Therefore the model is useful in further optimization of the device and feedback control of microsphere generation if any.

Keywords: CFD modeling, validation, microsphere generation, modified T-junction

Procedia PDF Downloads 674
1691 Computational Screening of Secretory Proteins with Brain-Specific Expression in Glioblastoma Multiforme

Authors: Sumera, Sanila Amber, Fatima Javed Mirza, Amjad Ali, Saadia Zahid

Abstract:

Glioblastoma multiforme (GBM) is a widely spread and fatal primary brain tumor with an increased risk of relapse in spite of aggressive treatment. The current procedures for GBM diagnosis include invasive procedures i.e. resection or biopsy, to acquire tumor mass. Implementation of negligibly invasive tests as a potential diagnostic technique and biofluid-based monitoring of GBM stresses on discovering biomarkers in CSF and blood. Therefore, we performed a comprehensive in silico analysis to identify potential circulating biomarkers for GBM. Initially, six gene and protein databases were utilized to mine brain-specific proteins. The resulting proteins were filtered using a channel of five tools to predict the secretory proteins. Subsequently, the expression profile of the secreted proteins was verified in the brain and blood using two databases. Additional verification of the resulting proteins was done using Plasma Proteome Database (PPD) to confirm their presence in blood. The final set of proteins was searched in literature for their relationship with GBM, keeping a special emphasis on secretome proteome. 2145 proteins were firstly mined as brain-specific, out of which 69 proteins were identified as secretory in nature. Verification of expression profile in brain and blood eliminated 58 proteins from the 69 proteins, providing a final list of 11 proteins. Further verification of these 11 proteins further eliminated 2 proteins, giving a final set of nine secretory proteins i.e. OPCML, NPTX1, LGI1, CNTN2, LY6H, SLIT1, CREG2, GDF1 and SERPINI1. Out of these 9 proteins, 7 were found to be linked to GBM, whereas 2 proteins are not investigated in GBM so far. We propose that these secretory proteins can serve as potential circulating biomarker signatures of GBM and will facilitate the development of minimally invasive diagnostic methods and novel therapeutic interventions for GBM.

Keywords: glioblastoma multiforme, secretory proteins, brain secretome, biomarkers

Procedia PDF Downloads 120
1690 Study of Wake Dynamics for a Rim-Driven Thruster Based on Numerical Method

Authors: Bao Liu, Maarten Vanierschot, Frank Buysschaert

Abstract:

The present work examines the wake dynamics of a rim-driven thruster (RDT) with Computational Fluid Dynamics (CFD). Unsteady Reynolds-averaged Navier-Stokes (URANS) equations were solved in the commercial solver ANSYS Fluent in combination with the SST k-ω turbulence model. The application of the moving reference frame (MRF) and sliding mesh (SM) approach to handling the rotational movement of the propeller were compared in the transient simulations. Validation and verification of the numerical model was performed to ensure numerical accuracy. Two representative scenarios were considered, i.e., the bollard condition (J=0) and a very light loading condition(J=0.7), respectively. From the results, it’s confirmed that compared to the SM method, the MRF method is not suitable for resolving the unsteady flow features as it only gives the general mean flow but smooths out lots of characteristic details in the flow field. By evaluating the simulation results with the SM technique, the instantaneous wake flow field under both conditions is presented and analyzed, most notably the helical vortex structure. It’s observed from the results that the tip vortices, blade shed vortices, and hub vortices are present in the wake flow field and convect downstream in a highly non-linear way. The shear layer vortices shedding from the duct displayed a strong interaction with the distorted tip vortices in an irregularmanner.

Keywords: computational fluid dynamics, rim-driven thruster, sliding mesh, wake dynamics

Procedia PDF Downloads 197
1689 Mathematical Modelling of Different Types of Body Support Surface for Pressure Ulcer Prevention

Authors: Mahbub C. Mishu, Venktesh N. Dubey, Tamas Hickish, Jonathan Cole

Abstract:

Pressure ulcer is a common problem for today's healthcare industry. It occurs due to external load applied to the skin. Also when the subject is immobile for a longer period of time and there is continuous load applied to a particular area of human body,blood flow gets reduced and as a result pressure ulcer develops. Body support surface has a significant role in preventing ulceration so it is important to know the characteristics of support surface under loading conditions. In this paper we have presented mathematical models of different types of viscoelastic materials and also we have shown the validation of our simulation results with experiments.

Keywords: pressure ulcer, viscoelastic material, mathematical model, experimental validation

Procedia PDF Downloads 283
1688 Reliability Verification of the Performance Evaluation of Multiphase Pump

Authors: Joon-Hyung Kim, Him-Chan Lee, Jin-Hyuk Kim, Yong-Kab Lee, Young-Seok Choi

Abstract:

The crude oil in an oil well exists in various phases such as gas, seawater, and sand, as well as oil. Therefore, a phase separator is needed at the front of a single-phase pump for pressurization and transfer. On the other hand, the application of a multiphase pump can provide such advantages as simplification of the equipment structure and cost savings, because there is no need for a phase separation process. Therefore, the crude oil transfer method using a multiphase pump is being applied to recently developed oil wells. Due to this increase in demand, technical demands for the development of multiphase pumps are sharply increasing, but the progress of research into related technologies is insufficient, due to the nature of multiphase pumps that require high levels of skills. This study was conducted to verify the reliability of pump performance evaluation using numerical analysis, which is the basis of the development of a multiphase pump. For this study, a model was designed by selecting the specifications of the pump under study. The performance of the designed model was evaluated through numerical analysis and experiment, and the results of the performance evaluation were compared to verify the reliability of the result using numerical analysis.

Keywords: multiphase pump, numerical analysis, experiment, performance evaluation, reliability verification

Procedia PDF Downloads 399
1687 Establishing Reference Intervals for Routine Coagulation Tests

Authors: Santina Sahibon, Sivasooriar Sivaneson, Martin Giddy, Nelson Nheu, Siti Sazeelah, Choo Kok Ming, Thuhairah Abdul Rahman, Fatmawati Binti Kamal

Abstract:

Introduction: Establishing population-based reference intervals (RI) are essential when evaluating laboratory test results and for method verification. Our laboratory initiated an exercise to establish RI for routine coagulation profile as part of the method verification procedure and to determine any differences in RI between three analyzers planned to be used in the laboratory. Methodology: 145 blood samples were collected and analysed for activated partial thromboplastin time (aPTT), prothrombin time (PT), international normalized ratio (INR), and fibrinogen] using three coagulation analysers which were CA104, CA660, and CS-2500 (Sysmex, USA). RI was established at 2.5th and 97.5th percentiles. Results: The RI for aPTT between C104, C660 and CS-2500 are (RI: 20.5-30.2 sec), (RI: 21.5-29.2 sec) and (RI: 22.7-30.3 sec) respectively. The RI for PT were (RI: 7.5-10.3 sec), (RI: 9.2- 11.1 sec) and (RI: 9.8-11.9 sec) for C104, CA660 and CS-2500 respectively. INR had an RI of (RI: 0.87- 1.16), (RI: 0.89-1.10) and (0.90-1.11) respectively on CA104, C660 and CS-2500. Fibrinogen RI was (RI: 2.04-4.62 g/L) and (2.05-4.76 g/L) on the CA660 and CS-2500, respectively. Conclusion: The RI was similar across the analytical platforms for aPTT, INR, and fibrinogen. However, CA104 showed lower RI compared to the other two analysers for PT. This highlights the potential variability in results between instruments that need to be addressed when verifying RI.

Keywords: coagulation, reference interval, APTT, PT, INR, fibrinogen

Procedia PDF Downloads 154
1686 Management Software for the Elaboration of an Electronic File in the Pharmaceutical Industry Following Mexican Regulations

Authors: M. Peña Aguilar Juan, Ríos Hernández Ezequiel, R. Valencia Luis

Abstract:

For certification, certain goods of public interest, such as medicines and food, it is required the preparation and delivery of a dossier. For its elaboration, legal and administrative knowledge must be taken, as well as organization of the documents of the process, and an order that allows the file verification. Therefore, a virtual platform was developed to support the process of management and elaboration of the dossier, providing accessibility to the information and interfaces that allow the user to know the status of projects. The development of dossier system on the cloud allows the inclusion of the technical requirements for the software management, including the validation and the manufacturing in the field industry. The platform guides and facilitates the dossier elaboration (report, file or history), considering Mexican legislation and regulations, it also has auxiliary tools for its management. This technological alternative provides organization support for documents and accessibility to the information required to specify the successful development of a dossier. The platform divides into the following modules: System control, catalog, dossier and enterprise management. The modules are designed per the structure required in a dossier in those areas. However, the structure allows for flexibility, as its goal is to become a tool that facilitates and does not obstruct processes. The architecture and development of the software allows flexibility for future work expansion to other fields, this would imply feeding the system with new regulations.

Keywords: electronic dossier, cloud management software, pharmaceutical industry, sanitary registration

Procedia PDF Downloads 266