Search results for: accuracy of payment time
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 20341

Search results for: accuracy of payment time

20191 Research on Development and Accuracy Improvement of an Explosion Proof Combustible Gas Leak Detector Using an IR Sensor

Authors: Gyoutae Park, Seungho Han, Byungduk Kim, Youngdo Jo, Yongsop Shim, Yeonjae Lee, Sangguk Ahn, Hiesik Kim, Jungil Park

Abstract:

In this paper, we presented not only development technology of an explosion proof type and portable combustible gas leak detector but also algorithm to improve accuracy for measuring gas concentrations. The presented techniques are to apply the flame-proof enclosure and intrinsic safe explosion proof to an infrared gas leak detector at first in Korea and to improve accuracy using linearization recursion equation and Lagrange interpolation polynomial. Together, we tested sensor characteristics and calibrated suitable input gases and output voltages. Then, we advanced the performances of combustible gaseous detectors through reflecting demands of gas safety management fields. To check performances of two company's detectors, we achieved the measurement tests with eight standard gases made by Korea Gas Safety Corporation. We demonstrated our instruments better in detecting accuracy other than detectors through experimental results.

Keywords: accuracy improvement, IR gas sensor, gas leak, detector

Procedia PDF Downloads 372
20190 A Study of ZY3 Satellite Digital Elevation Model Verification and Refinement with Shuttle Radar Topography Mission

Authors: Bo Wang

Abstract:

As the first high-resolution civil optical satellite, ZY-3 satellite is able to obtain high-resolution multi-view images with three linear array sensors. The images can be used to generate Digital Elevation Models (DEM) through dense matching of stereo images. However, due to the clouds, forest, water and buildings covered on the images, there are some problems in the dense matching results such as outliers and areas failed to be matched (matching holes). This paper introduced an algorithm to verify the accuracy of DEM that generated by ZY-3 satellite with Shuttle Radar Topography Mission (SRTM). Since the accuracy of SRTM (Internal accuracy: 5 m; External accuracy: 15 m) is relatively uniform in the worldwide, it may be used to improve the accuracy of ZY-3 DEM. Based on the analysis of mass DEM and SRTM data, the processing can be divided into two aspects. The registration of ZY-3 DEM and SRTM can be firstly performed using the conjugate line features and area features matched between these two datasets. Then the ZY-3 DEM can be refined by eliminating the matching outliers and filling the matching holes. The matching outliers can be eliminated based on the statistics on Local Vector Binning (LVB). The matching holes can be filled by the elevation interpolated from SRTM. Some works are also conducted for the accuracy statistics of the ZY-3 DEM.

Keywords: ZY-3 satellite imagery, DEM, SRTM, refinement

Procedia PDF Downloads 319
20189 Reliability of Diffusion Tensor Imaging in Differentiation of Salivary Gland Tumors

Authors: Sally Salah El Menshawy, Ghada M. Ahmed GabAllah, Doaa Khedr M. Khedr

Abstract:

Background: Our study aims to detect the diagnostic role of DTI in the differentiation of salivary glands benign and malignant lesions. Results: Our study included 50 patients (25males and 25 females) divided into 4 groups (benign lesions n=20, malignant tumors n=13, post-operative changes n=10 and normal n=7). 28 patients were with parotid gland lesions, 4 patients were with submandibular gland lesions and only 1 case with sublingual gland affection. The mean fractional anisotropy (FA) and apparent diffusion coefficient (ADC) of malignant salivary gland tumors (n = 13) (0.380±0.082 and 0.877±0.234× 10⁻³ mm² s⁻¹) were significantly different (P<0.001) than that of benign tumors (n = 20) (0.147±0.03 and 1.47±0.605 × 10⁻³ mm² s⁻¹), respectively. The mean FA and ADC of post-operative changes (n = 10) were (0.211±0.069 and 1.63±0.20× 10⁻³ mm² s⁻¹) while that of normal glands (n =7) was (0.251±0.034and 1.54±0.29× 10⁻³ mm² s⁻¹), respectively. Using ADC to differentiate malignant lesions from benign lesions has an (AUC) of 0.810, with an accuracy of 69.7%. ADC used to differentiate malignant lesions from post-operative changes has (AUC) of 1.0, and an accuracy of 95.7%. FA used to discriminate malignant from benign lesions has (AUC) of 1.0, and an accuracy of 93.9%. FA used to differentiate malignant from post-operative changes has (AUC) of 0.923, and an accuracy of 95.7%. Combined FA and ADC used to differentiate malignant from benign lesions has (AUC) of 1.0, and an accuracy of 100%. Combined FA and ADC used to differentiate malignant from post-operative changes has (AUC) of 1.0, and an accuracy of 100%. Conclusion: Combined FA and ADC can differentiate malignant tumors from benign salivary gland lesions.

Keywords: diffusion tensor imaging, MRI, salivary gland, tumors

Procedia PDF Downloads 87
20188 Government Credit Card in State Financial Management: Public Sector Innovation in Indonesia

Authors: Paramita Nur Kurniati, Stanislaus Riyanta

Abstract:

In the midst of the heightened usage of electronic money (e-money), Indonesian government expenditure is yet governed through cash-basis transactions. This conventional system brings about a number of potential risks and obstacles to operational conduct, including state financial liquidity issue. Consequently, Ministry of Finance is currently establishing the cashless payment methods for State Budget (APBN). Included in those advance methods is credit card facility as a government expenditure payment scheme. This policy is one of the innovations within the public sector learned from other countries’ best practices. Moreover, this particular method is already prominent within the private-sector realm. Qualitative descriptive analysis technique is implemented to evaluate the contemporary innovation of using government credit card in the path towards cashless society. This approach is expected to generate several benefits for the government, particularly in minimizing corruption within the state financial management. Effective coordination among policy makers and policy implementers is essential for the success of this policy’s exercise, without neglecting prudence and public transparency aspects. Government credit card usage shall be the potent resolution for enhancing the government’s overall public service performance.

Keywords: cashless basis, cashless society, government credit card, public sector innovation

Procedia PDF Downloads 131
20187 Heritage 3D Digitalization Combining High Definition Photogrammetry with Metrologic Grade Laser Scans

Authors: Sebastian Oportus, Fabrizio Alvarez

Abstract:

3D digitalization of heritage objects is widely used nowadays. However, the most advanced 3D scanners in the market that capture topology and texture at the same time, and are specifically made for this purpose, don’t deliver the accuracy that is needed for scientific research. In the last three years, we have developed a method that combines the use of Metrologic grade laser scans, that allows us to work with a high accuracy topology up to 15 times more precise and combine this mesh with a texture obtained from high definition photogrammetry with up to 100 times more pixel concentrations. The result is an accurate digitalization that promotes heritage preservation, scientific study, high detail reproduction, and digital restoration, among others. In Chile, we have already performed 478 digitalizations of high-value heritage pieces and compared the results with up to five different digitalization methods; the results obtained show a considerable better dimensional accuracy and texture resolution. We know the importance of high precision and resolution for academics and museology; that’s why our proposal is to set a worldwide standard using this open source methodology.

Keywords: 3D digitalization, digital heritage, heritage preservation, digital restauration, heritage reproduction

Procedia PDF Downloads 164
20186 Presenting a Model Based on Artificial Neural Networks to Predict the Execution Time of Design Projects

Authors: Hamed Zolfaghari, Mojtaba Kord

Abstract:

After feasibility study the design phase is started and the rest of other phases are highly dependent on this phase. forecasting the duration of design phase could do a miracle and would save a lot of time. This study provides a fast and accurate Machine learning (ML) and optimization framework, which allows a quick duration estimation of project design phase, hence improving operational efficiency and competitiveness of a design construction company. 3 data sets of three years composed of daily time spent for different design projects are used to train and validate the ML models to perform multiple projects. Our study concluded that Artificial Neural Network (ANN) performed an accuracy of 0.94.

Keywords: time estimation, machine learning, Artificial neural network, project design phase

Procedia PDF Downloads 64
20185 Economic Evaluation of Cataract Eye Surgery by Health Attendant of Doctor and Nurse through the Social Insurance Board Cadr at General Hospital Anutapura Palu Central Sulawesi Indonesia

Authors: Sitti Rahmawati

Abstract:

Payment system of cataract surgery implemented by professional attendant of doctor and nurse has been increasing, through health insurance program and this has become one of the factors that affects a lot of government in the budget establishment. This system has been implemented in purpose of quality and expenditure control, i.e., controlling health overpayment to obtain benefit (moral hazard) by the user of insurance or health service provider. The increasing health cost becomes the main issue that hampers the society to receive required health service in cash payment-system. One of the efforts that should be taken by the government in health payment is by securing health insurance through society's health insurance. The objective of the study is to learn the capability of a patient to pay cataract eye operation for the elders. Method of study sample population in this study was patients who obtain health insurance board card for the society that was started in the first of tri-semester (January-March) 2015 and claimed in Indonesian software-Case Based Group as a purposive sampling of 40 patients. Results of the study show that total unit cost analysis of surgery service unit was obtained $75 for unit cost without AFC and salary of nurse and doctor. The operation tariff that has been implemented today at Anutapura hospitals in eye department is tariff without AFC and the salary of the employee is $80. The operation tariff of the unit cost calculation with double distribution model at $65. Conclusion, the calculation result of actual unit cost that is much greater causes incentive distribution system provided to an ophthalmologist at $37 and nurse at $20 for one operation. The surgery service tariff is still low; consequently, the hospital receives low revenue and the quality of health insurance in eye operation department is relatively low. In purpose of increasing the service quality, it requires adequately high cost to equip medical equipment and increase the number of professional health attendant in serving patients in cataract eye operation at hospital.

Keywords: economic evaluation, cataract operation, health attendant, health insurance system

Procedia PDF Downloads 147
20184 Vehicle Detection and Tracking Using Deep Learning Techniques in Surveillance Image

Authors: Abe D. Desta

Abstract:

This study suggests a deep learning-based method for identifying and following moving objects in surveillance video. The proposed method uses a fast regional convolution neural network (F-RCNN) trained on a substantial dataset of vehicle images to first detect vehicles. A Kalman filter and a data association technique based on a Hungarian algorithm are then used to monitor the observed vehicles throughout time. However, in general, F-RCNN algorithms have been shown to be effective in achieving high detection accuracy and robustness in this research study. For example, in one study The study has shown that the vehicle detection and tracking, the system was able to achieve an accuracy of 97.4%. In this study, the F-RCNN algorithm was compared to other popular object detection algorithms and was found to outperform them in terms of both detection accuracy and speed. The presented system, which has application potential in actual surveillance systems, shows the usefulness of deep learning approaches in vehicle detection and tracking.

Keywords: artificial intelligence, computer vision, deep learning, fast-regional convolutional neural networks, feature extraction, vehicle tracking

Procedia PDF Downloads 85
20183 Accurate Calculation of the Penetration Depth of a Bullet Using ANSYS

Authors: Eunsu Jang, Kang Park

Abstract:

In developing an armored ground combat vehicle (AGCV), it is a very important step to analyze the vulnerability (or the survivability) of the AGCV against enemy’s attack. In the vulnerability analysis, the penetration equations are usually used to get the penetration depth and check whether a bullet can penetrate the armor of the AGCV, which causes the damage of internal components or crews. The penetration equations are derived from penetration experiments which require long time and great efforts. However, they usually hold only for the specific material of the target and the specific type of the bullet used in experiments. Thus, penetration simulation using ANSYS can be another option to calculate penetration depth. However, it is very important to model the targets and select the input parameters in order to get an accurate penetration depth. This paper performed a sensitivity analysis of input parameters of ANSYS on the accuracy of the calculated penetration depth. Two conflicting objectives need to be achieved in adopting ANSYS in penetration analysis: maximizing the accuracy of calculation and minimizing the calculation time. To maximize the calculation accuracy, the sensitivity analysis of the input parameters for ANSYS was performed and calculated the RMS error with the experimental data. The input parameters include mesh size, boundary condition, material properties, target diameter are tested and selected to minimize the error between the calculated result from simulation and the experiment data from the papers on the penetration equation. To minimize the calculation time, the parameter values obtained from accuracy analysis are adjusted to get optimized overall performance. As result of analysis, the followings were found: 1) As the mesh size gradually decreases from 0.9 mm to 0.5 mm, both the penetration depth and calculation time increase. 2) As diameters of the target decrease from 250mm to 60 mm, both the penetration depth and calculation time decrease. 3) As the yield stress which is one of the material property of the target decreases, the penetration depth increases. 4) The boundary condition with the fixed side surface of the target gives more penetration depth than that with the fixed side and rear surfaces. By using above finding, the input parameters can be tuned to minimize the error between simulation and experiments. By using simulation tool, ANSYS, with delicately tuned input parameters, penetration analysis can be done on computer without actual experiments. The data of penetration experiments are usually hard to get because of security reasons and only published papers provide them in the limited target material. The next step of this research is to generalize this approach to anticipate the penetration depth by interpolating the known penetration experiments. This result may not be accurate enough to be used to replace the penetration experiments, but those simulations can be used in the early stage of the design process of AGCV in modelling and simulation stage.

Keywords: ANSYS, input parameters, penetration depth, sensitivity analysis

Procedia PDF Downloads 373
20182 Exploitation behind the Development of Home Batik Industry in Lawean, Solo, Central Java

Authors: Mukhammad Fatkhullah, Ayla Karina Budita, Cut Rizka Al Usrah, Kanita Khoirun Nisa, Muhammad Alhada Fuadilah Habib, Siti Muslihatul Mukaromah

Abstract:

Batik industry has become one of the leading industries in the economy of Indonesia. Since the recognition of batik as one of cultural wealth and national identity of Indonesia by UNESCO, batik production keeps increasing as a result of increasing demands for batik, whether from domestically or abroad. One of the rapid development batik industries in Indonesia is batik industry in Lawean Village, Solo, Central Java, Indonesia. This batik industry generally uses putting-out system where batik workers work in their own houses. With the implementation of this system, therefore employers don’t have to prepare Environmental Impact Analysis (EIA), social security for workers, overtime payment, space for working, and equipment for working. The implementation of putting-out system causes many problems, starting from environmental pollution, the loss of social rights of workers, and even exploitation of workers by batik entrepreneurs. The data used to describe this reality is the primary data from qualitative research with in-depth interview data collection technique. Informants were determined purposively. The theory used to perform data interpretation is the phenomenology of Alfred Schutz. Both qualitative and phenomenology are used in this study to describe batik workers exploitation in terms of the implementation of putting-out system on home batik industry in Lawean. The research result showed that workers in batik industry sector in Lawean were exploited with the implementation of putting-out system. The workers were strictly employed by the entrepreneurs, so that their job cannot be called 'part-time' job anymore. In terms of labor and time, the workers often work more than 12 hours per day and they often work overtime without receiving any overtime payment. In terms of work safety, the workers often have contact with chemical substances contained in batik making materials without using any protection, such as clothes work, which is worsened by the lack of standard or procedure in work that can cause physical damage, such as burnt and peeled off skin. Moreover, exposure and contamination of chemical materials make the workers and their families vulnerable to various diseases. Meanwhile, batik entrepreneurs did not give any social security (including health cost aid). Besides that, the researchers found that batik industry in home industry sector is not environmentally friendly, even damaging ecosystem because industrial waste disposed without EIA.

Keywords: exploitation, home batik industry, occupational health and safety, putting-out system

Procedia PDF Downloads 294
20181 Nelder-Mead Parametric Optimization of Elastic Metamaterials with Artificial Neural Network Surrogate Model

Authors: Jiaqi Dong, Qing-Hua Qin, Yi Xiao

Abstract:

Some of the most fundamental challenges of elastic metamaterials (EMMs) optimization can be attributed to the high consumption of computational power resulted from finite element analysis (FEA) simulations that render the optimization process inefficient. Furthermore, due to the inherent mesh dependence of FEA, minuscule geometry features, which often emerge during the later stages of optimization, induce very fine elements, resulting in enormously high time consumption, particularly when repetitive solutions are needed for computing the objective function. In this study, a surrogate modelling algorithm is developed to reduce computational time in structural optimization of EMMs. The surrogate model is constructed based on a multilayer feedforward artificial neural network (ANN) architecture, trained with prepopulated eigenfrequency data prepopulated from FEA simulation and optimized through regime selection with genetic algorithm (GA) to improve its accuracy in predicting the location and width of the primary elastic band gap. With the optimized ANN surrogate at the core, a Nelder-Mead (NM) algorithm is established and its performance inspected in comparison to the FEA solution. The ANNNM model shows remarkable accuracy in predicting the band gap width and a reduction of time consumption by 47%.

Keywords: artificial neural network, machine learning, mechanical metamaterials, Nelder-Mead optimization

Procedia PDF Downloads 110
20180 A Trapezoidal-Like Integrator for the Numerical Solution of One-Dimensional Time Dependent Schrödinger Equation

Authors: Johnson Oladele Fatokun, I. P. Akpan

Abstract:

In this paper, the one-dimensional time dependent Schrödinger equation is discretized by the method of lines using a second order finite difference approximation to replace the second order spatial derivative. The evolving system of stiff ordinary differential equation (ODE) in time is solved numerically by an L-stable trapezoidal-like integrator. Results show accuracy of relative maximum error of order 10-4 in the interval of consideration. The performance of the method as compared to an existing scheme is considered favorable.

Keywords: Schrodinger’s equation, partial differential equations, method of lines (MOL), stiff ODE, trapezoidal-like integrator

Procedia PDF Downloads 399
20179 Real-Time Sensor Fusion for Mobile Robot Localization in an Oil and Gas Refinery

Authors: Adewole A. Ayoade, Marshall R. Sweatt, John P. H. Steele, Qi Han, Khaled Al-Wahedi, Hamad Karki, William A. Yearsley

Abstract:

Understanding the behavioral characteristics of sensors is a crucial step in fusing data from several sensors of different types. This paper introduces a practical, real-time approach to integrate heterogeneous sensor data to achieve higher accuracy than would be possible from any one individual sensor in localizing a mobile robot. We use this approach in both indoor and outdoor environments and it is especially appropriate for those environments like oil and gas refineries due to their sparse and featureless nature. We have studied the individual contribution of each sensor data to the overall combined accuracy achieved from the fusion process. A Sequential Update Extended Kalman Filter(EKF) using validation gates was used to integrate GPS data, Compass data, WiFi data, Inertial Measurement Unit(IMU) data, Vehicle Velocity, and pose estimates from Fiducial marker system. Results show that the approach can enable a mobile robot to navigate autonomously in any environment using a priori information.

Keywords: inspection mobile robot, navigation, sensor fusion, sequential update extended Kalman filter

Procedia PDF Downloads 448
20178 Estimation of Train Operation Using an Exponential Smoothing Method

Authors: Taiyo Matsumura, Kuninori Takahashi, Takashi Ono

Abstract:

The purpose of this research is to improve the convenience of waiting for trains at level crossings and stations and to prevent accidents resulting from forcible entry into level crossings, by providing level crossing users and passengers with information that tells them when the next train will pass through or arrive. For this paper, we proposed methods for estimating operation by means of an average value method, variable response smoothing method, and exponential smoothing method, on the basis of open data, which has low accuracy, but for which performance schedules are distributed in real time. We then examined the accuracy of the estimations. The results showed that the application of an exponential smoothing method is valid.

Keywords: exponential smoothing method, open data, operation estimation, train schedule

Procedia PDF Downloads 366
20177 Automated Tracking and Statistics of Vehicles at the Signalized Intersection

Authors: Qiang Zhang, Xiaojian Hu1

Abstract:

Intersection is the place where vehicles and pedestrians must pass through, turn and evacuate. Obtaining the motion data of vehicles near the intersection is of great significance for transportation research. Since there are usually many targets and there are more conflicts between targets, this makes it difficult to obtain vehicle motion parameters in traffic videos of intersections. According to the characteristics of traffic videos, this paper applies video technology to realize the automated track, count and trajectory extraction of vehicles to collect traffic data by roadside surveillance cameras installed near the intersections. Based on the video recognition method, the vehicles in each lane near the intersection are tracked with extracting trajectory and counted respectively in various degrees of occlusion and visibility. The performances are compared with current recognized CPU-based algorithms of real-time tracking-by-detection. The speed of the presented system is higher than the others and the system has a better real-time performance. The accuracy of direction has reached about 94.99% on average, and the accuracy of classification and statistics has reached about 75.12% on average.

Keywords: tracking and statistics, vehicle, signalized intersection, motion parameter, trajectory

Procedia PDF Downloads 198
20176 The Effect of the Construction Contract System by Simulating the Comparative Costs of Capital to the Financial Feasibility of the Construction of Toll Bali Mandara

Authors: Mas Pertiwi I. G. AG Istri, Sri Kristinayanti Wayan, Oka Aryawan I. Gede Made

Abstract:

Ability of government to meet the needs of infrastructure investment constrained by the size of the budget commitments for other sectors. Another barrier is the complexity of the process of land acquisition. Public Private Partnership can help bridge the investment gap by including the amount of funding from the private sector, shifted the responsibility of financing, construction of the asset, and the operation and post-project design and care to them. In principle, a construction project implementation always requires the investor as a party to provide resources in the form of funding which it must be contained in a successor agreement in the form of a contract. In general, construction contracts consist of contracts which passed in Indonesia and contract International. One source of funding used in the implementation of construction projects comes from funding that comes from the collaboration between the government and the private sector, for example with the system: BLT (Build Lease Transfer), BOT (Build Operate Transfer), BTO (Build Transfer Operate) and BOO (Build Operate Own). And form of payment under a construction contract can be distinguished several ways: monthly payment, payments based on progress and payment after completed projects (Turn Key). One of the tools used to analyze the feasibility of the investment is to use financial models. The financial model describes the relationship between different variables and assumptions used. From a financial model will be known how the cash flow structure of the project, which includes revenues, expenses, liabilities to creditors and the payment of taxes to the government. Net cash flow generated from the project will be used as a basis for analyzing the feasibility of investment source of project financing Public Private Partnership could come from equity or debt. The proportion of funding according to its source is a comparison of a number of investment funds originating from each source of financing for a total investment cost during the construction period by selected the contract system and several alternative financing percentage ratio determined according to sources will generate cash flow structure that is different. Of the various possibilities for the structure of the cash flow generated will be analyzed by software is to test T Paired to compared the contract system used by various alternatives comparison of financing to determine the effect of the contract system and the comparison of such financing for the feasibility of investment toll road construction project for the economic life of 20 (twenty) years. In this use case studies of toll road contruction project Bali Mandara. And in this analysis only covered two systems contracts, namely Build Operate Transfer and Turn Key. Based on the results obtained by analysis of the variable investment feasibility of the NPV, BCR and IRR between the contract system Build Operate Transfer and contract system Turn Key on the interest rate of 9%, 12% and 15%.

Keywords: contract system, financing, internal rate of return, net present value

Procedia PDF Downloads 210
20175 Speaker Identification by Atomic Decomposition of Learned Features Using Computational Auditory Scene Analysis Principals in Noisy Environments

Authors: Thomas Bryan, Veton Kepuska, Ivica Kostanic

Abstract:

Speaker recognition is performed in high Additive White Gaussian Noise (AWGN) environments using principals of Computational Auditory Scene Analysis (CASA). CASA methods often classify sounds from images in the time-frequency (T-F) plane using spectrograms or cochleargrams as the image. In this paper atomic decomposition implemented by matching pursuit performs a transform from time series speech signals to the T-F plane. The atomic decomposition creates a sparsely populated T-F vector in “weight space” where each populated T-F position contains an amplitude weight. The weight space vector along with the atomic dictionary represents a denoised, compressed version of the original signal. The arraignment or of the atomic indices in the T-F vector are used for classification. Unsupervised feature learning implemented by a sparse autoencoder learns a single dictionary of basis features from a collection of envelope samples from all speakers. The approach is demonstrated using pairs of speakers from the TIMIT data set. Pairs of speakers are selected randomly from a single district. Each speak has 10 sentences. Two are used for training and 8 for testing. Atomic index probabilities are created for each training sentence and also for each test sentence. Classification is performed by finding the lowest Euclidean distance between then probabilities from the training sentences and the test sentences. Training is done at a 30dB Signal-to-Noise Ratio (SNR). Testing is performed at SNR’s of 0 dB, 5 dB, 10 dB and 30dB. The algorithm has a baseline classification accuracy of ~93% averaged over 10 pairs of speakers from the TIMIT data set. The baseline accuracy is attributable to short sequences of training and test data as well as the overall simplicity of the classification algorithm. The accuracy is not affected by AWGN and produces ~93% accuracy at 0dB SNR.

Keywords: time-frequency plane, atomic decomposition, envelope sampling, Gabor atoms, matching pursuit, sparse dictionary learning, sparse autoencoder

Procedia PDF Downloads 276
20174 A Unified Fitting Method for the Set of Unified Constitutive Equations for Modelling Microstructure Evolution in Hot Deformation

Authors: Chi Zhang, Jun Jiang

Abstract:

Constitutive equations are very important in finite element (FE) modeling, and the accuracy of the material constants in the equations have significant effects on the accuracy of the FE models. A wide range of constitutive equations are available; however, fitting the material constants in the constitutive equations could be complex and time-consuming due to the strong non-linearity and relationship between the constants. This work will focus on the development of a set of unified MATLAB programs for fitting the material constants in the constitutive equations efficiently. Users will only need to supply experimental data in the required format and run the program without modifying functions or precisely guessing the initial values, or finding the parameters in previous works and will be able to fit the material constants efficiently.

Keywords: constitutive equations, FE modelling, MATLAB program, non-linear curve fitting

Procedia PDF Downloads 75
20173 The Decision-Making Mechanisms of Tax Regulations

Authors: Nino Pailodze, Malkhaz Sulashvili, Vladimer Kekenadze, Tea Khutsishvili, Irma Makharashvili, Aleksandre Kekenadze

Abstract:

In the nearest future among the important problems which Georgia has solve the most important is economic stability, that bases on fiscal policy and the proper definition of the its directions. The main source of the Budget revenue is the national income. The State uses taxes, loans and emission in order to create national income, were the principal weapon are taxes. As well as fiscal function of the fulfillment of the budget, tax systems successfully implement economic and social development and the regulatory functions of foreign economic relations. A tax is a mandatory, unconditional monetary payment to the budget made by a taxpayer in accordance with this Code, based on the necessary, nonequivalent and gratuitous character of the payment. Taxes shall be national and local. National taxes shall be the taxes provided for under this Code, the payment of which is mandatory across the whole territory of Georgia. Local taxes shall be the taxes provided for under this Code, introduced by normative acts of local self-government representative authorities (within marginal rates), the payment of which is mandatory within the territory of the relevant self-governing unit. National taxes have the leading role in tax systems, but also the local taxes have an importance role in tax systems. Exactly in the means of local taxes, the most part of the budget is formatted. National taxes shall be: income tax, profit tax, value added tax (VAT), excise tax, import duty, property tax shall be a local tax The property tax is one of the significant taxes in Georgia. The paper deals with the taxation mechanism that has been operated in Georgia. The above mention has the great influence in financial accounting. While comparing foreign legislation towards Georgian legislation we discuss the opportunity of using their experience. Also, we suggested recommendations in order to improve the tax system in financial accounting. In addition to accounting, which is regulated according the International Accounting Standards we have tax accounting, which is regulated by the Tax Code, various legal orders / regulations of the Minister of Finance. The rules are controlled by the tax authority, Revenue Service. The tax burden from the tax values are directly related to expenditures of the state from the emergence of the first day. Fiscal policy of the state is as well as expenditure of the state and decisions of taxation. In order to get the best and the most effective mobilization of funds, Government’s primary task is to decide the kind of taxation rules. Tax function is to reveal the substance of the act. Taxes have the following functions: distribution or the fiscal function; Control and regulatory functions. Foreign tax systems evolved in the different economic, political and social conditions influence. The tax systems differ greatly from each other: taxes, their structure, typing means, rates, the different levels of fiscal authority, the tax base, the tax sphere of action, the tax breaks.

Keywords: international accounting standards, financial accounting, tax systems, financial obligations

Procedia PDF Downloads 220
20172 Analyzing Time Lag in Seismic Waves and Its Effects on Isolated Structures

Authors: Faizan Ahmad, Jenna Wong

Abstract:

Time lag between peak values of horizontal and vertical seismic waves is a well-known phenomenon. Horizontal and vertical seismic waves, secondary and primary waves in nature respectively, travel through different layers of soil and the travel time is dependent upon the medium of wave transmission. In seismic analysis, many standardized codes do not require the actual vertical acceleration to be part of the analysis procedure. Instead, a factor load addition for a particular site is used to capture strength demands in case of vertical excitation. This study reviews the effects of vertical accelerations to analyze the behavior of a linearly rubber isolated structure in different time lag situations and frequency content by application of historical and simulated ground motions using SAP2000. The response of the structure is reviewed under multiple sets of ground motions and trends based on time lag and frequency variations are drawn. The accuracy of these results is discussed and evaluated to provide reasoning for use of real vertical excitations in seismic analysis procedures, especially for isolated structures.

Keywords: seismic analysis, vertical accelerations, time lag, isolated structures

Procedia PDF Downloads 314
20171 Neural Network-based Risk Detection for Dyslexia and Dysgraphia in Sinhala Language Speaking Children

Authors: Budhvin T. Withana, Sulochana Rupasinghe

Abstract:

The problem of Dyslexia and Dysgraphia, two learning disabilities that affect reading and writing abilities, respectively, is a major concern for the educational system. Due to the complexity and uniqueness of the Sinhala language, these conditions are especially difficult for children who speak it. The traditional risk detection methods for Dyslexia and Dysgraphia frequently rely on subjective assessments, making it difficult to cover a wide range of risk detection and time-consuming. As a result, diagnoses may be delayed and opportunities for early intervention may be lost. The project was approached by developing a hybrid model that utilized various deep learning techniques for detecting risk of Dyslexia and Dysgraphia. Specifically, Resnet50, VGG16 and YOLOv8 were integrated to detect the handwriting issues, and their outputs were fed into an MLP model along with several other input data. The hyperparameters of the MLP model were fine-tuned using Grid Search CV, which allowed for the optimal values to be identified for the model. This approach proved to be effective in accurately predicting the risk of Dyslexia and Dysgraphia, providing a valuable tool for early detection and intervention of these conditions. The Resnet50 model achieved an accuracy of 0.9804 on the training data and 0.9653 on the validation data. The VGG16 model achieved an accuracy of 0.9991 on the training data and 0.9891 on the validation data. The MLP model achieved an impressive training accuracy of 0.99918 and a testing accuracy of 0.99223, with a loss of 0.01371. These results demonstrate that the proposed hybrid model achieved a high level of accuracy in predicting the risk of Dyslexia and Dysgraphia.

Keywords: neural networks, risk detection system, Dyslexia, Dysgraphia, deep learning, learning disabilities, data science

Procedia PDF Downloads 67
20170 Using Deep Learning Real-Time Object Detection Convolution Neural Networks for Fast Fruit Recognition in the Tree

Authors: K. Bresilla, L. Manfrini, B. Morandi, A. Boini, G. Perulli, L. C. Grappadelli

Abstract:

Image/video processing for fruit in the tree using hard-coded feature extraction algorithms have shown high accuracy during recent years. While accurate, these approaches even with high-end hardware are computationally intensive and too slow for real-time systems. This paper details the use of deep convolution neural networks (CNNs), specifically an algorithm (YOLO - You Only Look Once) with 24+2 convolution layers. Using deep-learning techniques eliminated the need for hard-code specific features for specific fruit shapes, color and/or other attributes. This CNN is trained on more than 5000 images of apple and pear fruits on 960 cores GPU (Graphical Processing Unit). Testing set showed an accuracy of 90%. After this, trained data were transferred to an embedded device (Raspberry Pi gen.3) with camera for more portability. Based on correlation between number of visible fruits or detected fruits on one frame and the real number of fruits on one tree, a model was created to accommodate this error rate. Speed of processing and detection of the whole platform was higher than 40 frames per second. This speed is fast enough for any grasping/harvesting robotic arm or other real-time applications.

Keywords: artificial intelligence, computer vision, deep learning, fruit recognition, harvesting robot, precision agriculture

Procedia PDF Downloads 392
20169 The Pricing-Out Phenomenon in the U.S. Housing Market

Authors: Francesco Berald, Yunhui Zhao

Abstract:

The COVID-19 pandemic further extended the multi-year housing boom in advanced economies and emerging markets alike against massive monetary easing during the pandemic. In this paper, we analyze the pricing-out phenomenon in the U.S. residential housing market due to higher house prices associated with monetary easing. We first set up a stylized general equilibrium model and show that although monetary easing decreases the mortgage payment burden, it would raise house prices and lower housing affordability for first-time homebuyers (through the initial housing wealth channel and the liquidity constraint channel that increases repeat buyers’ housing demand), and increase housing wealth inequality between first-time and repeat homebuyers. We then use the U.S. household-level data to quantify the effect of the house price change on housing affordability relative to that of the interest rate change. We find evidence of the pricing-out effect for all homebuyers; moreover, we find that the pricing-out effect is stronger for first-time homebuyers than for repeat homebuyers. The paper highlights the importance of accounting for general equilibrium effects and distributional implications of monetary policy while assessing housing affordability. It also calls for complementing monetary easing with well-targeted policy measures that can boost housing affordability, particularly for first-time and lower-income households. Such measures are also needed during aggressive monetary tightening, given that the fall in house prices may be insufficient or too slow to fully offset the immediate adverse impact of higher rates on housing affordability.

Keywords: pricing-out, U.S. housing market, housing affordability, distributional effects, monetary policy

Procedia PDF Downloads 12
20168 Efficiency of Using E-Wallets as Payment Method in Marikina City During COVID-19 Pandemic

Authors: Noel Paolo Domingo, James Paul Menina, Laurente Ferrer

Abstract:

Most people were forced to stay at home and limit their physical contact during the COVID-19 pandemic. Due to the situation, strict implementation of government policies and safety protocols encouraged consumers to utilize cashless or digital transactions through e-wallets. In this study, the researchers aim to investigate the efficiency of using e-wallets as a payment method during the COVID-19 pandemic in Marikina City. The study examined the efficiency of e-wallets in terms of Usefulness, Convenience, and Safety and Security based on respondents’ assessment. Questionnaires developed by the researchers were distributed to a total of 400 e-wallet users in Marikina City aged 15 years old and above to gather data by using a purposive sampling technique. The data collected was processed using SPSS version 26. Frequency, percentage, and mean were utilized to describe the profile of respondents and their assessment of e-wallets in terms of the three constructs. ANOVA and t-tests were also employed to test the significant differences in the respondent’s assessment when the demographic profile was considered. The study revealed that when it comes to usefulness, e-wallet is efficient while in terms of convenience, and safety and security, e-wallet has been proven to be very efficient. During the COVID-19 pandemic, utilizing e-wallets has been embraced by most consumers. By enhancing its features, more people will be satisfied with using e-wallets.

Keywords: efficiency of e-wallets, usefulness, convenience, safety and security

Procedia PDF Downloads 98
20167 Creep Analysis and Rupture Evaluation of High Temperature Materials

Authors: Yuexi Xiong, Jingwu He

Abstract:

The structural components in an energy facility such as steam turbine machines are operated under high stress and elevated temperature in an endured time period and thus the creep deformation and creep rupture failure are important issues that need to be addressed in the design of such components. There are numerous creep models being used for creep analysis that have both advantages and disadvantages in terms of accuracy and efficiency. The Isochronous Creep Analysis is one of the simplified approaches in which a full-time dependent creep analysis is avoided and instead an elastic-plastic analysis is conducted at each time point. This approach has been established based on the rupture dependent creep equations using the well-known Larson-Miller parameter. In this paper, some fundamental aspects of creep deformation and the rupture dependent creep models are reviewed and the analysis procedures using isochronous creep curves are discussed. Four rupture failure criteria are examined from creep fundamental perspectives including criteria of Stress Damage, Strain Damage, Strain Rate Damage, and Strain Capability. The accuracy of these criteria in predicting creep life is discussed and applications of the creep analysis procedures and failure predictions of simple models will be presented. In addition, a new failure criterion is proposed to improve the accuracy and effectiveness of the existing criteria. Comparisons are made between the existing criteria and the new one using several examples materials. Both strain increase and stress relaxation form a full picture of the creep behaviour of a material under high temperature in an endured time period. It is important to bear this in mind when dealing with creep problems. Accordingly there are two sets of rupture dependent creep equations. While the rupture strength vs LMP equation shows how the rupture time depends on the stress level under load controlled condition, the strain rate vs rupture time equation reflects how the rupture time behaves under strain-controlled condition. Among the four existing failure criteria for rupture life predictions, the Stress Damage and Strain Damage Criteria provide the most conservative and non-conservative predictions, respectively. The Strain Rate and Strain Capability Criteria provide predictions in between that are believed to be more accurate because the strain rate and strain capability are more determined quantities than stress to reflect the creep rupture behaviour. A modified Strain Capability Criterion is proposed making use of the two sets of creep equations and therefore is considered to be more accurate than the original Strain Capability Criterion.

Keywords: creep analysis, high temperature mateials, rapture evalution, steam turbine machines

Procedia PDF Downloads 264
20166 Using Machine Learning to Classify Different Body Parts and Determine Healthiness

Authors: Zachary Pan

Abstract:

Our general mission is to solve the problem of classifying images into different body part types and deciding if each of them is healthy or not. However, for now, we will determine healthiness for only one-sixth of the body parts, specifically the chest. We will detect pneumonia in X-ray scans of those chest images. With this type of AI, doctors can use it as a second opinion when they are taking CT or X-ray scans of their patients. Another ad-vantage of using this machine learning classifier is that it has no human weaknesses like fatigue. The overall ap-proach to this problem is to split the problem into two parts: first, classify the image, then determine if it is healthy. In order to classify the image into a specific body part class, the body parts dataset must be split into test and training sets. We can then use many models, like neural networks or logistic regression models, and fit them using the training set. Now, using the test set, we can obtain a realistic accuracy the models will have on images in the real world since these testing images have never been seen by the models before. In order to increase this testing accuracy, we can also apply many complex algorithms to the models, like multiplicative weight update. For the second part of the problem, to determine if the body part is healthy, we can have another dataset consisting of healthy and non-healthy images of the specific body part and once again split that into the test and training sets. We then use another neural network to train on those training set images and use the testing set to figure out its accuracy. We will do this process only for the chest images. A major conclusion reached is that convolutional neural networks are the most reliable and accurate at image classification. In classifying the images, the logistic regression model, the neural network, neural networks with multiplicative weight update, neural networks with the black box algorithm, and the convolutional neural network achieved 96.83 percent accuracy, 97.33 percent accuracy, 97.83 percent accuracy, 96.67 percent accuracy, and 98.83 percent accuracy, respectively. On the other hand, the overall accuracy of the model that de-termines if the images are healthy or not is around 78.37 percent accuracy.

Keywords: body part, healthcare, machine learning, neural networks

Procedia PDF Downloads 76
20165 Radar-Based Classification of Pedestrian and Dog Using High-Resolution Raw Range-Doppler Signatures

Authors: C. Mayr, J. Periya, A. Kariminezhad

Abstract:

In this paper, we developed a learning framework for the classification of vulnerable road users (VRU) by their range-Doppler signatures. The frequency-modulated continuous-wave (FMCW) radar raw data is first pre-processed to obtain robust object range-Doppler maps per coherent time interval. The complex-valued range-Doppler maps captured from our outdoor measurements are further fed into a convolutional neural network (CNN) to learn the classification. This CNN has gone through a hyperparameter optimization process for improved learning. By learning VRU range-Doppler signatures, the three classes 'pedestrian', 'dog', and 'noise' are classified with an average accuracy of almost 95%. Interestingly, this classification accuracy holds for a combined longitudinal and lateral object trajectories.

Keywords: machine learning, radar, signal processing, autonomous driving

Procedia PDF Downloads 217
20164 Discussion as a Means to Improve Peer Assessment Accuracy

Authors: Jung Ae Park, Jooyong Park

Abstract:

Writing is an important learning activity that cultivates higher level thinking. Effective and immediate feedback is necessary to help improve students' writing skills. Peer assessment can be an effective method in writing tasks because it makes it possible for students not only to receive quick feedback on their writing but also to get a chance to examine different perspectives on the same topic. Peer assessment can be practiced frequently and has the advantage of immediate feedback. However, there is controversy about the accuracy of peer assessment. In this study, we tried to demonstrate experimentally how the accuracy of peer assessment could be improved. Participants (n=76) were randomly assigned to groups of 4 members. All the participant graded two sets of 4 essays on the same topic. They graded the first set twice, and the second set or the posttest once. After the first grading of the first set, each group in the experimental condition 1 (discussion group), were asked to discuss the results of the peer assessment and then to grade the essays again. Each group in the experimental condition 2 (reading group), were asked to read the assessment on each essay by an expert and then to grade the essays again. In the control group, the participants were asked to grade the 4 essays twice in different orders. Afterwards, all the participants graded the second set of 4 essays. The mean score from 4 participants was calculated for each essay. The accuracy of the peer assessment was measured by Pearson correlation with the scores of the expert. The results were analyzed by two-way repeated measure ANOVA. The main effect of grading was observed: Grading accuracy got better as the number of grading experience increased. Analysis of posttest accuracy revealed that the score variations within a group of 4 participants decreased in both discussion and reading conditions but not in the control condition. These results suggest that having students discuss their grading together can be an efficient means to improve peer assessment accuracy. By discussing, students can learn from others about what to consider in grading and whether their grading is too strict or lenient. Further research is needed to examine the exact cause of the grading accuracy.

Keywords: peer assessment, evaluation accuracy, discussion, score variations

Procedia PDF Downloads 251
20163 Application of a SubIval Numerical Solver for Fractional Circuits

Authors: Marcin Sowa

Abstract:

The paper discusses the subinterval-based numerical method for fractional derivative computations. It is now referred to by its acronym – SubIval. The basis of the method is briefly recalled. The ability of the method to be applied in time stepping solvers is discussed. The possibility of implementing a time step size adaptive solver is also mentioned. The solver is tested on a transient circuit example. In order to display the accuracy of the solver – the results have been compared with those obtained by means of a semi-analytical method called gcdAlpha. The time step size adaptive solver applying SubIval has been proven to be very accurate as the results are very close to the referential solution. The solver is currently able to solve FDE (fractional differential equations) with various derivative orders for each equation and any type of source time functions.

Keywords: numerical method, SubIval, fractional calculus, numerical solver, circuit analysis

Procedia PDF Downloads 184
20162 Developing a Hybrid Method to Diagnose and Predict Sports Related Concussions with Machine Learning

Authors: Melody Yin

Abstract:

Concussions impact a large amount of adolescents; they make up as much as half of the diagnosed concussions in America. This research proposes a hybrid machine learning model based on the combination of human/knowledge-based domains and computer-generated feature rankings to improve the accuracy of diagnosing sports related concussion (SRC). Using a data set of symptoms collected on the sideline post-SRC events, the symptom selection criteria method has been developed by using Google AutoML's important score function to identify the top 10 symptom features. In addition, symptom domains have been introduced as another parameter, categorizing the symptoms into physical, cognitive, sleep, and emotional domains. The hybrid machine learning model has been trained with a combination of the top 10 symptoms and 4 domains. From the results, the hybrid model was the best performer for symptom resolution time prediction in 2 and 4-week thresholds. This research is a proof of concept study in the use of domains along with machine learning in order to improve concussion prediction accuracy. It is also possible that the use of domains can make the model more efficient due to reduced training time. This research examines the use of a hybrid method in predicting sports-related concussion. This achievement is based on data preprocessing, using a hybrid method to select criteria to achieve high performance.

Keywords: hybrid model, machine learning, sports related concussion, symptom resolution time

Procedia PDF Downloads 150