Search results for: error minimization
582 Sequence Analysis and Structural Implications of Rotavirus Capsid Proteins
Authors: Nishal Parbhoo, John B. Dewar, Samantha Gildenhuys
Abstract:
Rotavirus is the major cause of severe gastroenteritis worldwide in children aged 5 and younger. Death rates are high particularly in developing countries. The mature rotavirus is a non-enveloped triple-layered nucleocapsid containing 11 double-stranded RNA segments. Here a global view on the sequence and structure of the three main capsid proteins, VP7, VP6, and VP2 is taken by generating a consensus sequence for each of these rotavirus proteins, for each species obtained from published data of representative rotavirus genotypes from across the world and across species. The degree of conservation between species was represented on homology models for each of the proteins. VP7 shows the highest level of variation with 14 - 45 amino acids showing conservation of less than 60%. These changes are localized to the outer surface which is exposed to antibodies alluding to a possible mechanism in evading the immune system. The middle layer, VP6 shows lower variability with only 14-32 sites having lower than 70% conservation. The inner structural layer made up of VP2 showed the lowest variability with only 1-16 sites having less than 70% conservation across species. The results correlate with proteins’ multiple structural roles. Although the nucleotide sequences vary due to an error-prone replication and lack of proofreading, the corresponding amino acid sequence of VP2, 6 and 7 remains conserved. Sequence conservation maintained for the virus results in stable protein structures, fit for function. This can be exploited in drug design, molecular studies and biotechnological applications.Keywords: amino acid sequence conservation, capsid protein, protein structure, vaccine candidate
Procedia PDF Downloads 291581 Estimation of Normalized Glandular Doses Using a Three-Layer Mammographic Phantom
Authors: Kuan-Jen Lai, Fang-Yi Lin, Shang-Rong Huang, Yun-Zheng Zeng, Po-Chieh Hsu, Jay Wu
Abstract:
The normalized glandular dose (DgN) estimates the energy deposition of mammography in clinical practice. The Monte Carlo simulations frequently use uniformly mixed phantom for calculating the conversion factor. However, breast tissues are not uniformly distributed, leading to errors of conversion factor estimation. This study constructed a three-layer phantom to estimated more accurate of normalized glandular dose. In this study, MCNP code (Monte Carlo N-Particles code) was used to create the geometric structure. We simulated three types of target/filter combinations (Mo/Mo, Mo/Rh, Rh/Rh), six voltages (25 ~ 35 kVp), six HVL parameters and nine breast phantom thicknesses (2 ~ 10 cm) for the three-layer mammographic phantom. The conversion factor for 25%, 50% and 75% glandularity was calculated. The error of conversion factors compared with the results of the American College of Radiology (ACR) was within 6%. For Rh/Rh, the difference was within 9%. The difference between the 50% average glandularity and the uniform phantom was 7.1% ~ -6.7% for the Mo/Mo combination, voltage of 27 kVp, half value layer of 0.34 mmAl, and breast thickness of 4 cm. According to the simulation results, the regression analysis found that the three-layer mammographic phantom at 0% ~ 100% glandularity can be used to accurately calculate the conversion factors. The difference in glandular tissue distribution leads to errors of conversion factor calculation. The three-layer mammographic phantom can provide accurate estimates of glandular dose in clinical practice.Keywords: Monte Carlo simulation, mammography, normalized glandular dose, glandularity
Procedia PDF Downloads 191580 The Effect of Macroeconomic Policies on Cambodia's Economy: ARDL and VECM Model
Authors: Siphat Lim
Abstract:
This study used Autoregressive Distributed Lag (ARDL) approach to cointegration. In the long-run the general price level and exchange rate have a positively significant effect on domestic output. The estimated result further revealed that fiscal stimulus help stimulate domestic output in the long-run, but not in the short-run, while monetary expansion help to stimulate output in both short-run and long-run. The result is complied with the theory which is the macroeconomic policies, fiscal and monetary policy; help to stimulate domestic output in the long-run. The estimated result of the Vector Error Correction Model (VECM) has indicated more clearly that the consumer price index has a positive effect on output with highly statistically significant. Increasing in the general price level would increase the competitiveness among producers than increase in the output. However, the exchange rate also has a positive effect and highly significant on the gross domestic product. The exchange rate depreciation might increase export since the purchasing power of foreigners has increased. More importantly, fiscal stimulus would help stimulate the domestic output in the long-run since the coefficient of government expenditure is positive. In addition, monetary expansion would also help stimulate the output and the result is highly significant. Thus, fiscal stimulus and monetary expansionary would help stimulate the domestic output in the long-run in Cambodia.Keywords: fiscal policy, monetary policy, ARDL, VECM
Procedia PDF Downloads 434579 Framework for Decision Support Tool for Quality Control and Management in Botswana Manufacturing Companies
Authors: Mogale Sabone, Thabiso Ntlole
Abstract:
The pressure from globalization has made manufacturing organizations to move towards three major competitive arenas: quality, cost, and responsiveness. Quality is a universal value and has become a global issue. In order to survive and be able to provide customers with good products, manufacturing organizations’ supporting systems, tools, and structures it uses must grow or evolve. The majority of quality management concepts and strategies that are practiced recently are aimed at detecting and correcting problems which already exist and serve to limit losses. In agile manufacturing environment there is no room for defect and error so it needs a quality management which is proactively directed at problem prevention. This proactive quality management avoids losses by focusing on failure prevention, virtual elimination of the possibility of premature failure, mistake-proofing, and assuring consistently high quality in the definition and design of creation processes. To achieve this, a decision support tool for quality control and management is suggested. Current decision support tools/methods used by most manufacturing companies in Botswana for quality management and control are not integrated, for example they are not consistent since some tests results data is recorded manually only whilst others are recorded electronically. It is only a set of procedures not a tool. These procedures cannot offer interactive decision support. This point brings to light the aim of this research which is to develop a framework which will help manufacturing companies in Botswana build a decision support tool for quality control and management.Keywords: decision support tool, manufacturing, quality control, quality management
Procedia PDF Downloads 568578 Artificial Neural Network to Predict the Optimum Performance of Air Conditioners under Environmental Conditions in Saudi Arabia
Authors: Amr Sadek, Abdelrahaman Al-Qahtany, Turkey Salem Al-Qahtany
Abstract:
In this study, a backpropagation artificial neural network (ANN) model has been used to predict the cooling and heating capacities of air conditioners (AC) under different conditions. Sufficiently large measurement results were obtained from the national energy-efficiency laboratories in Saudi Arabia and were used for the learning process of the ANN model. The parameters affecting the performance of the AC, including temperature, humidity level, specific heat enthalpy indoors and outdoors, and the air volume flow rate of indoor units, have been considered. These parameters were used as inputs for the ANN model, while the cooling and heating capacity values were set as the targets. A backpropagation ANN model with two hidden layers and one output layer could successfully correlate the input parameters with the targets. The characteristics of the ANN model including the input-processing, transfer, neurons-distance, topology, and training functions have been discussed. The performance of the ANN model was monitored over the training epochs and assessed using the mean squared error function. The model was then used to predict the performance of the AC under conditions that were not included in the measurement results. The optimum performance of the AC was also predicted under the different environmental conditions in Saudi Arabia. The uncertainty of the ANN model predictions has been evaluated taking into account the randomness of the data and lack of learning.Keywords: artificial neural network, uncertainty of model predictions, efficiency of air conditioners, cooling and heating capacities
Procedia PDF Downloads 78577 Ice Load Measurements on Known Structures Using Image Processing Methods
Authors: Azam Fazelpour, Saeed R. Dehghani, Vlastimil Masek, Yuri S. Muzychka
Abstract:
This study employs a method based on image analyses and structure information to detect accumulated ice on known structures. The icing of marine vessels and offshore structures causes significant reductions in their efficiency and creates unsafe working conditions. Image processing methods are used to measure ice loads automatically. Most image processing methods are developed based on captured image analyses. In this method, ice loads on structures are calculated by defining structure coordinates and processing captured images. A pyramidal structure is designed with nine cylindrical bars as the known structure of experimental setup. Unsymmetrical ice accumulated on the structure in a cold room represents the actual case of experiments. Camera intrinsic and extrinsic parameters are used to define structure coordinates in the image coordinate system according to the camera location and angle. The thresholding method is applied to capture images and detect iced structures in a binary image. The ice thickness of each element is calculated by combining the information from the binary image and the structure coordinate. Averaging ice diameters from different camera views obtains ice thicknesses of structure elements. Comparison between ice load measurements using this method and the actual ice loads shows positive correlations with an acceptable range of error. The method can be applied to complex structures defining structure and camera coordinates.Keywords: camera calibration, ice detection, ice load measurements, image processing
Procedia PDF Downloads 369576 State Estimator Performance Enhancement: Methods for Identifying Errors in Modelling and Telemetry
Authors: M. Ananthakrishnan, Sunil K Patil, Koti Naveen, Inuganti Hemanth Kumar
Abstract:
State estimation output of EMS forms the base case for all other advanced applications used in real time by a power system operator. Ensuring tuning of state estimator is a repeated process and cannot be left once a good solution is obtained. This paper attempts to demonstrate methods to improve state estimator solution by identifying incorrect modelling and telemetry inputs to the application. In this work, identification of database topology modelling error by plotting static network using node-to-node connection details is demonstrated with examples. Analytical methods to identify wrong transmission parameters, incorrect limits and mistakes in pseudo load and generator modelling are explained with various cases observed. Further, methods used for active and reactive power tuning using bus summation display, reactive power absorption summary, and transformer tap correction are also described. In a large power system, verifying all network static data and modelling parameter on regular basis is difficult .The proposed tuning methods can be easily used by operators to quickly identify errors to obtain the best possible state estimation performance. This, in turn, can lead to improved decision-support capabilities, ultimately enhancing the safety and reliability of the power grid.Keywords: active power tuning, database modelling, reactive power, state estimator
Procedia PDF Downloads 14575 Symbolic Partial Differential Equations Analysis Using Mathematica
Authors: Davit Shahnazaryan, Diogo Gomes, Mher Safaryan
Abstract:
Many symbolic computations and manipulations required in the analysis of partial differential equations (PDE) or systems of PDEs are tedious and error-prone. These computations arise when determining conservation laws, entropies or integral identities, which are essential tools for the study of PDEs. Here, we discuss a new Mathematica package for the symbolic analysis of PDEs that automate multiple tasks, saving time and effort. Methodologies: During the research, we have used concepts of linear algebra and partial differential equations. We have been working on creating algorithms based on theoretical mathematics to find results mentioned below. Major Findings: Our package provides the following functionalities; finding symmetry group of different PDE systems, generation of polynomials invariant with respect to different symmetry groups; simplification of integral quantities by integration by parts and null Lagrangian cleaning, computing general forms of expressions by integration by parts; finding equivalent forms of an integral expression that are simpler or more symmetric form; determining necessary and sufficient conditions on the coefficients for the positivity of a given symbolic expression. Conclusion: Using this package, we can simplify integral identities, find conserved and dissipated quantities of time-dependent PDE or system of PDEs. Some examples in the theory of mean-field games and semiconductor equations are discussed.Keywords: partial differential equations, symbolic computation, conserved and dissipated quantities, mathematica
Procedia PDF Downloads 164574 Monthly Labor Forces Surveys Portray Smooth Labor Markets and Bias Fixed Effects Estimation: Evidence from Israel’s Transition from Quarterly to Monthly Surveys
Authors: Haggay Etkes
Abstract:
This study provides evidence for the impact of monthly interviews conducted for the Israeli Labor Force Surveys (LFSs) on estimated flows between labor force (LF) statuses and on coefficients in fixed-effects estimations. The study uses the natural experiment of parallel interviews for the quarterly and the monthly LFSs in Israel in 2011 for demonstrating that the Labor Force Participation (LFP) rate of Jewish persons who participated in the monthly LFS increased between interviews, while in the quarterly LFS it decreased. Interestingly, the estimated impact on the LFP rate of self-reporting individuals is 2.6–3.5 percentage points while the impact on the LFP rate of individuals whose data was reported by another member of their household (a proxy), is lower and statistically insignificant. The relative increase of the LFP rate in the monthly survey is a result of a lower rate of exit from the LF and a somewhat higher rate of entry into the LF relative to these flows in the quarterly survey. These differing flows have a bearing on labor search models as the monthly survey portrays a labor market with less friction and a “steady state” LFP rate that is 5.9 percentage points higher than the quarterly survey. The study also demonstrates that monthly interviews affect a specific group (45–64 year-olds); thus the sign of coefficient of age as an explanatory variable in fixed-effects regressions on LFP is negative in the monthly survey and positive in the quarterly survey.Keywords: measurement error, surveys, search, LFSs
Procedia PDF Downloads 272573 Physics-Informed Machine Learning for Displacement Estimation in Solid Mechanics Problem
Authors: Feng Yang
Abstract:
Machine learning (ML), especially deep learning (DL), has been extensively applied to many applications in recently years and gained great success in solving different problems, including scientific problems. However, conventional ML/DL methodologies are purely data-driven which have the limitations, such as need of ample amount of labelled training data, lack of consistency to physical principles, and lack of generalizability to new problems/domains. Recently, there is a growing consensus that ML models need to further take advantage of prior knowledge to deal with these limitations. Physics-informed machine learning, aiming at integration of physics/domain knowledge into ML, has been recognized as an emerging area of research, especially in the recent 2 to 3 years. In this work, physics-informed ML, specifically physics-informed neural network (NN), is employed and implemented to estimate the displacements at x, y, z directions in a solid mechanics problem that is controlled by equilibrium equations with boundary conditions. By incorporating the physics (i.e. the equilibrium equations) into the learning process of NN, it is showed that the NN can be trained very efficiently with a small set of labelled training data. Experiments with different settings of the NN model and the amount of labelled training data were conducted, and the results show that very high accuracy can be achieved in fulfilling the equilibrium equations as well as in predicting the displacements, e.g. in setting the overall displacement of 0.1, a root mean square error (RMSE) of 2.09 × 10−4 was achieved.Keywords: deep learning, neural network, physics-informed machine learning, solid mechanics
Procedia PDF Downloads 152572 Contactless Attendance System along with Temperature Monitoring
Authors: Nalini C. Iyer, Shraddha H., Anagha B. Varahamurthy, Dikshith C. S., Ishwar G. Kubasad, Vinayak I. Karalatti, Pavan B. Mulimani
Abstract:
The current scenario of the pandemic due to COVID-19 has led to the awareness among the people to avoid unneces-sary contact in public places. There is a need to avoid contact with physical objects to stop the spreading of infection. The contactless feature has to be included in the systems in public places wherever possible. For example, attendance monitoring systems with fingerprint biometric can be replaced with a contactless feature. One more important protocol followed in the current situation is temperature monitoring and screening. The paper describes an attendance system with a contactless feature and temperature screening for the university. The system displays a QR code to scan, which redirects to the student login web page only if the location is valid (the location where the student scans the QR code should be the location of the display of the QR code). Once the student logs in, the temperature of the student is scanned by the contactless temperature sensor (mlx90614) with an error of 0.5°C. If the temperature falls in the range of the desired value (range of normal body temperature), then the attendance of the student is marked as present, stored in the database, and the door opens automatically. The attendance is marked as absent in the other case, alerted with the display of temperature, and the door remains closed. The door is automated with the help of a servomotor. To avoid the proxy, IR sensors are used to count the number of students in the classroom. The hardware system consisting of a contactless temperature sensor and IR sensor is implemented on the microcontroller, NodeMCU.Keywords: NodeMCU, IR sensor, attendance monitoring, contactless, temperature
Procedia PDF Downloads 188571 A Uniformly Convergent Numerical Scheme for a Singularly Perturbed Volterra Integrodifferential Equation
Authors: Nana Adjoah Mbroh, Suares Clovis Oukouomi Noutchie
Abstract:
Singularly perturbed problems are parameter dependent problems, and they play major roles in the modelling of real-life situational problems in applied sciences. Thus, designing efficient numerical schemes to solve these problems is of much interest since the exact solutions of such problems may not even exist. Generally, singularly perturbed problems are identified by a small parameter multiplying at least the highest derivative in the equation. The presence of this parameter causes the solution of these problems to be characterized by rapid oscillations. This unique feature renders classical numerical schemes inefficient since they are unable to capture the behaviour of the exact solution in the part of the domain where the rapid oscillations are present. In this paper, a numerical scheme is proposed to solve a singularly perturbed Volterra Integro-differential equation. The scheme is based on the midpoint rule and employs the non-standard finite difference scheme to solve the differential part whilst the composite trapezoidal rule is used for the integral part. A fully fledged error estimate is performed, and Richardson extrapolation is applied to accelerate the convergence of the scheme. Numerical simulations are conducted to confirm the theoretical findings before and after extrapolation.Keywords: midpoint rule, non-standard finite difference schemes, Richardson extrapolation, singularly perturbed problems, trapezoidal rule, uniform convergence
Procedia PDF Downloads 129570 The Identification of Combined Genomic Expressions as a Diagnostic Factor for Oral Squamous Cell Carcinoma
Authors: Ki-Yeo Kim
Abstract:
Trends in genetics are transforming in order to identify differential coexpressions of correlated gene expression rather than the significant individual gene. Moreover, it is known that a combined biomarker pattern improves the discrimination of a specific cancer. The identification of the combined biomarker is also necessary for the early detection of invasive oral squamous cell carcinoma (OSCC). To identify the combined biomarker that could improve the discrimination of OSCC, we explored an appropriate number of genes in a combined gene set in order to attain the highest level of accuracy. After detecting a significant gene set, including the pre-defined number of genes, a combined expression was identified using the weights of genes in a gene set. We used the Principal Component Analysis (PCA) for the weight calculation. In this process, we used three public microarray datasets. One dataset was used for identifying the combined biomarker, and the other two datasets were used for validation. The discrimination accuracy was measured by the out-of-bag (OOB) error. There was no relation between the significance and the discrimination accuracy in each individual gene. The identified gene set included both significant and insignificant genes. One of the most significant gene sets in the classification of normal and OSCC included MMP1, SOCS3 and ACOX1. Furthermore, in the case of oral dysplasia and OSCC discrimination, two combined biomarkers were identified. The combined genomic expression achieved better performance in the discrimination of different conditions than in a single significant gene. Therefore, it could be expected that accurate diagnosis for cancer could be possible with a combined biomarker.Keywords: oral squamous cell carcinoma, combined biomarker, microarray dataset, correlated genes
Procedia PDF Downloads 426569 Numerical Solution of Space Fractional Order Linear/Nonlinear Reaction-Advection Diffusion Equation Using Jacobi Polynomial
Authors: Shubham Jaiswal
Abstract:
During modelling of many physical problems and engineering processes, fractional calculus plays an important role. Those are greatly described by fractional differential equations (FDEs). So a reliable and efficient technique to solve such types of FDEs is needed. In this article, a numerical solution of a class of fractional differential equations namely space fractional order reaction-advection dispersion equations subject to initial and boundary conditions is derived. In the proposed approach shifted Jacobi polynomials are used to approximate the solutions together with shifted Jacobi operational matrix of fractional order and spectral collocation method. The main advantage of this approach is that it converts such problems in the systems of algebraic equations which are easier to be solved. The proposed approach is effective to solve the linear as well as non-linear FDEs. To show the reliability, validity and high accuracy of proposed approach, the numerical results of some illustrative examples are reported, which are compared with the existing analytical results already reported in the literature. The error analysis for each case exhibited through graphs and tables confirms the exponential convergence rate of the proposed method.Keywords: space fractional order linear/nonlinear reaction-advection diffusion equation, shifted Jacobi polynomials, operational matrix, collocation method, Caputo derivative
Procedia PDF Downloads 449568 Coding and Decoding versus Space Diversity for Rayleigh Fading Radio Frequency Channels
Authors: Ahmed Mahmoud Ahmed Abouelmagd
Abstract:
The diversity is the usual remedy of the transmitted signal level variations (Fading phenomena) in radio frequency channels. Diversity techniques utilize two or more copies of a signal and combine those signals to combat fading. The basic concept of diversity is to transmit the signal via several independent diversity branches to get independent signal replicas via time – frequency - space - and polarization diversity domains. Coding and decoding processes can be an alternative remedy for fading phenomena, it cannot increase the channel capacity, but it can improve the error performance. In this paper we propose the use of replication decoding with BCH code class, and Viterbi decoding algorithm with convolution coding; as examples of coding and decoding processes. The results are compared to those obtained from two optimized selection space diversity techniques. The performance of Rayleigh fading channel, as the model considered for radio frequency channels, is evaluated for each case. The evaluation results show that the coding and decoding approaches, especially the BCH coding approach with replication decoding scheme, give better performance compared to that of selection space diversity optimization approaches. Also, an approach for combining the coding and decoding diversity as well as the space diversity is considered, the main disadvantage of this approach is its complexity but it yields good performance results.Keywords: Rayleigh fading, diversity, BCH codes, Replication decoding, convolution coding, viterbi decoding, space diversity
Procedia PDF Downloads 445567 Allometric Models for Biomass Estimation in Savanna Woodland Area, Niger State, Nigeria
Authors: Abdullahi Jibrin, Aishetu Abdulkadir
Abstract:
The development of allometric models is crucial to accurate forest biomass/carbon stock assessment. The aim of this study was to develop a set of biomass prediction models that will enable the determination of total tree aboveground biomass for savannah woodland area in Niger State, Nigeria. Based on the data collected through biometric measurements of 1816 trees and destructive sampling of 36 trees, five species specific and one site specific models were developed. The sample size was distributed equally between the five most dominant species in the study site (Vitellaria paradoxa, Irvingia gabonensis, Parkia biglobosa, Anogeissus leiocarpus, Pterocarpus erinaceous). Firstly, the equations were developed for five individual species. Secondly these five species were mixed and were used to develop an allometric equation of mixed species. Overall, there was a strong positive relationship between total tree biomass and the stem diameter. The coefficient of determination (R2 values) ranging from 0.93 to 0.99 P < 0.001 were realised for the models; with considerable low standard error of the estimates (SEE) which confirms that the total tree above ground biomass has a significant relationship with the dbh. The F-test value for the biomass prediction models were also significant at p < 0.001 which indicates that the biomass prediction models are valid. This study recommends that for improved biomass estimates in the study site, the site specific biomass models should preferably be used instead of using generic models.Keywords: allometriy, biomass, carbon stock , model, regression equation, woodland, inventory
Procedia PDF Downloads 450566 An Efficient Backward Semi-Lagrangian Scheme for Nonlinear Advection-Diffusion Equation
Authors: Soyoon Bak, Sunyoung Bu, Philsu Kim
Abstract:
In this paper, a backward semi-Lagrangian scheme combined with the second-order backward difference formula is designed to calculate the numerical solutions of nonlinear advection-diffusion equations. The primary aims of this paper are to remove any iteration process and to get an efficient algorithm with the convergence order of accuracy 2 in time. In order to achieve these objects, we use the second-order central finite difference and the B-spline approximations of degree 2 and 3 in order to approximate the diffusion term and the spatial discretization, respectively. For the temporal discretization, the second order backward difference formula is applied. To calculate the numerical solution of the starting point of the characteristic curves, we use the error correction methodology developed by the authors recently. The proposed algorithm turns out to be completely iteration-free, which resolves the main weakness of the conventional backward semi-Lagrangian method. Also, the adaptability of the proposed method is indicated by numerical simulations for Burgers’ equations. Throughout these numerical simulations, it is shown that the numerical results are in good agreement with the analytic solution and the present scheme offer better accuracy in comparison with other existing numerical schemes. Semi-Lagrangian method, iteration-free method, nonlinear advection-diffusion equation, second-order backward difference formulaKeywords: Semi-Lagrangian method, iteration free method, nonlinear advection-diffusion equation, second-order backward difference formula
Procedia PDF Downloads 322565 Trusting the Eyes: The Changing Landscape of Eyewitness Testimony
Authors: Manveen Singh
Abstract:
Since the very advent of law enforcement, eyewitness testimony has played a pivotal role in identifying, arresting and convicting suspects. Reliant heavily on the accuracy of human memory, nothing seems to carry more weight with the judiciary than the testimony of an actual witness. The acceptance of eyewitness testimony as a substantive piece of evidence lies embedded in the assumption that the human mind is adept at recording and storing events. Research though, has proven otherwise. Having carried out extensive study in the field of eyewitness testimony for the past 40 years, psychologists have concluded that human memory is fragile and needs to be treated carefully. The question that arises then, is how reliable is eyewitness testimony? The credibility of eyewitness testimony, simply put, depends on several factors leaving it reliable at times while not so much at others. This is further substantiated by the fact that as per scientific research, over 75 percent of all eyewitness testimonies may stand in error; quite a few of these cases resulting in life sentences. Although the advancement of scientific techniques, especially DNA testing, helped overturn many of these eyewitness testimony-based convictions, yet eyewitness identifications continue to form the backbone of most police investigations and courtroom decisions till date. What then is the solution to this long standing concern regarding the accuracy of eyewitness accounts? The present paper shall analyze the linkage between human memory and eyewitness identification as well as look at the various factors governing the credibility of eyewitness testimonies. Furthermore, it shall elaborate upon some best practices developed over the years to help reduce mistaken identifications. Thus, in the process, trace out the changing landscape of eyewitness testimony amidst the evolution of DNA and trace evidence.Keywords: DNA, eyewitness, identification, testimony, evidence
Procedia PDF Downloads 330564 Numerical Study of Jet Impingement Heat Transfer
Authors: A. M. Tiara, Sudipto Chakraborty, S. K. Pal
Abstract:
Impinging jets and their different configurations are important from the viewpoint of the fluid flow characteristics and their influence on heat transfer from metal surfaces due to their complex flow characteristics. Such flow characteristics results in highly variable heat transfer from the surface, resulting in varying cooling rates which affects the mechanical properties including hardness and strength. The overall objective of the current research is to conduct a fundamental investigation of the heat transfer mechanisms for an impinging coolant jet. Numerical simulation of the cooling process gives a detailed analysis of the different parameters involved even though employing Computational Fluid Dynamics (CFD) to simulate the real time process, being a relatively new research area, poses many challenges. The heat transfer mechanism in the current research is actuated by jet cooling. The computational tool used in the ongoing research for simulation of the cooling process is ANSYS Workbench software. The temperature and heat flux distribution along the steel strip with the effect of various flow parameters on the heat transfer rate can be observed in addition to determination of the jet impingement patterns, which is the major aim of the present analysis. Modelling both jet and air atomized cooling techniques using CFD methodology and validating with those obtained experimentally- including trial and error with different models and comparison of cooling rates from both the techniques have been included in this work. Finally some concluding remarks are made that identify some gaps in the available literature that have influenced the path of the current investigation.Keywords: CFD, heat transfer, impinging jets, numerical simulation
Procedia PDF Downloads 238563 Commuters Trip Purpose Decision Tree Based Model of Makurdi Metropolis, Nigeria and Strategic Digital City Project
Authors: Emmanuel Okechukwu Nwafor, Folake Olubunmi Akintayo, Denis Alcides Rezende
Abstract:
Decision tree models are versatile and interpretable machine learning algorithms widely used for both classification and regression tasks, which can be related to cities, whether physical or digital. The aim of this research is to assess how well decision tree algorithms can predict trip purposes in Makurdi, Nigeria, while also exploring their connection to the strategic digital city initiative. The research methodology involves formalizing household demographic and trips information datasets obtained from extensive survey process. Modelling and Prediction were achieved using Python Programming Language and the evaluation metrics like R-squared and mean absolute error were used to assess the decision tree algorithm's performance. The results indicate that the model performed well, with accuracies of 84% and 68%, and low MAE values of 0.188 and 0.314, on training and validation data, respectively. This suggests the model can be relied upon for future prediction. The conclusion reiterates that This model will assist decision-makers, including urban planners, transportation engineers, government officials, and commuters, in making informed decisions on transportation planning and management within the framework of a strategic digital city. Its application will enhance the efficiency, sustainability, and overall quality of transportation services in Makurdi, Nigeria.Keywords: decision tree algorithm, trip purpose, intelligent transport, strategic digital city, travel pattern, sustainable transport
Procedia PDF Downloads 25562 Characterization of Bio-Inspired Thermoelastoplastic Composites Filled with Modified Cellulose Fibers
Authors: S. Cichosz, A. Masek
Abstract:
A new cellulose hybrid modification approach, which is undoubtedly a scientific novelty, is introduced. The study reports the properties of cellulose (Arbocel UFC100 – Ultra Fine Cellulose) and characterizes cellulose filled polymer composites based on an ethylene-norbornene copolymer (TOPAS Elastomer E-140). Moreover, the approach of physicochemical two-stage cellulose treatment is introduced: solvent exchange (to ethanol or hexane) and further chemical modification with maleic anhydride (MA). Furthermore, the impact of the drying process on cellulose properties was investigated. Suitable measurements were carried out to characterize cellulose fibers: spectroscopic investigation (Fourier Transform Infrared Spektrofotometer-FTIR, Near InfraRed spectroscopy-NIR), thermal analysis (Differential scanning calorimetry, Thermal gravimetric analysis ) and Karl Fischer titration. It should be emphasized that for all UFC100 treatments carried out, a decrease in moisture content was evidenced. FT-IR reveals a drop in absorption band intensity at 3334 cm-1, the peak is associated with both –OH moieties and water. Similar results were obtained with Karl Fischer titration. Based on the results obtained, it may be claimed that the employment of ethanol contributes greatly to the lowering of cellulose water absorption ability (decrease of moisture content to approximately 1.65%). Additionally, regarding polymer composite properties, crucial data has been obtained from the mechanical and thermal analysis. The highest material performance was noted in the case of the composite sample that contained cellulose modified with MA after a solvent exchange with ethanol. This specimen exhibited sufficient tensile strength, which is almost the same as that of the neat polymer matrix – in the region of 40 MPa. Moreover, both the Payne effect and filler efficiency factor, calculated based on dynamic mechanical analysis (DMA), reveal the possibility of the filler having a reinforcing nature. What is also interesting is that, according to the Payne effect results, fibers dried before the further chemical modification are assumed to allow more regular filler structure development in the polymer matrix (Payne effect maximum at 1.60 MPa), compared with those not dried (Payne effect in the range 0.84-1.26 MPa). Furthermore, taking into consideration the data gathered from DSC and TGA, higher thermal stability is obtained in case of the materials filled with fibers that were dried before the carried out treatments (degradation activation energy in the region of 195 kJ/mol) in comparison with the polymer composite samples filled with unmodified cellulose (degradation activation energy of approximately 180 kJ/mol). To author’s best knowledge this work results in the introduction of a novel, new filler hybrid treatment approach. Moreover, valuable data regarding the properties of composites filled with cellulose fibers of various moisture contents have been provided. It should be emphasized that plant fiber-based polymer bio-materials described in this research might contribute significantly to polymer waste minimization because they are more readily degraded.Keywords: cellulose fibers, solvent exchange, moisture content, ethylene-norbornene copolymer
Procedia PDF Downloads 118561 Mathematical Modelling of Ultrasound Pre-Treatment in Microwave Dried Strawberry (Fragaria L.) Slices
Authors: Hilal Uslu, Salih Eroglu, Betul Ozkan, Ozcan Bulantekin, Alper Kuscu
Abstract:
In this study, the strawberry (Fragaria L.) fruits, which were pretreated with ultrasound (US), were worked on in the microwave by using 90W power. Then mathematical modelling was applied to dried fruits by using different experimental thin layer models. The sliced fruits were subjected to ultrasound treatment at a frequency of 40 kHz for 10, 20, and 30 minutes, in an ultrasonic water bath, with a ratio of 1:4 to fruit/water. They are then dried in the microwave (90W). The drying process continued until the product moisture was below 10%. By analyzing the moisture change of the products at a certain time, eight different thin-layer drying models, (Newton, page, modified page, Midilli, Henderson and Pabis, logarithmic, two-term, Wang and Singh) were tested for verification of experimental data. MATLAB R2015a statistical program was used for the modelling, and the best suitable model was determined with R²adj (coefficient of determination of compatibility), and root mean square error (RMSE) values. According to analysis, the drying model that best describes the drying behavior for both drying conditions was determined as the Midilli model by high R²adj and low RMSE values. Control, 10, 20, and 30 min US for groups R²adj and RMSE values was established as respectively; 0,9997- 0,005298; 0,9998- 0,004735; 0,9995- 0,007031; 0,9917-0,02773. In addition, effective diffusion coefficients were calculated for each group and were determined as 3,80x 10⁻⁸, 3,71 x 10⁻⁸, 3,26 x10⁻⁸ ve 3,5 x 10⁻⁸ m/s, respectively.Keywords: mathematical modelling, microwave drying, strawberry, ultrasound
Procedia PDF Downloads 154560 Use of Statistical Correlations for the Estimation of Shear Wave Velocity from Standard Penetration Test-N-Values: Case Study of Algiers Area
Authors: Soumia Merat, Lynda Djerbal, Ramdane Bahar, Mohammed Amin Benbouras
Abstract:
Along with shear wave, many soil parameters are associated with the standard penetration test (SPT) as a dynamic in situ experiment. Both SPT-N data and geophysical data do not often exist in the same area. Statistical analysis of correlation between these parameters is an alternate method to estimate Vₛ conveniently and without additional investigations or data acquisition. Shear wave velocity is a basic engineering tool required to define dynamic properties of soils. In many instances, engineers opt for empirical correlations between shear wave velocity (Vₛ) and reliable static field test data like standard penetration test (SPT) N value, CPT (Cone Penetration Test) values, etc., to estimate shear wave velocity or dynamic soil parameters. The relation between Vs and SPT- N values of Algiers area is predicted using the collected data, and it is also compared with the previously suggested formulas of Vₛ determination by measuring Root Mean Square Error (RMSE) of each model. Algiers area is situated in high seismic zone (Zone III [RPA 2003: réglement parasismique algerien]), therefore the study is important for this region. The principal aim of this paper is to compare the field measurements of Down-hole test and the empirical models to show which one of these proposed formulas are applicable to predict and deduce shear wave velocity values.Keywords: empirical models, RMSE, shear wave velocity, standard penetration test
Procedia PDF Downloads 340559 Computational Fluid Dynamic Modeling of Mixing Enhancement by Stimulation of Ferrofluid under Magnetic Field
Authors: Neda Azimi, Masoud Rahimi, Faezeh Mohammadi
Abstract:
Computational fluid dynamics (CFD) simulation was performed to investigate the effect of ferrofluid stimulation on hydrodynamic and mass transfer characteristics of two immiscible liquid phases in a Y-micromixer. The main purpose of this work was to develop a numerical model that is able to simulate hydrodynamic of the ferrofluid flow under magnetic field and determine its effect on mass transfer characteristics. A uniform external magnetic field was applied perpendicular to the flow direction. The volume of fluid (VOF) approach was used for simulating the multiphase flow of ferrofluid and two-immiscible liquid flows. The geometric reconstruction scheme (Geo-Reconstruct) based on piecewise linear interpolation (PLIC) was used for reconstruction of the interface in the VOF approach. The mass transfer rate was defined via an equation as a function of mass concentration gradient of the transported species and added into the phase interaction panel using the user-defined function (UDF). The magnetic field was solved numerically by Fluent MHD module based on solving the magnetic induction equation method. CFD results were validated by experimental data and good agreements have been achieved, which maximum relative error for extraction efficiency was about 7.52 %. It was showed that ferrofluid actuation by a magnetic field can be considered as an efficient mixing agent for liquid-liquid two-phase mass transfer in microdevices.Keywords: CFD modeling, hydrodynamic, micromixer, ferrofluid, mixing
Procedia PDF Downloads 198558 A Hybrid Genetic Algorithm and Neural Network for Wind Profile Estimation
Authors: M. Saiful Islam, M. Mohandes, S. Rehman, S. Badran
Abstract:
Increasing necessity of wind power is directing us to have precise knowledge on wind resources. Methodical investigation of potential locations is required for wind power deployment. High penetration of wind energy to the grid is leading multi megawatt installations with huge investment cost. This fact appeals to determine appropriate places for wind farm operation. For accurate assessment, detailed examination of wind speed profile, relative humidity, temperature and other geological or atmospheric parameters are required. Among all of these uncertainty factors influencing wind power estimation, vertical extrapolation of wind speed is perhaps the most difficult and critical one. Different approaches have been used for the extrapolation of wind speed to hub height which are mainly based on Log law, Power law and various modifications of the two. This paper proposes a Artificial Neural Network (ANN) and Genetic Algorithm (GA) based hybrid model, namely GA-NN for vertical extrapolation of wind speed. This model is very simple in a sense that it does not require any parametric estimations like wind shear coefficient, roughness length or atmospheric stability and also reliable compared to other methods. This model uses available measured wind speeds at 10m, 20m and 30m heights to estimate wind speeds up to 100m. A good comparison is found between measured and estimated wind speeds at 30m and 40m with approximately 3% mean absolute percentage error. Comparisons with ANN and power law, further prove the feasibility of the proposed method.Keywords: wind profile, vertical extrapolation of wind, genetic algorithm, artificial neural network, hybrid machine learning
Procedia PDF Downloads 492557 Economic Valuation of Forest Landscape Function Using a Conditional Logit Model
Authors: A. J. Julius, E. Imoagene, O. A. Ganiyu
Abstract:
The purpose of this study is to estimate the economic value of the services and functions rendered by the forest landscape using a conditional logit model. For this study, attributes and levels of forest landscape were chosen; specifically, attributes include topographical forest type, forest type, forest density, recreational factor (side trip, accessibility of valley), and willingness to participate (WTP). Based on these factors, 48 choices sets with balanced and orthogonal form using statistical analysis system (SAS) 9.1 was adopted. The efficiency of the questionnaire was 6.02 (D-Error. 0.1), and choice set and socio-economic variables were analyzed. To reduce the cognitive load of respondents, the 48 choice sets were divided into 4 types in the questionnaire, so that respondents could respond to 12 choice sets, respectively. The study populations were citizens from seven metropolitan cities including Ibadan, Ilorin, Osogbo, etc. and annual WTP per household was asked by using the interview questionnaire, a total of 267 copies were recovered. As a result, Oshogbo had 0.45, and the statistical similarities could not be found except for urban forests, forest density, recreational factor, and level of WTP. Average annual WTP per household for forest landscape was 104,758 Naira (Nigerian currency) based on the outcome from this model, total economic value of the services and functions enjoyed from Nigerian forest landscape has reached approximately 1.6 trillion Naira.Keywords: economic valuation, urban cities, services, forest landscape, logit model, nigeria
Procedia PDF Downloads 135556 Modelling Biological Treatment of Dye Wastewater in SBR Systems Inoculated with Bacteria by Artificial Neural Network
Authors: Yasaman Sanayei, Alireza Bahiraie
Abstract:
This paper presents a systematic methodology based on the application of artificial neural networks for sequencing batch reactor (SBR). The SBR is a fill-and-draw biological wastewater technology, which is specially suited for nutrient removal. Employing reactive dye by Sphingomonas paucimobilis bacteria at sequence batch reactor is a novel approach of dye removal. The influent COD, MLVSS, and reaction time were selected as the process inputs and the effluent COD and BOD as the process outputs. The best possible result for the discrete pole parameter was a= 0.44. In orderto adjust the parameters of ANN, the Levenberg-Marquardt (LM) algorithm was employed. The results predicted by the model were compared to the experimental data and showed a high correlation with R2> 0.99 and a low mean absolute error (MAE). The results from this study reveal that the developed model is accurate and efficacious in predicting COD and BOD parameters of the dye-containing wastewater treated by SBR. The proposed modeling approach can be applied to other industrial wastewater treatment systems to predict effluent characteristics. Note that SBR are normally operated with constant predefined duration of the stages, thus, resulting in low efficient operation. Data obtained from the on-line electronic sensors installed in the SBR and from the control quality laboratory analysis have been used to develop the optimal architecture of two different ANN. The results have shown that the developed models can be used as efficient and cost-effective predictive tools for the system analysed.Keywords: artificial neural network, COD removal, SBR, Sphingomonas paucimobilis
Procedia PDF Downloads 415555 Anthropomorphism in the Primate Mind-Reading Debate: A Critique of Sober's Justification Argument
Authors: Boyun Lee
Abstract:
This study aims to discuss whether anthropomorphism some scientists tend to use in cross-species comparison can be justified epistemologically, especially in the primate mind-reading debate. Concretely, this study critically analyzes Elliott Sober’s argument about mind-reading hypothesis (MRH), an anthropomorphic hypothesis which states that nonhuman primates (e.g., chimpanzee) are mind-readers like humans. Although many scientists consider anthropomorphism as an error and choosing anthropomorphic hypothesis like MRH without any definite evidence invalid, Sober advocates that anthropomorphism is supported by cladistic parsimony that suggests choosing the simplest hypothesis postulating the minimum number of evolutionary changes, which can be justified epistemologically in the mind-reading debate. However, his argument has several problems. First, Reichenbach’s theorem which Sober uses in process of showing that MRH has the higher likelihood than its competing hypothesis, behavior-reading hypothesis (BRH), does not fit in the context of inferring the evolutionary relationship. Second, the phylogenetic tree Sober supports is one of the possible scenarios of MRH, and even without this problem, it is difficult to prove that the possibility nonhuman primate species and human share mind-reading ability is higher than the possibility of the other case, considering how evolution occurs. Consequently, it seems hard to justify anthropomorphism of MRH under Sober’s argument. Some scientists and philosophers say that anthropomorphism sometimes helps observe interesting phenomena or make hypotheses in comparative biology. Nonetheless, we cannot determine that it provides answers about why and how the interesting phenomena appear or which of the hypotheses is better, at least the mind-reading debate, under the current state.Keywords: anthropomorphism, cladistic parsimony, comparative biology, mind-reading debate
Procedia PDF Downloads 173554 Optimizing and Evaluating Performance Quality Control of the Production Process of Disposable Essentials Using Approach Vague Goal Programming
Authors: Hadi Gholizadeh, Ali Tajdin
Abstract:
To have effective production planning, it is necessary to control the quality of processes. This paper aims at improving the performance of the disposable essentials process using statistical quality control and goal programming in a vague environment. That is expressed uncertainty because there is always a measurement error in the real world. Therefore, in this study, the conditions are examined in a vague environment that is a distance-based environment. The disposable essentials process in Kach Company was studied. Statistical control tools were used to characterize the existing process for four factor responses including the average of disposable glasses’ weights, heights, crater diameters, and volumes. Goal programming was then utilized to find the combination of optimal factors setting in a vague environment which is measured to apply uncertainty of the initial information when some of the parameters of the models are vague; also, the fuzzy regression model is used to predict the responses of the four described factors. Optimization results show that the process capability index values for disposable glasses’ average of weights, heights, crater diameters and volumes were improved. Such increasing the quality of the products and reducing the waste, which will reduce the cost of the finished product, and ultimately will bring customer satisfaction, and this satisfaction, will mean increased sales.Keywords: goal programming, quality control, vague environment, disposable glasses’ optimization, fuzzy regression
Procedia PDF Downloads 227553 A Comparative Soft Computing Approach to Supplier Performance Prediction Using GEP and ANN Models: An Automotive Case Study
Authors: Seyed Esmail Seyedi Bariran, Khairul Salleh Mohamed Sahari
Abstract:
In multi-echelon supply chain networks, optimal supplier selection significantly depends on the accuracy of suppliers’ performance prediction. Different methods of multi criteria decision making such as ANN, GA, Fuzzy, AHP, etc have been previously used to predict the supplier performance but the “black-box” characteristic of these methods is yet a major concern to be resolved. Therefore, the primary objective in this paper is to implement an artificial intelligence-based gene expression programming (GEP) model to compare the prediction accuracy with that of ANN. A full factorial design with %95 confidence interval is initially applied to determine the appropriate set of criteria for supplier performance evaluation. A test-train approach is then utilized for the ANN and GEP exclusively. The training results are used to find the optimal network architecture and the testing data will determine the prediction accuracy of each method based on measures of root mean square error (RMSE) and correlation coefficient (R2). The results of a case study conducted in Supplying Automotive Parts Co. (SAPCO) with more than 100 local and foreign supply chain members revealed that, in comparison with ANN, gene expression programming has a significant preference in predicting supplier performance by referring to the respective RMSE and R-squared values. Moreover, using GEP, a mathematical function was also derived to solve the issue of ANN black-box structure in modeling the performance prediction.Keywords: Supplier Performance Prediction, ANN, GEP, Automotive, SAPCO
Procedia PDF Downloads 422