Search results for: improvement of model accuracy and reliability
23027 Reliability Qualification Test Plan Derivation Method for Weibull Distributed Products
Authors: Ping Jiang, Yunyan Xing, Dian Zhang, Bo Guo
Abstract:
The reliability qualification test (RQT) is widely used in product development to qualify whether the product meets predetermined reliability requirements, which are mainly described in terms of reliability indices, for example, MTBF (Mean Time Between Failures). It is widely exercised in product development. In engineering practices, RQT plans are mandatorily referred to standards, such as MIL-STD-781 or GJB899A-2009. But these conventional RQT plans in standards are not preferred, as the test plans often require long test times or have high risks for both producer and consumer due to the fact that the methods in the standards only use the test data of the product itself. And the standards usually assume that the product is exponentially distributed, which is not suitable for a complex product other than electronics. So it is desirable to develop an RQT plan derivation method that safely shortens test time while keeping the two risks under control. To meet this end, for the product whose lifetime follows Weibull distribution, an RQT plan derivation method is developed. The merit of the method is that expert judgment is taken into account. This is implemented by applying the Bayesian method, which translates the expert judgment into prior information on product reliability. Then producer’s risk and the consumer’s risk are calculated accordingly. The procedures to derive RQT plans are also proposed in this paper. As extra information and expert judgment are added to the derivation, the derived test plans have the potential to shorten the required test time and have satisfactory low risks for both producer and consumer, compared with conventional test plans. A case study is provided to prove that when using expert judgment in deriving product test plans, the proposed method is capable of finding ideal test plans that not only reduce the two risks but also shorten the required test time as well.Keywords: expert judgment, reliability qualification test, test plan derivation, producer’s risk, consumer’s risk
Procedia PDF Downloads 13723026 Dynamic Fault Tree Analysis of Dynamic Positioning System through Monte Carlo Approach
Authors: A. S. Cheliyan, S. K. Bhattacharyya
Abstract:
Dynamic Positioning System (DPS) is employed in marine vessels of the offshore oil and gas industry. It is a computer controlled system to automatically maintain a ship’s position and heading by using its own thrusters. Reliability assessment of the same can be analyzed through conventional fault tree. However, the complex behaviour like sequence failure, redundancy management and priority of failing of events cannot be analyzed by the conventional fault trees. The Dynamic Fault Tree (DFT) addresses these shortcomings of conventional Fault Tree by defining additional gates called dynamic gates. Monte Carlo based simulation approach has been adopted for the dynamic gates. This method of realistic modeling of DPS gives meaningful insight into the system reliability and the ability to improve the same.Keywords: dynamic positioning system, dynamic fault tree, Monte Carlo simulation, reliability assessment
Procedia PDF Downloads 77323025 Flood Predicting in Karkheh River Basin Using Stochastic ARIMA Model
Authors: Karim Hamidi Machekposhti, Hossein Sedghi, Abdolrasoul Telvari, Hossein Babazadeh
Abstract:
Floods have huge environmental and economic impact. Therefore, flood prediction is given a lot of attention due to its importance. This study analysed the annual maximum streamflow (discharge) (AMS or AMD) of Karkheh River in Karkheh River Basin for flood predicting using ARIMA model. For this purpose, we use the Box-Jenkins approach, which contains four-stage method model identification, parameter estimation, diagnostic checking and forecasting (predicting). The main tool used in ARIMA modelling was the SAS and SPSS software. Model identification was done by visual inspection on the ACF and PACF. SAS software computed the model parameters using the ML, CLS and ULS methods. The diagnostic checking tests, AIC criterion, RACF graph and RPACF graphs, were used for selected model verification. In this study, the best ARIMA models for Annual Maximum Discharge (AMD) time series was (4,1,1) with their AIC value of 88.87. The RACF and RPACF showed residuals’ independence. To forecast AMD for 10 future years, this model showed the ability of the model to predict floods of the river under study in the Karkheh River Basin. Model accuracy was checked by comparing the predicted and observation series by using coefficient of determination (R2).Keywords: time series modelling, stochastic processes, ARIMA model, Karkheh river
Procedia PDF Downloads 28723024 A Type-2 Fuzzy Model for Link Prediction in Social Network
Authors: Mansoureh Naderipour, Susan Bastani, Mohammad Fazel Zarandi
Abstract:
Predicting links that may occur in the future and missing links in social networks is an attractive problem in social network analysis. Granular computing can help us to model the relationships between human-based system and social sciences in this field. In this paper, we present a model based on granular computing approach and Type-2 fuzzy logic to predict links regarding nodes’ activity and the relationship between two nodes. Our model is tested on collaboration networks. It is found that the accuracy of prediction is significantly higher than the Type-1 fuzzy and crisp approach.Keywords: social network, link prediction, granular computing, type-2 fuzzy sets
Procedia PDF Downloads 32523023 Nonlinear Modeling of the PEMFC Based on NNARX Approach
Authors: Shan-Jen Cheng, Te-Jen Chang, Kuang-Hsiung Tan, Shou-Ling Kuo
Abstract:
Polymer Electrolyte Membrane Fuel Cell (PEMFC) is such a time-vary nonlinear dynamic system. The traditional linear modeling approach is hard to estimate structure correctly of PEMFC system. From this reason, this paper presents a nonlinear modeling of the PEMFC using Neural Network Auto-regressive model with eXogenous inputs (NNARX) approach. The multilayer perception (MLP) network is applied to evaluate the structure of the NNARX model of PEMFC. The validity and accuracy of NNARX model are tested by one step ahead relating output voltage to input current from measured experimental of PEMFC. The results show that the obtained nonlinear NNARX model can efficiently approximate the dynamic mode of the PEMFC and model output and system measured output consistently.Keywords: PEMFC, neural network, nonlinear modeling, NNARX
Procedia PDF Downloads 38123022 Port Governance Model by International Freight Forwarders’ Point of View: A Study at Port of Santos - Brazil
Authors: Guilherme B. B. Vieira, Rafael M. da Silva, Eliana T. P. Senna, Luiz A. S. Senna, Francisco J. Kliemann Neto
Abstract:
Due to the importance of ports to trade and economic development of the regions in which they are inserted, in recent decades the number of studies devoted to this subject has increased. Part of these studies consider the ports as business agglomerations and focuses on port governance. This is an important approach since the port performance is the result of activities performed by actors belonging to the port-logistics chain, which need to be properly coordinated. This coordination takes place through a port governance model. Given this context, this study aims to analyze the governance model of the port of Santos from the perspective of port customers. To do this, a closed-ended questionnaire based on a conceptual model that considers the key dimensions associated with port governance was applied to the international freight forwarders that operate in the port. The results show the applicability of the considered model and highlight improvement opportunities to be implemented at the port of Santos.Keywords: port governance, model, Port of Santos, customers’ perception
Procedia PDF Downloads 45223021 The Direct Deconvolutional Model in the Large-Eddy Simulation of Turbulence
Authors: Ning Chang, Zelong Yuan, Yunpeng Wang, Jianchun Wang
Abstract:
The utilization of Large Eddy Simulation (LES) has been extensive in turbulence research. LES concentrates on resolving the significant grid-scale motions while representing smaller scales through subfilter-scale (SFS) models. The deconvolution model, among the available SFS models, has proven successful in LES of engineering and geophysical flows. Nevertheless, the thorough investigation of how sub-filter scale dynamics and filter anisotropy affect SFS modeling accuracy remains lacking. The outcomes of LES are significantly influenced by filter selection and grid anisotropy, factors that have not been adequately addressed in earlier studies. This study examines two crucial aspects of LES: Firstly, the accuracy of direct deconvolution models (DDM) is evaluated concerning sub-filter scale (SFS) dynamics across varying filter-to-grid ratios (FGR) in isotropic turbulence. Various invertible filters are employed, including Gaussian, Helmholtz I and II, Butterworth, Chebyshev I and II, Cauchy, Pao, and rapidly decaying filters. The importance of FGR becomes evident as it plays a critical role in controlling errors for precise SFS stress prediction. When FGR is set to 1, the DDM models struggle to faithfully reconstruct SFS stress due to inadequate resolution of SFS dynamics. Notably, prediction accuracy improves when FGR is set to 2, leading to accurate reconstruction of SFS stress, except for cases involving Helmholtz I and II filters. Remarkably high precision, nearly 100%, is achieved at an FGR of 4 for all DDM models. Furthermore, the study extends to filter anisotropy and its impact on SFS dynamics and LES accuracy. By utilizing the dynamic Smagorinsky model (DSM), dynamic mixed model (DMM), and direct deconvolution model (DDM) with anisotropic filters, aspect ratios (AR) ranging from 1 to 16 are examined in LES filters. The results emphasize the DDM’s proficiency in accurately predicting SFS stresses under highly anisotropic filtering conditions. Notably high correlation coefficients exceeding 90% are observed in the a priori study for the DDM’s reconstructed SFS stresses, surpassing those of the DSM and DMM models. However, these correlations tend to decrease as filter anisotropy increases. In the a posteriori analysis, the DDM model consistently outperforms the DSM and DMM models across various turbulence statistics, including velocity spectra, probability density functions related to vorticity, SFS energy flux, velocity increments, strainrate tensors, and SFS stress. It is evident that as filter anisotropy intensifies, the results of DSM and DMM deteriorate, while the DDM consistently delivers satisfactory outcomes across all filter-anisotropy scenarios. These findings underscore the potential of the DDM framework as a valuable tool for advancing the development of sophisticated SFS models for LES in turbulence research.Keywords: deconvolution model, large eddy simulation, subfilter scale modeling, turbulence
Procedia PDF Downloads 7523020 Predicting Success and Failure in Drug Development Using Text Analysis
Authors: Zhi Hao Chow, Cian Mulligan, Jack Walsh, Antonio Garzon Vico, Dimitar Krastev
Abstract:
Drug development is resource-intensive, time-consuming, and increasingly expensive with each developmental stage. The success rates of drug development are also relatively low, and the resources committed are wasted with each failed candidate. As such, a reliable method of predicting the success of drug development is in demand. The hypothesis was that some examples of failed drug candidates are pushed through developmental pipelines based on false confidence and may possess common linguistic features identifiable through sentiment analysis. Here, the concept of using text analysis to discover such features in research publications and investor reports as predictors of success was explored. R studios were used to perform text mining and lexicon-based sentiment analysis to identify affective phrases and determine their frequency in each document, then using SPSS to determine the relationship between our defined variables and the accuracy of predicting outcomes. A total of 161 publications were collected and categorised into 4 groups: (i) Cancer treatment, (ii) Neurodegenerative disease treatment, (iii) Vaccines, and (iv) Others (containing all other drugs that do not fit into the 3 categories). Text analysis was then performed on each document using 2 separate datasets (BING and AFINN) in R within the category of drugs to determine the frequency of positive or negative phrases in each document. A relative positivity and negativity value were then calculated by dividing the frequency of phrases with the word count of each document. Regression analysis was then performed with SPSS statistical software on each dataset (values from using BING or AFINN dataset during text analysis) using a random selection of 61 documents to construct a model. The remaining documents were then used to determine the predictive power of the models. Model constructed from BING predicts the outcome of drug performance in clinical trials with an overall percentage of 65.3%. AFINN model had a lower accuracy at predicting outcomes compared to the BING model at 62.5% but was not effective at predicting the failure of drugs in clinical trials. Overall, the study did not show significant efficacy of the model at predicting outcomes of drugs in development. Many improvements may need to be made to later iterations of the model to sufficiently increase the accuracy.Keywords: data analysis, drug development, sentiment analysis, text-mining
Procedia PDF Downloads 15723019 The Impact of Major Accounting Events on Managerial Ability and the Accuracy of Environmental Capital Expenditure Projections of the Environmentally Sensitive Industries
Authors: Jason Chen, Jennifer Chen, Shiyu Li
Abstract:
We examine whether managerial ability (MA), the passing of Sarbanes-Oxley in 2002 (SOX), and corporate operational complexity affect the accuracy of environmental capital expenditure projections of the environmentally sensitive industries (ESI). Prior studies found that firms in the ESI manipulated their projected environmental capital expenditures as a tool to achieve corporate legitimation and suggested that human factors must be examined to determine whether they are part of the determinants. We use MA to proxy for the latent human factors to examine whether MA affects the accuracy of financial disclosures in the ESI. To expand Chen and Chen (2020), we further investigate whether (1) SOX and (2) firms with complex operations and financial reporting in conjunction with MA affect firms’ projection accuracy. We find, overall, that MA is positively correlated with firm’s projection accuracy in the annual 10-Ks. Furthermore, results suggest that SOX has a positive, yet temporary, effect on MA, and that leads to better accuracy. Finally, MA matters for firms with more complex operations and financial reporting to make less projection errors than their less-complex counterparts. These results suggest that MA is a determinant that affects the accuracy of environmental capital expenditure projections for the firms in the ESI.Keywords: managerial ability, environmentally sensitive industries, sox, corporate operational complexity
Procedia PDF Downloads 14523018 Forecasting Unemployment Rate in Selected European Countries Using Smoothing Methods
Authors: Ksenija Dumičić, Anita Čeh Časni, Berislav Žmuk
Abstract:
The aim of this paper is to select the most accurate forecasting method for predicting the future values of the unemployment rate in selected European countries. In order to do so, several forecasting techniques adequate for forecasting time series with trend component, were selected, namely: double exponential smoothing (also known as Holt`s method) and Holt-Winters` method which accounts for trend and seasonality. The results of the empirical analysis showed that the optimal model for forecasting unemployment rate in Greece was Holt-Winters` additive method. In the case of Spain, according to MAPE, the optimal model was double exponential smoothing model. Furthermore, for Croatia and Italy the best forecasting model for unemployment rate was Holt-Winters` multiplicative model, whereas in the case of Portugal the best model to forecast unemployment rate was Double exponential smoothing model. Our findings are in line with European Commission unemployment rate estimates.Keywords: European Union countries, exponential smoothing methods, forecast accuracy unemployment rate
Procedia PDF Downloads 36923017 Developing a Multidimensional Adjustment Scale
Authors: Nadereh Sohrabi Shegefti, Siamak Samani
Abstract:
Level of adjustment is the first index to check mental health. The aim of this study was developing a valid and reliable Multidimensional Adjustment Scale (MAS). The sample consisted of 150 college students. Multidimensional adjustment scale and Depression, Anxiety, and stress scale (DASS) were used in this study. Principle factor analysis, Pearson correlation coefficient, and Cornbach's Alpha were used to check the validity and reliability of the MAS. Principle component factor analysis showed a 5 factor solution for the MAS. Alpha coefficients for the MAS sub scales were ranged between .69 to .83. Test-retest reliability for MAS was .88 and the mean of sub scales- total score correlation was .88. All these indexes revealed an acceptable reliability and validity for the MAS. The MAS is a short assessment instrument with good acceptable psychometric properties to use in clinical filed.Keywords: psychological adjustment, psychometric properties, validity, Pearson correlation
Procedia PDF Downloads 63423016 Flood-prone Urban Area Mapping Using Machine Learning, a Case Sudy of M'sila City (Algeria)
Authors: Medjadj Tarek, Ghribi Hayet
Abstract:
This study aims to develop a flood sensitivity assessment tool using machine learning (ML) techniques and geographic information system (GIS). The importance of this study is integrating the geographic information systems (GIS) and machine learning (ML) techniques for mapping flood risks, which help decision-makers to identify the most vulnerable areas and take the necessary precautions to face this type of natural disaster. To reach this goal, we will study the case of the city of M'sila, which is among the areas most vulnerable to floods. This study drew a map of flood-prone areas based on the methodology where we have made a comparison between 3 machine learning algorithms: the xGboost model, the Random Forest algorithm and the K Nearest Neighbour algorithm. Each of them gave an accuracy respectively of 97.92 - 95 - 93.75. In the process of mapping flood-prone areas, the first model was relied upon, which gave the greatest accuracy (xGboost).Keywords: Geographic information systems (GIS), machine learning (ML), emergency mapping, flood disaster management
Procedia PDF Downloads 9523015 Comparison of Various Landfill Ground Improvement Techniques for Redevelopment of Closed Landfills to Cater Transport Infrastructure
Authors: Michael D. Vinod, Hadi Khabbaz
Abstract:
Construction of infrastructure above or adjacent to landfills is becoming more common to capitalize on the limited space available within urban areas. However, development above landfills is a challenging task due to large voids, the presence of organic matter, heterogeneous nature of waste and ambiguity surrounding landfill settlement prediction. Prior to construction of infrastructure above landfills, ground improvement techniques are being employed to improve the geotechnical properties of landfill material. Although the ground improvement techniques have little impact on long term biodegradation and creep related landfill settlement, they have shown some notable short term success with a variety of techniques, including methods for verifying the level of effectiveness of ground improvement techniques. This paper provides geotechnical and landfill engineers a guideline for selection of landfill ground improvement techniques and their suitability to project-specific sites. Ground improvement methods assessed and compared in this paper include concrete injected columns (CIC), dynamic compaction, rapid impact compaction (RIC), preloading, high energy impact compaction (HEIC), vibro compaction, vibro replacement, chemical stabilization and the inclusion of geosynthetics such as geocells. For each ground improvement technique a summary of the existing theory, benefits, limitations, suitable modern ground improvement monitoring methods, the applicability of ground improvement techniques for landfills and supporting case studies are provided. The authors highlight the importance of implementing cost-effective monitoring techniques to allow observation and necessary remediation of the subsidence effects associated with long term landfill settlement. These ground improvement techniques are primarily for the purpose of construction above closed landfills to cater for transport infrastructure loading.Keywords: closed landfills, ground improvement, monitoring, settlement, transport infrastructure
Procedia PDF Downloads 22423014 Random Forest Classification for Population Segmentation
Authors: Regina Chua
Abstract:
To reduce the costs of re-fielding a large survey, a Random Forest classifier was applied to measure the accuracy of classifying individuals into their assigned segments with the fewest possible questions. Given a long survey, one needed to determine the most predictive ten or fewer questions that would accurately assign new individuals to custom segments. Furthermore, the solution needed to be quick in its classification and usable in non-Python environments. In this paper, a supervised Random Forest classifier was modeled on a dataset with 7,000 individuals, 60 questions, and 254 features. The Random Forest consisted of an iterative collection of individual decision trees that result in a predicted segment with robust precision and recall scores compared to a single tree. A random 70-30 stratified sampling for training the algorithm was used, and accuracy trade-offs at different depths for each segment were identified. Ultimately, the Random Forest classifier performed at 87% accuracy at a depth of 10 with 20 instead of 254 features and 10 instead of 60 questions. With an acceptable accuracy in prioritizing feature selection, new tools were developed for non-Python environments: a worksheet with a formulaic version of the algorithm and an embedded function to predict the segment of an individual in real-time. Random Forest was determined to be an optimal classification model by its feature selection, performance, processing speed, and flexible application in other environments.Keywords: machine learning, supervised learning, data science, random forest, classification, prediction, predictive modeling
Procedia PDF Downloads 9423013 Root Mean Square-Based Method for Fault Diagnosis and Fault Detection and Isolation of Current Fault Sensor in an Induction Machine
Authors: Ahmad Akrad, Rabia Sehab, Fadi Alyoussef
Abstract:
Nowadays, induction machines are widely used in industry thankful to their advantages comparing to other technologies. Indeed, there is a big demand because of their reliability, robustness and cost. The objective of this paper is to deal with diagnosis, detection and isolation of faults in a three-phase induction machine. Among the faults, Inter-turn short-circuit fault (ITSC), current sensors fault and single-phase open circuit fault are selected to deal with. However, a fault detection method is suggested using residual errors generated by the root mean square (RMS) of phase currents. The application of this method is based on an asymmetric nonlinear model of Induction Machine considering the winding fault of the three axes frame state space. In addition, current sensor redundancy and sensor fault detection and isolation (FDI) are adopted to ensure safety operation of induction machine drive. Finally, a validation is carried out by simulation in healthy and faulty operation modes to show the benefit of the proposed method to detect and to locate with, a high reliability, the three types of faults.Keywords: induction machine, asymmetric nonlinear model, fault diagnosis, inter-turn short-circuit fault, root mean square, current sensor fault, fault detection and isolation
Procedia PDF Downloads 19823012 Automatic Reporting System for Transcriptome Indel Identification and Annotation Based on Snapshot of Next-Generation Sequencing Reads Alignment
Authors: Shuo Mu, Guangzhi Jiang, Jinsa Chen
Abstract:
The analysis of Indel for RNA sequencing of clinical samples is easily affected by sequencing experiment errors and software selection. In order to improve the efficiency and accuracy of analysis, we developed an automatic reporting system for Indel recognition and annotation based on image snapshot of transcriptome reads alignment. This system includes sequence local-assembly and realignment, target point snapshot, and image-based recognition processes. We integrated high-confidence Indel dataset from several known databases as a training set to improve the accuracy of image processing and added a bioinformatical processing module to annotate and filter Indel artifacts. Subsequently, the system will automatically generate data, including data quality levels and images results report. Sanger sequencing verification of the reference Indel mutation of cell line NA12878 showed that the process can achieve 83% sensitivity and 96% specificity. Analysis of the collected clinical samples showed that the interpretation accuracy of the process was equivalent to that of manual inspection, and the processing efficiency showed a significant improvement. This work shows the feasibility of accurate Indel analysis of clinical next-generation sequencing (NGS) transcriptome. This result may be useful for RNA study for clinical samples with microsatellite instability in immunotherapy in the future.Keywords: automatic reporting, indel, next-generation sequencing, NGS, transcriptome
Procedia PDF Downloads 19123011 The Relation between Earnings Management with the Financial Reporting
Authors: Anocha Rojanapanich
Abstract:
The objective of this research is to investigate the effects of earnings management on corporate transparency of the company in Dusit area workplace via financial reporting reliability and stakeholder acceptance as independent variable. And the company in Dusit are are taken as the population and sample. The questionnaire is used to collect data. Exploratory Factor Analysis is implemented to ensure construct validity, and correlation statistic is selected to test the relationship among all variable and the ordinary least squares regression is used to explore the hypothesized. The results show that earnings management has a significant and negative impact on financial reporting reliability, stakeholder acceptance, and corporate transparency. Both financial reporting reliability and stakeholder acceptance have an important and positive effect on corporate transparency, and they are then mediators of the earnings management-corporate transparency relationships.Keywords: dusit area workplace, earnings management, financial report, business and marketing management
Procedia PDF Downloads 40623010 Optimization of Heat Insulation Structure and Heat Flux Calculation Method of Slug Calorimeter
Authors: Zhu Xinxin, Wang Hui, Yang Kai
Abstract:
Heat flux is one of the most important test parameters in the ground thermal protection test. Slug calorimeter is selected as the main sensor measuring heat flux in arc wind tunnel test due to the convenience and low cost. However, because of excessive lateral heat transfer and the disadvantage of the calculation method, the heat flux measurement error of the slug calorimeter is large. In order to enhance measurement accuracy, the heat insulation structure and heat flux calculation method of slug calorimeter were improved. The heat transfer model of the slug calorimeter was built according to the energy conservation principle. Based on the heat transfer model, the insulating sleeve of the hollow structure was designed, which helped to greatly decrease lateral heat transfer. And the slug with insulating sleeve of hollow structure was encapsulated using a package shell. The improved insulation structure reduced heat loss and ensured that the heat transfer characteristics were almost the same when calibrated and tested. The heat flux calibration test was carried out in arc lamp system for heat flux sensor calibration, and the results show that test accuracy and precision of slug calorimeter are improved greatly. In the meantime, the simulation model of the slug calorimeter was built. The heat flux values in different temperature rise time periods were calculated by the simulation model. The results show that extracting the data of the temperature rise rate as soon as possible can result in a smaller heat flux calculation error. Then the different thermal contact resistance affecting calculation error was analyzed by the simulation model. The contact resistance between the slug and the insulating sleeve was identified as the main influencing factor. The direct comparison calibration correction method was proposed based on only heat flux calibration. The numerical calculation correction method was proposed based on the heat flux calibration and simulation model of slug calorimeter after the simulation model was solved by solving the contact resistance between the slug and the insulating sleeve. The simulation and test results show that two methods can greatly reduce the heat flux measurement error. Finally, the improved slug calorimeter was tested in the arc wind tunnel. And test results show that the repeatability accuracy of improved slug calorimeter is less than 3%. The deviation of measurement value from different slug calorimeters is less than 3% in the same fluid field. The deviation of measurement value between slug calorimeter and Gordon Gage is less than 4% in the same fluid field.Keywords: correction method, heat flux calculation, heat insulation structure, heat transfer model, slug calorimeter
Procedia PDF Downloads 11723009 A Physically-Based Analytical Model for Reduced Surface Field Laterally Double Diffused MOSFETs
Authors: M. Abouelatta, A. Shaker, M. El-Banna, G. T. Sayah, C. Gontrand, A. Zekry
Abstract:
In this paper, a methodology for physically modeling the intrinsic MOS part and the drift region of the n-channel Laterally Double-diffused MOSFET (LDMOS) is presented. The basic physical effects like velocity saturation, mobility reduction, and nonuniform impurity concentration in the channel are taken into consideration. The analytical model is implemented using MATLAB. A comparison of the simulations from technology computer aided design (TCAD) and that from the proposed analytical model, at room temperature, shows a satisfactory accuracy which is less than 5% for the whole voltage domain.Keywords: LDMOS, MATLAB, RESURF, modeling, TCAD
Procedia PDF Downloads 19823008 Selection of Variogram Model for Environmental Variables
Authors: Sheikh Samsuzzhan Alam
Abstract:
The present study investigates the selection of variogram model in analyzing spatial variations of environmental variables with the trend. Sometimes, the autofitted theoretical variogram does not really capture the true nature of the empirical semivariogram. So proper exploration and analysis are needed to select the best variogram model. For this study, an open source data collected from California Soil Resource Lab1 is used to explain the problems when fitting a theoretical variogram. Five most commonly used variogram models: Linear, Gaussian, Exponential, Matern, and Spherical were fitted to the experimental semivariogram. Ordinary kriging methods were considered to evaluate the accuracy of the selected variograms through cross-validation. This study is beneficial for selecting an appropriate theoretical variogram model for environmental variables.Keywords: anisotropy, cross-validation, environmental variables, kriging, variogram models
Procedia PDF Downloads 33423007 Evaluation of the Impact of Reducing the Traffic Light Cycle for Cars to Improve Non-Vehicular Transportation: A Case of Study in Lima
Authors: Gheyder Concha Bendezu, Rodrigo Lescano Loli, Aldo Bravo Lizano
Abstract:
In big urbanized cities of Latin America, motor vehicles have priority over non-motor vehicles and pedestrians. There is an important problem that affects people's health and quality of life; lack of inclusion towards pedestrians makes it difficult for them to move smoothly and safely since the city has been planned for the transit of motor vehicles. Faced with the new trend for sustainable and economical transport, the city is forced to develop infrastructure in order to incorporate pedestrians and users with non-motorized vehicles in the transport system. The present research aims to study the influence of non-motorized vehicles on an avenue, the optimization of a cycle using traffic lights based on simulation in Synchro software, to improve the flow of non-motor vehicles. The evaluation is of the microscopic type; for this reason, field data was collected, such as vehicular, pedestrian, and non-motor vehicle user demand. With the values of speed and travel time, it is represented in the current scenario that contains the existing problem. These data allow to create a microsimulation model in Vissim software, later to be calibrated and validated so that it has a behavior similar to reality. The results of this model are compared with the efficiency parameters of the proposed model; these parameters are the queue length, the travel speed, and mainly the travel times of the users at this intersection. The results reflect a reduction of 27% in travel time, that is, an improvement between the proposed model and the current one for this great avenue. The tail length of motor vehicles is also reduced by 12.5%, a considerable improvement. All this represents an improvement in the level of service and in the quality of life of users.Keywords: bikeway, microsimulation, pedestrians, queue length, traffic light cycle, travel time
Procedia PDF Downloads 17323006 Experimental and Numerical Analyses of Tehran Research Reactor
Authors: A. Lashkari, H. Khalafi, H. Khazeminejad, S. Khakshourniya
Abstract:
In this paper, a numerical model is presented. The model is used to analyze a steady state thermo-hydraulic and reactivity insertion transient in TRR reference cores respectively. The model predictions are compared with the experiments and PARET code results. The model uses the piecewise constant and lumped parameter methods for the coupled point kinetics and thermal-hydraulics modules respectively. The advantages of the piecewise constant method are simplicity, efficiency and accuracy. A main criterion on the applicability range of this model is that the exit coolant temperature remains below the saturation temperature, i.e. no bulk boiling occurs in the core. The calculation values of power and coolant temperature, in steady state and positive reactivity insertion scenario, are in good agreement with the experiment values. However, the model is a useful tool for the transient analysis of most research reactor encountered in practice. The main objective of this work is using simple calculation methods and benchmarking them with experimental data. This model can be used for training proposes.Keywords: thermal-hydraulic, research reactor, reactivity insertion, numerical modeling
Procedia PDF Downloads 40123005 Designing a Model for Measuring the Components of Good Governance in the Iranian Higher Education System
Authors: Maria Ghorbanian, Mohammad Ghahramani, Mahmood Abolghasemi
Abstract:
Universities and institutions of higher education in Iran, like other higher education institutions in the world, have a heavy mission and task to educate students based on the needs of the country. Taking on such a serious responsibility requires having a good governance system for planning, formulating executive plans, evaluating, and finally modifying them in accordance with the current conditions and challenges ahead. In this regard, the present study was conducted with the aim of identifying the components of good governance in the Iranian higher education system by survey method and with a quantitative approach. In order to collect data, a researcher-made questionnaire was used, which includes two parts: personal and professional characteristics (5 questions) and the three components of good governance in the Iranian higher education system, including good management and leadership (8 items), continuous evaluation and effective (university performance, finance, and university appointments) (8 items) and civic responsibility and sustainable development (7 items). These variables were measured and coded in the form of a five-level Likert scale from "Very Low = 1" to "Very High = 5". First, the validity and reliability of the research model were examined. In order to calculate the reliability of the questionnaire, two methods of Cronbach's alpha and combined reliability were used. Fornell-Larker interaction and criterion were also used to determine the degree of diagnostic validity. The statistical population of this study included all faculty members of public universities in Tehran (N = 4429). The sample size was estimated to be 340 using the Cochran's formula. These numbers were studied using a randomized method with a proportional assignment. The data were analyzed by the structural equation method with the least-squares approach. The results showed that the component of civil responsibility and sustainable development with a factor load of 0.827 is the most important element of good governance.Keywords: good governance, higher education, sustainable, development
Procedia PDF Downloads 17123004 Synthetic Data-Driven Prediction Using GANs and LSTMs for Smart Traffic Management
Authors: Srinivas Peri, Siva Abhishek Sirivella, Tejaswini Kallakuri, Uzair Ahmad
Abstract:
Smart cities and intelligent transportation systems rely heavily on effective traffic management and infrastructure planning. This research tackles the data scarcity challenge by generating realistically synthetic traffic data from the PeMS-Bay dataset, enhancing predictive modeling accuracy and reliability. Advanced techniques like TimeGAN and GaussianCopula are utilized to create synthetic data that mimics the statistical and structural characteristics of real-world traffic. The future integration of Spatial-Temporal Generative Adversarial Networks (ST-GAN) is anticipated to capture both spatial and temporal correlations, further improving data quality and realism. Each synthetic data generation model's performance is evaluated against real-world data to identify the most effective models for accurately replicating traffic patterns. Long Short-Term Memory (LSTM) networks are employed to model and predict complex temporal dependencies within traffic patterns. This holistic approach aims to identify areas with low vehicle counts, reveal underlying traffic issues, and guide targeted infrastructure interventions. By combining GAN-based synthetic data generation with LSTM-based traffic modeling, this study facilitates data-driven decision-making that improves urban mobility, safety, and the overall efficiency of city planning initiatives.Keywords: GAN, long short-term memory (LSTM), synthetic data generation, traffic management
Procedia PDF Downloads 1423003 Simulation and Analysis of Passive Parameters of Building in eQuest: A Case Study in Istanbul, Turkey
Authors: Mahdiyeh Zafaranchi
Abstract:
With rapid development of urbanization and improvement of living standards in the world, energy consumption and carbon emissions of the building sector are expected to increase in the near future; because of that, energy-saving issues have become more important among the engineers. Besides, the building sector is a major contributor to energy consumption and carbon emissions. The concept of efficient building appeared as a response to the need for reducing energy demand in this sector which has the main purpose of shifting from standard buildings to low-energy buildings. Although energy-saving should happen in all steps of a building during the life cycle (material production, construction, demolition), the main concept of efficient energy building is saving energy during the life expectancy of a building by using passive and active systems, and should not sacrifice comfort and quality to reach these goals. The main aim of this study is to investigate passive strategies (do not need energy consumption or use renewable energy) to achieve energy-efficient buildings. Energy retrofit measures were explored by eQuest software using a case study as a base model. The study investigates predictive accuracy for the major factors like thermal transmittance (U-value) of the material, windows, shading devices, thermal insulation, rate of the exposed envelope, window/wall ration, lighting system in the energy consumption of the building. The base model was located in Istanbul, Turkey. The impact of eight passive parameters on energy consumption had been indicated. After analyzing the base model by eQuest, a final scenario was suggested which had a good energy performance. The results showed a decrease in the U-values of materials, the rate of exposing buildings, and windows had a significant effect on energy consumption. Finally, savings in electric consumption of about 10.5%, and gas consumption by about 8.37% in the suggested model were achieved annually.Keywords: efficient building, electric and gas consumption, eQuest, Passive parameters
Procedia PDF Downloads 11123002 Applications of Out-of-Sequence Thrust Movement for Earthquake Mitigation: A Review
Authors: Rajkumar Ghosh
Abstract:
The study presents an overview of the many uses and approaches for estimating out-of-sequence thrust movement in earthquake mitigation. The study investigates how knowing and forecasting thrust movement during seismic occurrences might assist to effective earthquake mitigation measures. The review begins by discussing out-of-sequence thrust movement and its importance in earthquake mitigation strategies. It explores how typical techniques of estimating thrust movement may not capture the full complexity of seismic occurrences and emphasizes the benefits of include out-of-sequence data in the analysis. A thorough review of existing research and studies on out-of-sequence thrust movement estimates for earthquake mitigation. The study demonstrates how to estimate out-of-sequence thrust movement using multiple data sources such as GPS measurements, satellite imagery, and seismic recordings. The study also examines the use of out-of-sequence thrust movement estimates in earthquake mitigation measures. It investigates how precise calculation of thrust movement may help improve structural design, analyse infrastructure risk, and develop early warning systems. The potential advantages of using out-of-sequence data in these applications to improve the efficiency of earthquake mitigation techniques. The difficulties and limits of estimating out-of-sequence thrust movement for earthquake mitigation. It addresses data quality difficulties, modelling uncertainties, and computational complications. To address these obstacles and increase the accuracy and reliability of out-of-sequence thrust movement estimates, the authors recommend topics for additional study and improvement. The study is a helpful resource for seismic monitoring and earthquake risk assessment researchers, engineers, and policymakers, supporting innovations in earthquake mitigation measures based on a better knowledge of thrust movement dynamics.Keywords: earthquake mitigation, out-of-sequence thrust, satellite imagery, seismic recordings, GPS measurements
Procedia PDF Downloads 8423001 Design Modification in CNC Milling Machine to Reduce the Weight of Structure
Authors: Harshkumar K. Desai, Anuj K. Desai, Jay P. Patel, Snehal V. Trivedi, Yogendrasinh Parmar
Abstract:
The need of continuous improvement in a product or process in this era of global competition leads to apply value engineering for functional and aesthetic improvement in consideration with economic aspect too. Solar industries located at G.I.D.C., Makarpura, Vadodara, Gujarat, India; a manufacturer of variety of CNC Machines had a challenge to analyze the structural design of column, base, carriage and table of CNC Milling Machine in the account of reduction of overall weight of a machine without affecting the rigidity and accuracy at the time of operation. The identified task is the first attempt to validate and optimize the proposed design of ribbed structure statically using advanced modeling and analysis tools in a systematic way. Results of stress and deformation obtained using analysis software are validated with theoretical analysis and found quite satisfactory. Such optimized results offer a weight reduction of the final assembly which is desired by manufacturers in favor of reduction of material cost, processing cost and handling cost finally.Keywords: CNC milling machine, optimization, finite element analysis (FEA), weight reduction
Procedia PDF Downloads 27623000 Blind Hybrid ARQ Retransmissions with Different Multiplexing between Time and Frequency for Ultra-Reliable Low-Latency Communications in 5G
Authors: Mohammad Tawhid Kawser, Ishrak Kabir, Sadia Sultana, Tanjim Ahmad
Abstract:
A promising service category of 5G, popularly known as Ultra-Reliable Low-Latency Communications (URLLC), is devoted to providing users with the staunchest fail-safe connections in the splits of a second. The reliability of data transfer, as offered by Hybrid ARQ (HARQ), should be employed as URLLC applications are highly error-sensitive. However, the delay added by HARQ ACK/NACK and retransmissions can degrade performance as URLLC applications are highly delay-sensitive too. To improve latency while maintaining reliability, this paper proposes the use of blind transmissions of redundancy versions exploiting the frequency diversity of wide bandwidth of 5G. The blind HARQ retransmissions proposed so far consider narrow bandwidth cases, for example, dedicated short range communication (DSRC), shared channels for device-to-device (D2D) communication, etc., and thus, do not gain much from the frequency diversity. The proposal also combines blind and ACK/NACK based retransmissions for different multiplexing options between time and frequency depending on the current radio channel quality and stringency of latency requirements. The wide bandwidth of 5G justifies that the proposed blind retransmission, without waiting for ACK/NACK, is not palpably extravagant. A simulation is performed to demonstrate the improvement in latency of the proposed scheme.Keywords: 5G, URLLC, HARQ, latency, frequency diversity
Procedia PDF Downloads 3622999 Harnessing Nigeria's Forestry Potential for Structural Applications: Structural Reliability of Nigerian Grown Opepe Timber
Authors: J. I. Aguwa, S. Sadiku, M. Abdullahi
Abstract:
This study examined the structural reliability of the Nigerian grown Opepe timber as bridge beam material. The strength of a particular specie of timber depends so much on some factors such as soil and environment in which it is grown. The steps involved are collection of the Opepe timber samples, seasoning/preparation of the test specimens, determination of the strength properties/statistical analysis, development of a computer programme in FORTRAN language and finally structural reliability analysis using FORM 5 software. The result revealed that the Nigerian grown Opepe is a reliable and durable structural bridge beam material for span of 5000mm, depth of 400mm, breadth of 250mm and end bearing length of 150mm. The probabilities of failure in bending parallel to the grain, compression perpendicular to the grain, shear parallel to the grain and deflection are 1.61 x 10-7, 1.43 x 10-8, 1.93 x 10-4 and 1.51 x 10-15 respectively. The paper recommends establishment of Opepe plantation in various Local Government Areas in Nigeria for structural applications such as in bridges, railway sleepers, generation of income to the nation as well as creating employment for the numerous unemployed youths.Keywords: bending and deflection, bridge beam, compression, Nigerian Opepe, shear, structural reliability
Procedia PDF Downloads 46722998 Metamorphic Computer Virus Classification Using Hidden Markov Model
Authors: Babak Bashari Rad
Abstract:
A metamorphic computer virus uses different code transformation techniques to mutate its body in duplicated instances. Characteristics and function of new instances are mostly similar to their parents, but they cannot be easily detected by the majority of antivirus in market, as they depend on string signature-based detection techniques. The purpose of this research is to propose a Hidden Markov Model for classification of metamorphic viruses in executable files. In the proposed solution, portable executable files are inspected to extract the instructions opcodes needed for the examination of code. A Hidden Markov Model trained on portable executable files is employed to classify the metamorphic viruses of the same family. The proposed model is able to generate and recognize common statistical features of mutated code. The model has been evaluated by examining the model on a test data set. The performance of the model has been practically tested and evaluated based on False Positive Rate, Detection Rate and Overall Accuracy. The result showed an acceptable performance with high average of 99.7% Detection Rate.Keywords: malware classification, computer virus classification, metamorphic virus, metamorphic malware, Hidden Markov Model
Procedia PDF Downloads 315