Search results for: gaussian selection operator
2620 Cache Analysis and Software Optimizations for Faster on-Chip Network Simulations
Authors: Khyamling Parane, B. M. Prabhu Prasad, Basavaraj Talawar
Abstract:
Fast simulations are critical in reducing time to market in CMPs and SoCs. Several simulators have been used to evaluate the performance and power consumed by Network-on-Chips. Researchers and designers rely upon these simulators for design space exploration of NoC architectures. Our experiments show that simulating large NoC topologies take hours to several days for completion. To speed up the simulations, it is necessary to investigate and optimize the hotspots in simulator source code. Among several simulators available, we choose Booksim2.0, as it is being extensively used in the NoC community. In this paper, we analyze the cache and memory system behaviour of Booksim2.0 to accurately monitor input dependent performance bottlenecks. Our measurements show that cache and memory usage patterns vary widely based on the input parameters given to Booksim2.0. Based on these measurements, the cache configuration having least misses has been identified. To further reduce the cache misses, we use software optimization techniques such as removal of unused functions, loop interchanging and replacing post-increment operator with pre-increment operator for non-primitive data types. The cache misses were reduced by 18.52%, 5.34% and 3.91% by employing above technology respectively. We also employ thread parallelization and vectorization to improve the overall performance of Booksim2.0. The OpenMP programming model and SIMD are used for parallelizing and vectorizing the more time-consuming portions of Booksim2.0. Speedups of 2.93x and 3.97x were observed for the Mesh topology with 30 × 30 network size by employing thread parallelization and vectorization respectively.Keywords: cache behaviour, network-on-chip, performance profiling, vectorization
Procedia PDF Downloads 1972619 A Comparative Study of Additive and Nonparametric Regression Estimators and Variable Selection Procedures
Authors: Adriano Z. Zambom, Preethi Ravikumar
Abstract:
One of the biggest challenges in nonparametric regression is the curse of dimensionality. Additive models are known to overcome this problem by estimating only the individual additive effects of each covariate. However, if the model is misspecified, the accuracy of the estimator compared to the fully nonparametric one is unknown. In this work the efficiency of completely nonparametric regression estimators such as the Loess is compared to the estimators that assume additivity in several situations, including additive and non-additive regression scenarios. The comparison is done by computing the oracle mean square error of the estimators with regards to the true nonparametric regression function. Then, a backward elimination selection procedure based on the Akaike Information Criteria is proposed, which is computed from either the additive or the nonparametric model. Simulations show that if the additive model is misspecified, the percentage of time it fails to select important variables can be higher than that of the fully nonparametric approach. A dimension reduction step is included when nonparametric estimator cannot be computed due to the curse of dimensionality. Finally, the Boston housing dataset is analyzed using the proposed backward elimination procedure and the selected variables are identified.Keywords: additive model, nonparametric regression, variable selection, Akaike Information Criteria
Procedia PDF Downloads 2642618 Application of Data Mining Techniques for Tourism Knowledge Discovery
Authors: Teklu Urgessa, Wookjae Maeng, Joong Seek Lee
Abstract:
Application of five implementations of three data mining classification techniques was experimented for extracting important insights from tourism data. The aim was to find out the best performing algorithm among the compared ones for tourism knowledge discovery. Knowledge discovery process from data was used as a process model. 10-fold cross validation method is used for testing purpose. Various data preprocessing activities were performed to get the final dataset for model building. Classification models of the selected algorithms were built with different scenarios on the preprocessed dataset. The outperformed algorithm tourism dataset was Random Forest (76%) before applying information gain based attribute selection and J48 (C4.5) (75%) after selection of top relevant attributes to the class (target) attribute. In terms of time for model building, attribute selection improves the efficiency of all algorithms. Artificial Neural Network (multilayer perceptron) showed the highest improvement (90%). The rules extracted from the decision tree model are presented, which showed intricate, non-trivial knowledge/insight that would otherwise not be discovered by simple statistical analysis with mediocre accuracy of the machine using classification algorithms.Keywords: classification algorithms, data mining, knowledge discovery, tourism
Procedia PDF Downloads 2952617 The Discussion on the Composition of Feng Shui by the Environmental Planning Viewpoint
Authors: Jhuang Jin-Jhong, Hsieh Wei-Fan
Abstract:
Climate change causes natural disasters persistently. Therefore, nowadays environmental planning objective tends to the issues of respecting nature and coexisting with nature. As a result, the natural environment analysis, e.g., the analysis of topography, soil, hydrology, climate, vegetation, is highly emphasized. On the other hand, Feng Shui has been a criterion of site selection for residence in Eastern since the ancient times and has had farther influence on site selection for castles and even for temples and tombs. The primary criterion of site selection is judging the quality of Long: mountain range, Sha: nearby mountains, Shui: hydrology, Xue: foundation, Xiang: aspect, which are similar to the environmental variables of mountain range, topography, hydrology and aspect. For the reason, a lot researchers attempt to probe into the connection between the criterion of Feng Shui and environmental planning factors. Most researches only discussed with the composition and theory of space of Feng Shui, but there is no research which explained Feng Shui through the environmental field. Consequently, this study reviewed the theory of Feng Shui through the environmental planning viewpoint and assembled essential composition factors of Feng Shui. The results of this study point. From literature review and comparison of theoretical meanings, we find that the ideal principles for planning the Feng Shui environment can also be used for environmental planning. Therefore, this article uses 12 ideal environmental features used in Feng Shui to contrast the natural aspects of the environment and make comparisons with previous research and classifies the environmental factors into climate, topography, hydrology, vegetation, and soil.Keywords: the composition of Feng Shui, environmental planning, site selection, main components of the Feng Shui environment
Procedia PDF Downloads 5092616 Investment Decision among Public Sector Retirees: A Behavioural Finance View
Authors: Bisi S. Olawoyin
Abstract:
This study attempts an exploration into behavioural finance in which the traditional assumptions of expected utility maximization with rational investors in efficient markets are dropped. It reviews prior research and evidence about how psychological biases affect investors behaviour and stock selection. This study examined the relationship between demographic variables and financial behaviour biases among public sector retirees who invested in the Nigerian Stock Exchange prior to their retirement. By using questionnaire survey method, a total of 214 valid convenient samples were collected in order to determine how specific demographic and psychological trait affect stock selection between dividend paying and non-dividend paying stocks. Descriptive statistics and OLS were used to analyse the results. Findings showed that most of the retirees prefer dividend paying stocks in few years preceding their retirement but still hold on to their non-dividend paying stock on retirement. A significant difference also exists between senior and junior retirees in preference for non-dividend paying stocks. These findings are consistent with the clientele theories of dividend.Keywords: behavioural finance, clientele theories, dividend paying stocks, stock selection
Procedia PDF Downloads 1412615 Fuzzy Population-Based Meta-Heuristic Approaches for Attribute Reduction in Rough Set Theory
Authors: Mafarja Majdi, Salwani Abdullah, Najmeh S. Jaddi
Abstract:
One of the global combinatorial optimization problems in machine learning is feature selection. It concerned with removing the irrelevant, noisy, and redundant data, along with keeping the original meaning of the original data. Attribute reduction in rough set theory is an important feature selection method. Since attribute reduction is an NP-hard problem, it is necessary to investigate fast and effective approximate algorithms. In this paper, we proposed two feature selection mechanisms based on memetic algorithms (MAs) which combine the genetic algorithm with a fuzzy record to record travel algorithm and a fuzzy controlled great deluge algorithm to identify a good balance between local search and genetic search. In order to verify the proposed approaches, numerical experiments are carried out on thirteen datasets. The results show that the MAs approaches are efficient in solving attribute reduction problems when compared with other meta-heuristic approaches.Keywords: rough set theory, attribute reduction, fuzzy logic, memetic algorithms, record to record algorithm, great deluge algorithm
Procedia PDF Downloads 4542614 Multiscale Entropy Analysis of Electroencephalogram (EEG) of Alcoholic and Control Subjects
Authors: Lal Hussain, Wajid Aziz, Imtiaz Ahmed Awan, Sharjeel Saeed
Abstract:
Multiscale entropy analysis (MSE) is a useful technique recently developed to quantify the dynamics of physiological signals at different time scales. This study is aimed at investigating the electroencephalogram (EEG) signals to analyze the background activity of alcoholic and control subjects by inspecting various coarse-grained sequences formed at different time scales. EEG recordings of alcoholic and control subjects were taken from the publically available machine learning repository of University of California (UCI) acquired using 64 electrodes. The MSE analysis was performed on the EEG data acquired from all the electrodes of alcoholic and control subjects. Mann-Whitney rank test was used to find significant differences between the groups and result were considered statistically significant for p-values<0.05. The area under receiver operator curve was computed to find the degree separation between the groups. The mean ranks of MSE values at all the times scales for all electrodes were higher control subject as compared to alcoholic subjects. Higher mean ranks represent higher complexity and vice versa. The finding indicated that EEG signals acquired through electrodes C3, C4, F3, F7, F8, O1, O2, P3, T7 showed significant differences between alcoholic and control subjects at time scales 1 to 5. Moreover, all electrodes exhibit significance level at different time scales. Likewise, the highest accuracy and separation was obtained at the central region (C3 and C4), front polar regions (P3, O1, F3, F7, F8 and T8) while other electrodes such asFp1, Fp2, P4 and F4 shows no significant results.Keywords: electroencephalogram (EEG), multiscale sample entropy (MSE), Mann-Whitney test (MMT), Receiver Operator Curve (ROC), complexity analysis
Procedia PDF Downloads 3762613 Timetabling for Interconnected LRT Lines: A Package Solution Based on a Real-world Case
Authors: Huazhen Lin, Ruihua Xu, Zhibin Jiang
Abstract:
In this real-world case, timetabling the LRT network as a whole is rather challenging for the operator: they are supposed to create a timetable to avoid various route conflicts manually while satisfying a given interval and the number of rolling stocks, but the outcome is not satisfying. Therefore, the operator adopts a computerised timetabling tool, the Train Plan Maker (TPM), to cope with this problem. However, with various constraints in the dual-line network, it is still difficult to find an adequate pairing of turnback time, interval and rolling stocks’ number, which requires extra manual intervention. Aiming at current problems, a one-off model for timetabling is presented in this paper to simplify the procedure of timetabling. Before the timetabling procedure starts, this paper presents how the dual-line system with a ring and several branches is turned into a simpler structure. Then, a non-linear programming model is presented in two stages. In the first stage, the model sets a series of constraints aiming to calculate a proper timing for coordinating two lines by adjusting the turnback time at termini. Then, based on the result of the first stage, the model introduces a series of inequality constraints to avoid various route conflicts. With this model, an analysis is conducted to reveal the relation between the ratio of trains in different directions and the possible minimum interval, observing that the more imbalance the ratio is, the less possible to provide frequent service under such strict constraints.Keywords: light rail transit (LRT), non-linear programming, railway timetabling, timetable coordination
Procedia PDF Downloads 872612 Switching Losses in Power Electronic Converter of Switched Reluctance Motor
Authors: Ali Asghar Memon
Abstract:
A cautious and astute selection of switching devices used in power electronic converters of a switched reluctance (SR) motor is required. It is a matter of choice of best switching devices with respect to their switching ability rather than fulfilling the number of switches. This paper highlights the computational determination of switching losses comprising of switch-on, switch-off and conduction losses respectively by using experimental data in simulation model of a SR machine. The finding of this research is helpful for proper selection of electronic switches and suitable converter topology for switched reluctance motor.Keywords: converter, operating modes, switched reluctance motor, switching losses
Procedia PDF Downloads 5072611 Assessment of Relationships between Agro-Morphological Traits and Cold Tolerance in Faba Bean (vicia faba l.) and Wild Relatives
Authors: Nisa Ertoy Inci, Cengiz Toker
Abstract:
Winter or autumn-sown faba bean (Vicia faba L.) is one the most efficient ways to overcome drought since faba bean is usually grown under rainfed where drought and high-temperature stresses are the main growth constraints. The objectives of this study were assessment of (i) relationships between cold tolerance and agro-morphological traits, and (ii) the most suitable agro-morphological trait(s) under cold conditions. Three species of the genus Vicia L. includes 109 genotypes of faba bean (Vicia faba L.), three genotypes of narbon bean (V. narbonensis L.) and two genotypes of V. montbretii Fisch. & C.A. Mey. Davis and Plitmann were sown in autumn at highland of Mediterranean region of Turkey. All relatives of faba bean were more cold-tolerant than the faba bean genotypes. Three faba bean genotypes, ACV-42, ACV-84 and ACV-88, were selected as sources of cold tolerance under field conditions. Path and correlation coefficients and factor and principal component analyses indicated that biological yield should be evaluated in selection for cold tolerance under cold conditions ahead of many agro-morphological traits. The seed weight should be considered for selection in early breeding generations because they had the highest heritability.Keywords: cold tolerance, faba bean, narbon bean, selection
Procedia PDF Downloads 3982610 Porul: Option Generation and Selection and Scoring Algorithms for a Tamil Flash Card Game
Authors: Anitha Narasimhan, Aarthy Anandan, Madhan Karky, C. N. Subalalitha
Abstract:
Games can be the excellent tools for teaching a language. There are few e-learning games in Indian languages like word scrabble, cross word, quiz games etc., which were developed mainly for educational purposes. This paper proposes a Tamil word game called, “Porul”, which focuses on education as well as on players’ thinking and decision-making skills. Porul is a multiple choice based quiz game, in which the players attempt to answer questions correctly from the given multiple options that are generated using a unique algorithm called the Option Selection algorithm which explores the semantics of the question in various dimensions namely, synonym, rhyme and Universal Networking Language semantic category. This kind of semantic exploration of the question not only increases the complexity of the game but also makes it more interesting. The paper also proposes a Scoring Algorithm which allots a score based on the popularity score of the question word. The proposed game has been tested using 20,000 Tamil words.Keywords: Porul game, Tamil word game, option selection, flash card, scoring, algorithm
Procedia PDF Downloads 4022609 Double Gaussian Distribution of Nonhomogeneous Barrier Height in Metal/n-type GaN Schottky Contacts
Authors: M. Mamor
Abstract:
GaN-based compounds have attracted much interest in the fabrication of high-power, high speed and high-frequency electronic devices. Other examples of GaN-based applications are blue and ultraviolet (UV) light-emitting diodes (LEDs). All these devices require high-quality ohmic and Schottky contacts. Gaining an understanding of the electrical characteristics of metal/GaN contacts is of fundamental and technological importance for developing GaN-based devices. In this work, the barrier characteristics of Pt and Pd Schottky contacts on n-type GaN were studied using temperature-dependent forward current-voltage (I-V) measurements over a wide temperature range 80–400 K. Our results show that the barrier height and ideality factor, extracted from the forward I-V characteristics based on thermionic emission (TE) model, exhibit an abnormal dependence with temperature; i.e., by increasing temperature, the barrier height increases whereas the ideality factor decreases. This abnormal behavior has been explained based on the TE model by considering the presence of double Gaussian distribution (GD) of nonhomogeneous barrier height at the metal/GaN interface. However, in the high-temperature range (160-400 K), the extracted value for the effective Richardson constant A* based on the barrier inhomogeneity (BHi) model is found in fair agreement with the theoretically predicted value of about 26.9 A.cm-2 K-2 for n-type GaN. This result indicates that in this temperature range, the conduction current transport is dominated by the thermionic emission mode. On the other hand, in the lower temperature range (80-160 K), the corresponding effective Richardson constant value according to the BHi model is lower than the theoretical value, suggesting the presence of other current transport, such as tunneling-assisted mode at lower temperatures.Keywords: Schottky diodes, inhomogeneous barrier height, GaN semiconductors, Schottky barrier heights
Procedia PDF Downloads 552608 A Framework for an Automated Decision Support System for Selecting Safety-Conscious Contractors
Authors: Rawan A. Abdelrazeq, Ahmed M. Khalafallah, Nabil A. Kartam
Abstract:
Selection of competent contractors for construction projects is usually accomplished through competitive bidding or negotiated contracting in which the contract bid price is the basic criterion for selection. The evaluation of contractor’s safety performance is still not a typical criterion in the selection process, despite the existence of various safety prequalification procedures. There is a critical need for practical and automated systems that enable owners and decision makers to evaluate contractor safety performance, among other important contractor selection criteria. These systems should ultimately favor safety-conscious contractors to be selected by the virtue of their past good safety records and current safety programs. This paper presents an exploratory sequential mixed-methods approach to develop a framework for an automated decision support system that evaluates contractor safety performance based on a multitude of indicators and metrics that have been identified through a comprehensive review of construction safety research, and a survey distributed to domain experts. The framework is developed in three phases: (1) determining the indicators that depict contractor current and past safety performance; (2) soliciting input from construction safety experts regarding the identified indicators, their metrics, and relative significance; and (3) designing a decision support system using relational database models to integrate the identified indicators and metrics into a system that assesses and rates the safety performance of contractors. The proposed automated system is expected to hold several advantages including: (1) reducing the likelihood of selecting contractors with poor safety records; (2) enhancing the odds of completing the project safely; and (3) encouraging contractors to exert more efforts to improve their safety performance and practices in order to increase their bid winning opportunities which can lead to significant safety improvements in the construction industry. This should prove useful to decision makers and researchers, alike, and should help improve the safety record of the construction industry.Keywords: construction safety, contractor selection, decision support system, relational database
Procedia PDF Downloads 2802607 Design of Target Selection for Pedestrian Autonomous Emergency Braking System
Authors: Tao Song, Hao Cheng, Guangfeng Tian, Chuang Xu
Abstract:
An autonomous emergency braking system is an advanced driving assistance system that enables vehicle collision avoidance and pedestrian collision avoidance to improve vehicle safety. At present, because the pedestrian target is small, and the mobility is large, the pedestrian AEB system is faced with more technical difficulties and higher functional requirements. In this paper, a method of pedestrian target selection based on a variable width funnel is proposed. Based on the current position and predicted position of pedestrians, the relative position of vehicle and pedestrian at the time of collision is calculated, and different braking strategies are adopted according to the hazard level of pedestrian collisions. In the CNCAP standard operating conditions, comparing the method of considering only the current position of pedestrians and the method of considering pedestrian prediction position, as well as the method based on fixed width funnel and variable width funnel, the results show that, based on variable width funnel, the choice of pedestrian target will be more accurate and the opportunity of the intervention of AEB system will be more reasonable by considering the predicted position of the pedestrian target and vehicle's lateral motion.Keywords: automatic emergency braking system, pedestrian target selection, TTC, variable width funnel
Procedia PDF Downloads 1572606 Normalizing Flow to Augmented Posterior: Conditional Density Estimation with Interpretable Dimension Reduction for High Dimensional Data
Authors: Cheng Zeng, George Michailidis, Hitoshi Iyatomi, Leo L. Duan
Abstract:
The conditional density characterizes the distribution of a response variable y given other predictor x and plays a key role in many statistical tasks, including classification and outlier detection. Although there has been abundant work on the problem of Conditional Density Estimation (CDE) for a low-dimensional response in the presence of a high-dimensional predictor, little work has been done for a high-dimensional response such as images. The promising performance of normalizing flow (NF) neural networks in unconditional density estimation acts as a motivating starting point. In this work, the authors extend NF neural networks when external x is present. Specifically, they use the NF to parameterize a one-to-one transform between a high-dimensional y and a latent z that comprises two components [zₚ, zₙ]. The zₚ component is a low-dimensional subvector obtained from the posterior distribution of an elementary predictive model for x, such as logistic/linear regression. The zₙ component is a high-dimensional independent Gaussian vector, which explains the variations in y not or less related to x. Unlike existing CDE methods, the proposed approach coined Augmented Posterior CDE (AP-CDE) only requires a simple modification of the common normalizing flow framework while significantly improving the interpretation of the latent component since zₚ represents a supervised dimension reduction. In image analytics applications, AP-CDE shows good separation of 𝑥-related variations due to factors such as lighting condition and subject id from the other random variations. Further, the experiments show that an unconditional NF neural network based on an unsupervised model of z, such as a Gaussian mixture, fails to generate interpretable results.Keywords: conditional density estimation, image generation, normalizing flow, supervised dimension reduction
Procedia PDF Downloads 962605 Instance Selection for MI-Support Vector Machines
Authors: Amy M. Kwon
Abstract:
Support vector machine (SVM) is a well-known algorithm in machine learning due to its superior performance, and it also functions well in multiple-instance (MI) problems. Our study proposes a schematic algorithm to select instances based on Hausdorff distance, which can be adapted to SVMs as input vectors under the MI setting. Based on experiments on five benchmark datasets, our strategy for adapting representation outperformed in comparison with original approach. In addition, task execution times (TETs) were reduced by more than 80% based on MissSVM. Hence, it is noteworthy to consider this representation adaptation to SVMs under MI-setting.Keywords: support vector machine, Margin, Hausdorff distance, representation selection, multiple-instance learning, machine learning
Procedia PDF Downloads 342604 The Economic Value of Mastitis Resistance in Dairy Cattle in Kenya
Authors: Caleb B. Sagwa, Tobias O. Okeno, Alexander K. Kahi
Abstract:
Dairy cattle production plays an important role in the Kenyan economy. However, high incidences of mastitis is a major setback to the productivity in this industry. The current dairy cattle breeding objective in Kenya does not include mastitis resistance, mainly because the economic value of mastitis resistance has not been determined. Therefore this study aimed at estimating the economic value of mastitis resistance in dairy cattle in Kenya. Initial input parameters were obtained from literature on dairy cattle production systems in the tropics. Selection index methodology was used to derive the economic value of mastitis resistance. Somatic cell count (SCC) was used an indicator trait for mastitis resistance. The economic value was estimated relative to milk yield (MY). Economic values were assigned to SCC in a selection index such that the overall gain in the breeding goal trait was maximized. The option of estimating the economic value for SCC by equating the response in the trait of interest to its index response was considered. The economic value of mastitis resistance was US $23.64 while maximum response to selection for MY was US $66.01. The findings of this study provide vital information that is a pre-requisite for the inclusion of mastitis resistance in the current dairy cattle breeding goal in Kenya.Keywords: somatic cell count, milk quality, payment system, breeding goal
Procedia PDF Downloads 2612603 Analysis of Particle Reinforced Metal Matrix Composite Crankshaft
Authors: R. S. Vikaash, S. Vinodh, T. S. Sai Prashanth
Abstract:
Six sigma is a defect reduction strategy enabling modern organizations to achieve business prosperity. The practitioners are in need to select best six sigma project among the available alternatives to achieve customer satisfaction. In this circumstance, this article presents a study in which six sigma project selection is formulated as Multi-Criteria Decision-Making(MCDM) problem and the best project has been found using AHP. Five main governing criteria and 14 sub criteria are being formulated. The decision maker’s inputs were gathered and computations were performed. The project with the high values from the set of projects is selected as the best project. Based on calculations, Project “P1”is found to be the best and further deployment actions have been undertaken in the organization.Keywords: six Sigma, project selection, MCDM, analytic hierarchy process, business prosperity
Procedia PDF Downloads 3422602 Development of Building Information Modeling in Property Industry: Beginning with Building Information Modeling Construction
Authors: B. Godefroy, D. Beladjine, K. Beddiar
Abstract:
In France, construction BIM actors commonly evoke the BIM gains for exploitation by integrating of the life cycle of a building. The standardization of level 7 of development would achieve this stage of the digital model. The householders include local public authorities, social landlords, public institutions (health and education), enterprises, facilities management companies. They have a dual role: owner and manager of their housing complex. In a context of financial constraint, the BIM of exploitation aims to control costs, make long-term investment choices, renew the portfolio and enable environmental standards to be met. It assumes a knowledge of the existing buildings, marked by its size and complexity. The information sought must be synthetic and structured, it concerns, in general, a real estate complex. We conducted a study with professionals about their concerns and ways to use it to see how householders could benefit from this development. To obtain results, we had in mind the recurring interrogation of the project management, on the needs of the operators, we tested the following stages: 1) Inculcate a minimal culture of BIM with multidisciplinary teams of the operator then by business, 2) Learn by BIM tools, the adaptation of their trade in operations, 3) Understand the place and creation of a graphic and technical database management system, determine the components of its library so their needs, 4) Identify the cross-functional interventions of its managers by business (operations, technical, information system, purchasing and legal aspects), 5) Set an internal protocol and define the BIM impact in their digital strategy. In addition, continuity of management by the integration of construction models in the operation phase raises the question of interoperability in the control of the production of IFC files in the operator’s proprietary format and the export and import processes, a solution rivaled by the traditional method of vectorization of paper plans. Companies that digitize housing complex and those in FM produce a file IFC, directly, according to their needs without recourse to the model of construction, they produce models business for the exploitation. They standardize components, equipment that are useful for coding. We observed the consequences resulting from the use of the BIM in the property industry and, made the following observations: a) The value of data prevail over the graphics, 3D is little used b) The owner must, through his organization, promote the feedback of technical management information during the design phase c) The operator's reflection on outsourcing concerns the acquisition of its information system and these services, observing the risks and costs related to their internal or external developments. This study allows us to highlight: i) The need for an internal organization of operators prior to a response to the construction management ii) The evolution towards automated methods for creating models dedicated to the exploitation, a specialization would be required iii) A review of the communication of the project management, management continuity not articulating around his building model, it must take into account the environment of the operator and reflect on its scope of action.Keywords: information system, interoperability, models for exploitation, property industry
Procedia PDF Downloads 1442601 Design Components and Reliability Aspects of Municipal Waste Water and SEIG Based Micro Hydro Power Plant
Authors: R. K. Saket
Abstract:
This paper presents design aspects and probabilistic approach for generation reliability evaluation of an alternative resource: municipal waste water based micro hydro power generation system. Annual and daily flow duration curves have been obtained for design, installation, development, scientific analysis and reliability evaluation of the MHPP. The hydro potential of the waste water flowing through sewage system of the BHU campus has been determined to produce annual flow duration and daily flow duration curves by ordering the recorded water flows from maximum to minimum values. Design pressure, the roughness of the pipe’s interior surface, method of joining, weight, ease of installation, accessibility to the sewage system, design life, maintenance, weather conditions, availability of material, related cost and likelihood of structural damage have been considered for design of a particular penstock for reliable operation of the MHPP. A MHPGS based on MWW and SEIG is designed, developed, and practically implemented to provide reliable electric energy to suitable load in the campus of the Banaras Hindu University, Varanasi, (UP), India. Generation reliability evaluation of the developed MHPP using Gaussian distribution approach, safety factor concept, peak load consideration and Simpson 1/3rd rule has presented in this paper.Keywords: self excited induction generator, annual and daily flow duration curve, sewage system, municipal waste water, reliability evaluation, Gaussian distribution, Simpson 1/3rd rule
Procedia PDF Downloads 5582600 Motion Capture Based Wizard of Oz Technique for Humanoid Robot
Authors: Rafal Stegierski, Krzysztof Dmitruk
Abstract:
The paper focuses on robotic tele-presence system build around humanoid robot operated with controller-less Wizard of Oz technique. Proposed solution gives possibility to quick start acting as a operator with short, if any, initial training.Keywords: robotics, motion capture, Wizard of Oz, humanoid robots, human robot interaction
Procedia PDF Downloads 4812599 Auto-Tuning of CNC Parameters According to the Machining Mode Selection
Authors: Jenq-Shyong Chen, Ben-Fong Yu
Abstract:
CNC(computer numerical control) machining centers have been widely used for machining different metal components for various industries. For a specific CNC machine, its everyday job is assigned to cut different products with quite different attributes such as material type, workpiece weight, geometry, tooling, and cutting conditions. Theoretically, the dynamic characteristics of the CNC machine should be properly tuned match each machining job in order to get the optimal machining performance. However, most of the CNC machines are set with only a standard set of CNC parameters. In this study, we have developed an auto-tuning system which can automatically change the CNC parameters and in hence change the machine dynamic characteristics according to the selection of machining modes which are set by the mixed combination of three machine performance indexes: the HO (high surface quality) index, HP (high precision) index and HS (high speed) index. The acceleration, jerk, corner error tolerance, oscillation and dynamic bandwidth of machine’s feed axes have been changed according to the selection of the machine performance indexes. The proposed auto-tuning system of the CNC parameters has been implemented on a PC-based CNC controller and a three-axis machining center. The measured experimental result have shown the promising of our proposed auto-tuning system.Keywords: auto-tuning, CNC parameters, machining mode, high speed, high accuracy, high surface quality
Procedia PDF Downloads 3802598 Implicit U-Net Enhanced Fourier Neural Operator for Long-Term Dynamics Prediction in Turbulence
Authors: Zhijie Li, Wenhui Peng, Zelong Yuan, Jianchun Wang
Abstract:
Turbulence is a complex phenomenon that plays a crucial role in various fields, such as engineering, atmospheric science, and fluid dynamics. Predicting and understanding its behavior over long time scales have been challenging tasks. Traditional methods, such as large-eddy simulation (LES), have provided valuable insights but are computationally expensive. In the past few years, machine learning methods have experienced rapid development, leading to significant improvements in computational speed. However, ensuring stable and accurate long-term predictions remains a challenging task for these methods. In this study, we introduce the implicit U-net enhanced Fourier neural operator (IU-FNO) as a solution for stable and efficient long-term predictions of the nonlinear dynamics in three-dimensional (3D) turbulence. The IU-FNO model combines implicit re-current Fourier layers to deepen the network and incorporates the U-Net architecture to accurately capture small-scale flow structures. We evaluate the performance of the IU-FNO model through extensive large-eddy simulations of three types of 3D turbulence: forced homogeneous isotropic turbulence (HIT), temporally evolving turbulent mixing layer, and decaying homogeneous isotropic turbulence. The results demonstrate that the IU-FNO model outperforms other FNO-based models, including vanilla FNO, implicit FNO (IFNO), and U-net enhanced FNO (U-FNO), as well as the dynamic Smagorinsky model (DSM), in predicting various turbulence statistics. Specifically, the IU-FNO model exhibits improved accuracy in predicting the velocity spectrum, probability density functions (PDFs) of vorticity and velocity increments, and instantaneous spatial structures of the flow field. Furthermore, the IU-FNO model addresses the stability issues encountered in long-term predictions, which were limitations of previous FNO models. In addition to its superior performance, the IU-FNO model offers faster computational speed compared to traditional large-eddy simulations using the DSM model. It also demonstrates generalization capabilities to higher Taylor-Reynolds numbers and unseen flow regimes, such as decaying turbulence. Overall, the IU-FNO model presents a promising approach for long-term dynamics prediction in 3D turbulence, providing improved accuracy, stability, and computational efficiency compared to existing methods.Keywords: data-driven, Fourier neural operator, large eddy simulation, fluid dynamics
Procedia PDF Downloads 742597 Weyl Type Theorem and the Fuglede Property
Authors: M. H. M. Rashid
Abstract:
Given H a Hilbert space and B(H) the algebra of bounded linear operator in H, let δAB denote the generalized derivation defined by A and B. The main objective of this article is to study Weyl type theorems for generalized derivation for (A,B) satisfying a couple of Fuglede.Keywords: Fuglede Property, Weyl’s theorem, generalized derivation, Aluthge transform
Procedia PDF Downloads 1282596 Supplier Selection in a Scenario Based Stochastic Model with Uncertain Defectiveness and Delivery Lateness Rates
Authors: Abeer Amayri, Akif A. Bulgak
Abstract:
Due to today’s globalization as well as outsourcing practices of the companies, the Supply Chain (SC) performances have become more dependent on the efficient movement of material among places that are geographically dispersed, where there is more chance for disruptions. One such disruption is the quality and delivery uncertainties of outsourcing. These uncertainties could lead the products to be unsafe and, as is the case in a number of recent examples, companies may have to end up in recalling their products. As a result of these problems, there is a need to develop a methodology for selecting suppliers globally in view of risks associated with low quality and late delivery. Accordingly, we developed a two-stage stochastic model that captures the risks associated with uncertainty in quality and delivery as well as a solution procedure for the model. The stochastic model developed simultaneously optimizes supplier selection and purchase quantities under price discounts over a time horizon. In particular, our target is the study of global organizations with multiple sites and multiple overseas suppliers, where the pricing is offered in suppliers’ local currencies. Our proposed methodology is applied to a case study for a US automotive company having two assembly plants and four potential global suppliers to illustrate how the proposed model works in practice.Keywords: global supply chains, quality, stochastic programming, supplier selection
Procedia PDF Downloads 4582595 A Hybrid Feature Selection and Deep Learning Algorithm for Cancer Disease Classification
Authors: Niousha Bagheri Khulenjani, Mohammad Saniee Abadeh
Abstract:
Learning from very big datasets is a significant problem for most present data mining and machine learning algorithms. MicroRNA (miRNA) is one of the important big genomic and non-coding datasets presenting the genome sequences. In this paper, a hybrid method for the classification of the miRNA data is proposed. Due to the variety of cancers and high number of genes, analyzing the miRNA dataset has been a challenging problem for researchers. The number of features corresponding to the number of samples is high and the data suffer from being imbalanced. The feature selection method has been used to select features having more ability to distinguish classes and eliminating obscures features. Afterward, a Convolutional Neural Network (CNN) classifier for classification of cancer types is utilized, which employs a Genetic Algorithm to highlight optimized hyper-parameters of CNN. In order to make the process of classification by CNN faster, Graphics Processing Unit (GPU) is recommended for calculating the mathematic equation in a parallel way. The proposed method is tested on a real-world dataset with 8,129 patients, 29 different types of tumors, and 1,046 miRNA biomarkers, taken from The Cancer Genome Atlas (TCGA) database.Keywords: cancer classification, feature selection, deep learning, genetic algorithm
Procedia PDF Downloads 1112594 Bayesian Structural Identification with Systematic Uncertainty Using Multiple Responses
Authors: André Jesus, Yanjie Zhu, Irwanda Laory
Abstract:
Structural health monitoring is one of the most promising technologies concerning aversion of structural risk and economic savings. Analysts often have to deal with a considerable variety of uncertainties that arise during a monitoring process. Namely the widespread application of numerical models (model-based) is accompanied by a widespread concern about quantifying the uncertainties prevailing in their use. Some of these uncertainties are related with the deterministic nature of the model (code uncertainty) others with the variability of its inputs (parameter uncertainty) and the discrepancy between a model/experiment (systematic uncertainty). The actual process always exhibits a random behaviour (observation error) even when conditions are set identically (residual variation). Bayesian inference assumes that parameters of a model are random variables with an associated PDF, which can be inferred from experimental data. However in many Bayesian methods the determination of systematic uncertainty can be problematic. In this work systematic uncertainty is associated with a discrepancy function. The numerical model and discrepancy function are approximated by Gaussian processes (surrogate model). Finally, to avoid the computational burden of a fully Bayesian approach the parameters that characterise the Gaussian processes were estimated in a four stage process (modular Bayesian approach). The proposed methodology has been successfully applied on fields such as geoscience, biomedics, particle physics but never on the SHM context. This approach considerably reduces the computational burden; although the extent of the considered uncertainties is lower (second order effects are neglected). To successfully identify the considered uncertainties this formulation was extended to consider multiple responses. The efficiency of the algorithm has been tested on a small scale aluminium bridge structure, subjected to a thermal expansion due to infrared heaters. Comparison of its performance with responses measured at different points of the structure and associated degrees of identifiability is also carried out. A numerical FEM model of the structure was developed and the stiffness from its supports is considered as a parameter to calibrate. Results show that the modular Bayesian approach performed best when responses of the same type had the lowest spatial correlation. Based on previous literature, using different types of responses (strain, acceleration, and displacement) should also improve the identifiability problem. Uncertainties due to parametric variability, observation error, residual variability, code variability and systematic uncertainty were all recovered. For this example the algorithm performance was stable and considerably quicker than Bayesian methods that account for the full extent of uncertainties. Future research with real-life examples is required to fully access the advantages and limitations of the proposed methodology.Keywords: bayesian, calibration, numerical model, system identification, systematic uncertainty, Gaussian process
Procedia PDF Downloads 3252593 A Range of Steel Production in Japan towards 2050
Authors: Reina Kawase
Abstract:
Japan set the goal of 80% reduction in GHG emissions by 2050. To consider countermeasures for reducing GHG emission, the production estimation of energy intensive materials, such as steel, is essential. About 50% of steel production is exported in Japan, so it is necessary to consider steel production including export. Steel productions from 2005-2050 in Japan were estimated under various global assumptions based on combination of scenarios such as goods trade scenarios and steel making process selection scenarios. Process selection scenarios decide volume of steel production by process (basic oxygen furnace and electric arc furnace) with considering steel consumption projection, supply-demand balance of steel, and scrap surplus. The range of steel production by process was analyzed. Maximum steel production was estimated under the scenario which consumes scrap in domestic steel production at maximum level. In 2035, steel production reaches 149 million ton because of increase in electric arc furnace steel. However, it decreases towards 2050 and amounts to 120 million ton, which is almost same as a current level. Minimum steel production is under the scenario which assumes technology progress in steel making and supply-demand balance consideration in each region. Steel production decreases from base year and is 44 million ton in 2050.Keywords: goods trade scenario, steel making process selection scenario, steel production, global warming
Procedia PDF Downloads 3832592 Factors Influencing the Logistics Services Providers' Performance: A Literature Overview
Authors: A. Aguezzoul
Abstract:
The Logistics Services Providers (LSPs) selection and performance is a strategic decision that affects the overall performance of any company as well as its supply chain. It is a complex process, which takes into account various conflicting quantitative and qualitative factors, as well as outsourced logistics activities. This article focuses on the evolution of the weights associated to these factors over the last years in order to better understand the change in the importance that logistics professionals place on them criteria when choosing their LSPs. For that, an analysis of 17 main studies published during 2014-2017 period was carried out and the results are compared to those of a previous literature review on this subject. Our analysis allowed us to deduce the following observations: 1) the LSPs selection is a multi-criteria process; 2) the empirical character of the majority of studies, conducted particularly in Asian countries; 3) the criteria importance has undergone significant changes following the emergence of information technologies that have favored the work in close collaboration and in partnership between the LSPs and their customers, even on a worldwide scale; 4) the cost criterion is relatively less important than in the past; and finally 5) with the development of sustainable supply chains, the factors associated with the logistic activities of return and waste processing (reverse logistics) are becoming increasingly important in this multi-criteria process of selection and evaluation of LSPs performance.Keywords: logistics outsourcing, logistics providers, multi-criteria decision making, performance
Procedia PDF Downloads 1542591 Flat-Top Apodization of Laser Beams by Means of Acousto-Optics
Authors: Sergey I. Chizhikov, Vladimir Y. Molchanov, Konstantin B. Yushkov
Abstract:
We demonstrate a method for adaptive spatial shaping of laser beams by means of acousto-optic Bragg diffraction. Transformation of the angular spectrum during Bragg diffraction is used to convert Gaussian intensity distribution into a flat-top one. Theoretical model is supported by the experiment.Keywords: acousto-optics, flat top, beam shaping, Bragg diffraction
Procedia PDF Downloads 625