Search results for: Early Decision.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2028

Search results for: Early Decision.

678 Welding Process Selection for Storage Tank by Integrated Data Envelopment Analysis and Fuzzy Credibility Constrained Programming Approach

Authors: Rahmad Wisnu Wardana, Eakachai Warinsiriruk, Sutep Joy-A-Ka

Abstract:

Selecting the most suitable welding process usually depends on experiences or common application in similar companies. However, this approach generally ignores many criteria that can be affecting the suitable welding process selection. Therefore, knowledge automation through knowledge-based systems will significantly improve the decision-making process. The aims of this research propose integrated data envelopment analysis (DEA) and fuzzy credibility constrained programming approach for identifying the best welding process for stainless steel storage tank in the food and beverage industry. The proposed approach uses fuzzy concept and credibility measure to deal with uncertain data from experts' judgment. Furthermore, 12 parameters are used to determine the most appropriate welding processes among six competitive welding processes.

Keywords: Welding process selection, data envelopment analysis, fuzzy credibility constrained programming, storage tank.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 793
677 WebAppShield: An Approach Exploiting Machine Learning to Detect SQLi Attacks in an Application Layer in Run-Time

Authors: Ahmed Abdulla Ashlam, Atta Badii, Frederic Stahl

Abstract:

In recent years, SQL injection attacks have been identified as being prevalent against web applications. They affect network security and user data, which leads to a considerable loss of money and data every year. This paper presents the use of classification algorithms in machine learning using a method to classify the login data filtering inputs into "SQLi" or "Non-SQLi,” thus increasing the reliability and accuracy of results in terms of deciding whether an operation is an attack or a valid operation. A method as a Web-App is developed for auto-generated data replication to provide a twin of the targeted data structure. Shielding against SQLi attacks (WebAppShield) that verifies all users and prevents attackers (SQLi attacks) from entering and or accessing the database, which the machine learning module predicts as "Non-SQLi", has been developed. A special login form has been developed with a special instance of the data validation; this verification process secures the web application from its early stages. The system has been tested and validated, and up to 99% of SQLi attacks have been prevented.

Keywords: SQL injection, attacks, web application, accuracy, database, WebAppShield.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 425
676 Topology Optimization of Structures with Web-Openings

Authors: D. K. Lee, S. M. Shin, J. H. Lee

Abstract:

Topology optimization technique utilizes constant element densities as design parameters. Finally, optimal distribution contours of the material densities between voids (0) and solids (1) in design domain represent the determination of topology. It means that regions with element density values become occupied by solids in design domain, while there are only void phases in regions where no density values exist. Therefore the void regions of topology optimization results provide design information to decide appropriate depositions of web-opening in structure. Contrary to the basic objective of the topology optimization technique which is to obtain optimal topology of structures, this present study proposes a new idea that topology optimization results can be also utilized for decision of proper web-opening’s position. Numerical examples of linear elastostatic structures demonstrate efficiency of methodological design processes using topology optimization in order to determinate the proper deposition of web-openings.

Keywords: Topology optimization, web-opening, structure, element density, material.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1416
675 A Simplified Single Correlator Rake Receiver for CDMA Communications

Authors: K. Murali Krishna, Abhijit Mitra, C. Ardil

Abstract:

This paper presents a single correlator RAKE receiver for direct sequence code division multiple access (DS-CDMA) systems. In conventional RAKE receivers, multiple correlators are used to despread the multipath signals and then to align and combine those signals in a later stage before making a bit decision. The simplified receiver structure presented here uses a single correlator and single code sequence generator to recover the multipaths. Modified Walsh- Hadamard codes are used here for data spreading that provides better uncorrelation properties for the multipath signals. The main advantage of this receiver structure is that it requires only a single correlator and a code generator in contrary to the conventional RAKE receiver concept with multiple correlators. It is shown in results that the proposed receiver achieves better bit error rates in comparison with the conventional one for more than one multipaths.

Keywords: RAKE receiver, Code division multiple access, ModifiedWalsh-Hadamard codes, Single correlator.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3636
674 Financial Portfolio Optimization in Electricity Markets: Evaluation via Sharpe Ratio

Authors: F. Gökgöz, M. E. Atmaca

Abstract:

Electricity plays an indispensable role in human life and the economy. It is a unique product or service that must be balanced instantaneously, as electricity is not stored, generation and consumption should be proportional. Effective and efficient use of electricity is very important not only for society, but also for the environment. A competitive electricity market is one of the best ways to provide a suitable platform for effective and efficient use of electricity. On the other hand, it carries some risks that should be carefully managed by the market players. Risk management is an essential part in market players’ decision making. In this paper, risk management through diversification is applied with the help of Markowitz’s Mean-variance, Down-side and Semi-variance methods for a case study. Performance of optimal electricity sale solutions are measured and evaluated via Sharpe-Ratio, and the optimal portfolio solutions are improved. Two years of historical weekdays’ price data of the Turkish Day Ahead Market are used to demonstrate the approach.

Keywords: Electricity market, portfolio optimization, risk management in electricity market, Sharpe ratio.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1473
673 Multi-Sensor Target Tracking Using Ensemble Learning

Authors: Bhekisipho Twala, Mantepu Masetshaba, Ramapulana Nkoana

Abstract:

Multiple classifier systems combine several individual classifiers to deliver a final classification decision. However, an increasingly controversial question is whether such systems can outperform the single best classifier, and if so, what form of multiple classifiers system yields the most significant benefit. Also, multi-target tracking detection using multiple sensors is an important research field in mobile techniques and military applications. In this paper, several multiple classifiers systems are evaluated in terms of their ability to predict a system’s failure or success for multi-sensor target tracking tasks. The Bristol Eden project dataset is utilised for this task. Experimental and simulation results show that the human activity identification system can fulfil requirements of target tracking due to improved sensors classification performances with multiple classifier systems constructed using boosting achieving higher accuracy rates.

Keywords: Single classifier, machine learning, ensemble learning, multi-sensor target tracking.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 583
672 An Auxiliary Technique for Coronary Heart Disease Prediction by Analyzing ECG Based on ResNet and Bi-LSTM

Authors: Yang Zhang, Jian He

Abstract:

Heart disease is one of the leading causes of death in the world, and coronary heart disease (CHD) is one of the major heart diseases. Electrocardiogram (ECG) is widely used in the detection of heart diseases, but the traditional manual method for CHD prediction by analyzing ECG requires lots of professional knowledge for doctors. This paper presents sliding window and continuous wavelet transform (CWT) to transform ECG signals into images, and then ResNet and Bi-LSTM are introduced to build the ECG feature extraction network (namely ECGNet). At last, an auxiliary system for CHD prediction was developed based on modified ResNet18 and Bi-LSTM, and the public ECG dataset of CHD from MIMIC-3 was used to train and test the system. The experimental results show that the accuracy of the method is 83%, and the F1-score is 83%. Compared with the available methods for CHD prediction based on ECG, such as kNN, decision tree, VGGNet, etc., this method not only improves the prediction accuracy but also could avoid the degradation phenomenon of the deep learning network.

Keywords: Bi-LSTM, CHD, coronary heart disease, ECG, electrocardiogram, ResNet, sliding window.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 290
671 Design and Implementation of an AI-Enabled Task Assistance and Management System

Authors: Arun Prasad Jaganathan

Abstract:

In today's dynamic industrial world, traditional task allocation methods often fall short in adapting to evolving operational conditions. This paper presents an AI-enabled task assistance and management system designed to overcome the limitations of conventional approaches. By using artificial intelligence (AI) and machine learning (ML), the system intelligently interprets user instructions, analyzes tasks, and allocates resources based on real-time data and environmental factors. Additionally, geolocation tracking enables proactive identification of potential delays, ensuring timely interventions. With its transparent reporting mechanisms, the system provides stakeholders with clear insights into task progress, fostering accountability and informed decision-making. The paper presents a comprehensive overview of the system architecture, algorithm, and implementation, highlighting its potential to revolutionize task management across diverse industries.

Keywords: Artificial intelligence, machine learning, task allocation, operational efficiency, resource optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 34
670 CompleX-Machine: An Automated Testing Tool Using X-Machine Theory

Authors: E. K. A. Ogunshile

Abstract:

This paper is aimed at creating an Automatic Java X-Machine testing tool for software development. The nature of software development is changing; thus, the type of software testing tools required is also changing. Software is growing increasingly complex and, in part due to commercial impetus for faster software releases with new features and value, increasingly in danger of containing faults. These faults can incur huge cost for software development organisations and users; Cambridge Judge Business School’s research estimated the cost of software bugs to the global economy is $312 billion. Beyond the cost, faster software development methodologies and increasing expectations on developers to become testers is driving demand for faster, automated, and effective tools to prevent potential faults as early as possible in the software development lifecycle. Using X-Machine theory, this paper will explore a new tool to address software complexity, changing expectations on developers, faster development pressures and methodologies, with a view to reducing the huge cost of fixing software bugs.

Keywords: Conformance testing, finite state machine, software testing, X-Machine.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1226
669 Machine Learning Approach for Identifying Dementia from MRI Images

Authors: S. K. Aruna, S. Chitra

Abstract:

This research paper presents a framework for classifying Magnetic Resonance Imaging (MRI) images for Dementia. Dementia, an age-related cognitive decline is indicated by degeneration of cortical and sub-cortical structures. Characterizing morphological changes helps understand disease development and contributes to early prediction and prevention of the disease. Modelling, that captures the brain’s structural variability and which is valid in disease classification and interpretation is very challenging. Features are extracted using Gabor filter with 0, 30, 60, 90 orientations and Gray Level Co-occurrence Matrix (GLCM). It is proposed to normalize and fuse the features. Independent Component Analysis (ICA) selects features. Support Vector Machine (SVM) classifier with different kernels is evaluated, for efficiency to classify dementia. This study evaluates the presented framework using MRI images from OASIS dataset for identifying dementia. Results showed that the proposed feature fusion classifier achieves higher classification accuracy.

Keywords: Magnetic resonance imaging, dementia, Gabor filter, gray level co-occurrence matrix, support vector machine.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2105
668 Hybridized Technique to Analyze Workstress Related Data via the StressCafé

Authors: Anusua Ghosh, Andrew Nafalski, Jeffery Tweedale, Maureen Dollard

Abstract:

This paper presents anapproach of hybridizing two or more artificial intelligence (AI) techniques which arebeing used to fuzzify the workstress level ranking and categorize the rating accordingly. The use of two or more techniques (hybrid approach) has been considered in this case, as combining different techniques may lead to neutralizing each other-s weaknesses generating a superior hybrid solution. Recent researches have shown that there is a need for a more valid and reliable tools, for assessing work stress. Thus artificial intelligence techniques have been applied in this instance to provide a solution to a psychological application. An overview about the novel and autonomous interactive model for analysing work-stress that has been developedusing multi-agent systems is also presented in this paper. The establishment of the intelligent multi-agent decision analyser (IMADA) using hybridized technique of neural networks and fuzzy logic within the multi-agent based framework is also described.

Keywords: Fuzzy logic, intelligent agent, multi-agent systems, neural network, workplace stress.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3957
667 Intelligent Multi-Agent Middleware for Ubiquitous Home Networking Environments

Authors: Minwoo Son, Seung-Hun Lee, Dongkyoo Shin, Dongil Shin

Abstract:

The next stage of the home networking environment is supposed to be ubiquitous, where each piece of material is equipped with an RFID (Radio Frequency Identification) tag. To fully support the ubiquitous environment, home networking middleware should be able to recommend home services based on a user-s interests and efficiently manage information on service usage profiles for the users. Therefore, USN (Ubiquitous Sensor Network) technology, which recognizes and manages a appliance-s state-information (location, capabilities, and so on) by connecting RFID tags is considered. The Intelligent Multi-Agent Middleware (IMAM) architecture was proposed to intelligently manage the mobile RFID-based home networking and to automatically supply information about home services that match a user-s interests. Evaluation results for personalization services for IMAM using Bayesian-Net and Decision Trees are presented.

Keywords: Intelligent Agents, Home Network, Mobile RFID, Intelligent Middleware.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1435
666 An Extended Model for Sustainable Food and Nutrition Security in the Agrifood Sector

Authors: Ioannis Manikas

Abstract:

The increased consumer demand for environmentally friendly production and distribution practices and the stricter environmental regulations turned environmental aspects into important criteria in business decision-making. On the other hand, Food and Nutrition Security (FNS) has evolved dramatically during the last decades in theory and practice serving as a reference point for exchanging experiences among all agents involved in programs and projects to fostering policy and strategy development. Global pressures make it more important than ever to gain a better understanding of the contribution that agrifood businesses make to FNS and to examine ways to make them more resilient in an increasingly globalized and uncertain world. This study extends the standard three-dimensional model of sustainability to include two more dimensions: A technological dimension and a policy/political dimension. Apart from the economic, environmental and social dimensions regularly used in sustainability literature, the extended model will accurately represent the measures and policies addressing food and nutrition security.

Keywords: Food and nutrition security, sustainability, food safety, resilience.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1464
665 Holomorphic Prioritization of Sets within Decagram of Strategic Decision Making of POSM Using Operational Research (OR): Analytic Hierarchy Process (AHP) Analysis

Authors: Elias O. Tembe, Hussain A. Al-Salamin

Abstract:

There is decagram of strategic decisions of operations and production/service management (POSM) within operational research (OR) which must collate, namely: design, inventory, quality, location, process and capacity, layout, scheduling, maintain ace, and supply chain. This paper presents an architectural configuration conceptual framework of a decagram of sets decisions in a form of mathematical complete graph and abelian graph. Mathematically, a complete graph is undirected (UDG), and directed (DG) a relationship where every pair of vertices is connected, collated, confluent, and holomorphic. There has not been any study conducted which, however, prioritizes the holomorphic sets which of POMS within OR field of study. The study utilizes OR structured technique known as The Analytic Hierarchy Process (AHP) analysis for organizing, sorting and prioritizing(ranking) the sets within the decagram of POMS according to their attribution (propensity), and provides an analysis how the prioritization has real-world application within the 21st century.

Keywords: AHP analysis, Decagram, Decagon, Holomorphic.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1995
664 Socio-Economic Determinants of Physical Activity of Non-Manual Workers, Including the Early Senior Group, from the City of Wroclaw in Poland

Authors: Daniel Puciato, Piotr Oleśniewicz, Julita Markiewicz-Patkowska, Krzysztof Widawski, Michał Rozpara, Władysław Mynarski, Agnieszka Gawlik, Małgorzata Dębska, Soňa Jandová

Abstract:

Physical activity as a part of people’s everyday life reduces the risk of many diseases, including those induced by lifestyle, e.g. obesity, type 2 diabetes, osteoporosis, coronary heart disease, degenerative arthritis, and certain types of cancer. That refers particularly to professionally active people, including the early senior group working on non-manual positions. The aim of the study is to evaluate the relationship between physical activity and the socio-economic status of non-manual workers from Wroclaw—one of the biggest cities in Poland, a model setting for such investigations in this part of Europe. The crucial problem in the research is to find out the percentage of respondents who meet the health-related recommendations of the World Health Organization (WHO) concerning the volume, frequency, and intensity of physical activity, as well as to establish if the most important socio-economic factors, such as gender, age, education, marital status, per capita income, savings and debt, determine the compliance with the WHO physical activity recommendations. During the research, conducted in 2013, 1,170 people (611 women and 559 men) aged 21–60 years were examined. A diagnostic poll method was applied to collect the data. Physical activity was measured with the use of the short form of the International Physical Activity Questionnaire with extended socio-demographic questions, i.e. concerning gender, age, education, marital status, income, savings or debts. To evaluate the relationship between physical activity and selected socio-economic factors, logistic regression was used (odds ratio statistics). Statistical inference was conducted on the adopted ex ante probability level of p<0.05. The majority of respondents met the volume of physical effort recommended for health benefits. It was particularly noticeable in the case of the examined men. The probability of compliance with the WHO physical activity recommendations was highest for workers aged 21–30 years with secondary or higher education who were single, received highest incomes and had savings. The results indicate the relations between physical activity and socio-economic status in the examined women and men. People with lower socio-economic status (e.g. manual workers) are physically active primarily at work, whereas those better educated and wealthier implement physical effort primarily in their leisure time. Among the investigated subjects, the youngest group of non-manual workers have the best chances to meet the WHO standards of physical activity. The study also confirms that secondary education has a positive effect on the public awareness on the role of physical activity in human life. In general, the analysis of the research indicates that there is a relationship between physical activity and some socio-economic factors of the respondents, such as gender, age, education, marital status, income per capita, and the possession of savings. Although the obtained results cannot be applied for the general population, they show some important trends that will be verified in subsequent studies conducted by the authors of the paper.

Keywords: International physical activity questionnaire, non-manual workers, physical activity, socio-economic factors, WHO.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1244
663 Economic Optimization of Shell and Tube Heat Exchanger Using Nanofluid

Authors: Hassan Hajabdollahi

Abstract:

Economic optimization of shell and tube heat exchanger (STHE) is presented in this paper. To increase the rate of heat transfer, copper oxide (CuO) nanoparticle is added into the tube side fluid and their optimum results are compared with the case of without additive nanoparticle. Total annual cost (TAC) is selected as fitness function and nine decision variables related to the heat exchanger parameters as well as concentration of nanoparticle are considered. Optimization results reveal the noticeable improvement in the TAC and in the case of heat exchanger working with nanofluid compared with the case of base fluid (8.9%). Comparison of the results between two studied cases also reveal that the lower tube diameter, tube number, and baffle spacing are needed in the case of heat exchanger working with nanofluid compared with the case of base fluid.

Keywords: Shell and tube heat exchanger, nanoparticles additive, total annual cost, particle volumetric concentration.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1110
662 Cross Project Software Fault Prediction at Design Phase

Authors: Pradeep Singh, Shrish Verma

Abstract:

Software fault prediction models are created by using the source code, processed metrics from the same or previous version of code and related fault data. Some company do not store and keep track of all artifacts which are required for software fault prediction. To construct fault prediction model for such company, the training data from the other projects can be one potential solution. Earlier we predicted the fault the less cost it requires to correct. The training data consists of metrics data and related fault data at function/module level. This paper investigates fault predictions at early stage using the cross-project data focusing on the design metrics. In this study, empirical analysis is carried out to validate design metrics for cross project fault prediction. The machine learning techniques used for evaluation is Naïve Bayes. The design phase metrics of other projects can be used as initial guideline for the projects where no previous fault data is available. We analyze seven datasets from NASA Metrics Data Program which offer design as well as code metrics. Overall, the results of cross project is comparable to the within company data learning.

Keywords: Software Metrics, Fault prediction, Cross project, Within project.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2534
661 Using Different Aspects of the Signings for Appearance-based Sign Language Recognition

Authors: Morteza Zahedi, Philippe Dreuw, Thomas Deselaers, Hermann Ney

Abstract:

Sign language is used by the deaf and hard of hearing people for communication. Automatic sign language recognition is a challenging research area since sign language often is the only way of communication for the deaf people. Sign language includes different components of visual actions made by the signer using the hands, the face, and the torso, to convey his/her meaning. To use different aspects of signs, we combine the different groups of features which have been extracted from the image frames recorded directly by a stationary camera. We combine the features in two levels by employing three techniques. At the feature level, an early feature combination can be performed by concatenating and weighting different feature groups, or by concatenating feature groups over time and using LDA to choose the most discriminant elements. At the model level, a late fusion of differently trained models can be carried out by a log-linear model combination. In this paper, we investigate these three combination techniques in an automatic sign language recognition system and show that the recognition rate can be significantly improved.

Keywords: American sign language, appearance-based features, Feature combination, Sign language recognition

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1388
660 Financial Portfolio Optimization in Turkish Electricity Market via Value at Risk

Authors: F. Gökgöz, M. E. Atmaca

Abstract:

Electricity has an indispensable role in human daily life, technological development and economy. It is a special product or service that should be instantaneously generated and consumed. Sources of the world are limited so that effective and efficient use of them is very important not only for human life and environment but also for technological and economic development. Competitive electricity market is one of the important way that provides suitable platform for effective and efficient use of electricity. Besides benefits, it brings along some risks that should be carefully managed by a market player like Electricity Generation Company. Risk management is an essential part in market players’ decision making. In this paper, risk management through diversification is applied with the help of Value at Risk methods for case studies. Performance of optimal electricity sale solutions are measured and the portfolio performance has been evaluated via Sharpe-Ratio, and compared with conventional approach. Biennial historical electricity price data of Turkish Day Ahead Market are used to demonstrate the approach.

Keywords: Electricity market, portfolio optimization, risk management, Sharpe ratio, value at risk.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1045
659 Segmentation of Lungs from CT Scan Images for Early Diagnosis of Lung Cancer

Authors: Nisar Ahmed Memon, Anwar Majid Mirza, S.A.M. Gilani

Abstract:

Segmentation is an important step in medical image analysis and classification for radiological evaluation or computer aided diagnosis. The CAD (Computer Aided Diagnosis ) of lung CT generally first segment the area of interest (lung) and then analyze the separately obtained area for nodule detection in order to diagnosis the disease. For normal lung, segmentation can be performed by making use of excellent contrast between air and surrounding tissues. However this approach fails when lung is affected by high density pathology. Dense pathologies are present in approximately a fifth of clinical scans, and for computer analysis such as detection and quantification of abnormal areas it is vital that the entire and perfectly lung part of the image is provided and no part, as present in the original image be eradicated. In this paper we have proposed a lung segmentation technique which accurately segment the lung parenchyma from lung CT Scan images. The algorithm was tested against the 25 datasets of different patients received from Ackron Univeristy, USA and AGA Khan Medical University, Karachi, Pakistan.

Keywords: Computer Aided Diagnosis, Medical ImageProcessing, Region Growing, Segmentation, Thresholding,

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2584
658 Effect of Rotation Rate on Chemical Segragation during Phase Change

Authors: Nouri Sabrina, Benzeghiba Mohamed, Ghezal Abderrahmane

Abstract:

Numerical parametric study is conducted to study the effects of ampoule rotation on the flows and the dopant segregation in vertical bridgman (vb) crystal growth. Calculations were performed in unsteady state. The extended darcy model, which includes the time derivative and coriolis terms, has been employed in the momentum equation. It’s found that the convection, and dopant segregation can be affected significantly by ampoule rotation, and the effect is similar to that by an axial magnetic field. Ampoule rotation decreases the intensity of convection and stretches the flow cell axially. When the convection is weak, the flow can be suppressed almost completely by moderate ampoule rotation and the dopant segregation becomes diffusion-controlled. For stronger convection, the elongated flow cell by ampoule rotation may bring dopant mixing into the bulk melt reducing axial segregation at the early stage of the growth. However, if the cellular flow cannot be suppressed completely, ampoule rotation may induce larger radial segregation due to poor mixing.

Keywords: Numerical Simulation, Heat and mass transfer, vertical solidification, chemical segregation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1745
657 Performance Evaluation of Faculties of Islamic Azad University of Zahedan Branch Based-On Two-Component DEA

Authors: Ali Payan

Abstract:

The aim of this paper is to evaluate the performance of the faculties of Islamic Azad University of Zahedan Branch based on two-component (teaching and research) decision making units (DMUs) in data envelopment analysis (DEA). Nowadays it is obvious that most of the systems as DMUs do not act as a simple inputoutput structure. Instead, if they have been studied more delicately, they include network structure. University is such a network in which different sections i.e. teaching, research, students and office work as a parallel structure. They consume some inputs of university commonly and some others individually. Then, they produce both dependent and independent outputs. These DMUs are called two-component DMUs with network structure. In this paper, performance of the faculties of Zahedan branch is calculated by using relative efficiency model and also, a formula to compute relative efficiencies teaching and research components based on DEA are offered.

Keywords: Data envelopment analysis, faculties of Islamic Azad University of Zahedan branch, two-component DMUs.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1651
656 Probabilistic Characteristics of older PR Frames in the Mid-America Earthquake Region

Authors: Do-Hwan Kim, Roberto Leon

Abstract:

Probabilistic characteristics of seismic responses of the Partially Restrained connection rotation (PRCR) and panel zone deformation (PZD) installed in older steel moment frames were investigated in accordance with statistical inference in decision-making process. The 4, 6 and 8 story older steel moment frames with clip angle and T-stub connections were designed and analyzed using 2%/50yrs ground motions in four cities of the Mid-America earthquake region. The probability density function and cumulative distribution function of PRCR and PZD were determined by the goodness-of-fit tests based on probabilistic parameters measured from the results of the nonlinear time-history analyses. The obtained probabilistic parameters and distributions can be used to find out what performance level mainly PR connections and panel zones satisfy and how many PR connections and panel zones experience a serious damage under the Mid-America ground motions.

Keywords: Mid-America earthquake, Panel zone, PR connection, Probabilistic characteristics, seismic performance

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1402
655 Roadmapping as a Collaborative Strategic Decision-Making Process: Shaping Social Dialogue Options for the European Banking Sector

Authors: Christos A. Ioannou, Panagiotis Panagiotopoulos, Lampros Stergioulas

Abstract:

The new status generated by technological advancements and changes in the global economy raises important issues on how communities and organisations need to innovate upon their traditional processes in order to adapt to the challenges of the Knowledge Society. The DialogoS+ European project aims to study the role of and promote social dialogue in the banking sector, strengthen the link between old and new members and make social dialogue at the European level a force for innovation and change, also given the context of the international crisis emerging in 2008- 2009. Under the scope of DialogoS+, this paper describes how the community of Europe-s banking sector trade unions attempted to adapt to the challenges of the Knowledge Society by exploiting the benefits of new channels of communication, learning, knowledge generation and diffusion focusing on the concept of roadmapping. Important dimensions of social dialogue such as collective bargaining and working conditions are addressed.

Keywords: Banking sector, knowledge society, road mapping, social dialogue.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2122
654 A Quality Optimization Approach: An Application on Next Generation Networks

Authors: Gülfem I. Alptekin, S. Emre Alptekin

Abstract:

The next generation wireless systems, especially the cognitive radio networks aim at utilizing network resources more efficiently. They share a wide range of available spectrum in an opportunistic manner. In this paper, we propose a quality management model for short-term sub-lease of unutilized spectrum bands to different service providers. We built our model on competitive secondary market architecture. To establish the necessary conditions for convergent behavior, we utilize techniques from game theory. Our proposed model is based on potential game approach that is suitable for systems with dynamic decision making. The Nash equilibrium point tells the spectrum holders the ideal price values where profit is maximized at the highest level of customer satisfaction. Our numerical results show that the price decisions of the network providers depend on the price and QoS of their own bands as well as the prices and QoS levels of their opponents- bands.

Keywords: cognitive radio networks, game theory, nextgeneration wireless networks, spectrum management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1504
653 Low Power and Less Area Architecture for Integer Motion Estimation

Authors: C Hisham, K Komal, Amit K Mishra

Abstract:

Full search block matching algorithm is widely used for hardware implementation of motion estimators in video compression algorithms. In this paper we are proposing a new architecture, which consists of a 2D parallel processing unit and a 1D unit both working in parallel. The proposed architecture reduces both data access power and computational power which are the main causes of power consumption in integer motion estimation. It also completes the operations with nearly the same number of clock cycles as compared to a 2D systolic array architecture. In this work sum of absolute difference (SAD)-the most repeated operation in block matching, is calculated in two steps. The first step is to calculate the SAD for alternate rows by a 2D parallel unit. If the SAD calculated by the parallel unit is less than the stored minimum SAD, the SAD of the remaining rows is calculated by the 1D unit. Early termination, which stops avoidable computations has been achieved with the help of alternate rows method proposed in this paper and by finding a low initial SAD value based on motion vector prediction. Data reuse has been applied to the reference blocks in the same search area which significantly reduced the memory access.

Keywords: Sum of absolute difference, high speed DSP.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1483
652 Inadequate Requirements Engineering Process: A Key Factor for Poor Software Development in Developing Nations: A Case Study

Authors: K. Adu Michael, K. Alese Boniface

Abstract:

Developing a reliable and sustainable software products is today a big challenge among up–coming software developers in Nigeria. The inability to develop a comprehensive problem statement needed to execute proper requirements engineering process is missing. The need to describe the ‘what’ of a system in one document, written in a natural language is a major step in the overall process of Software Engineering. Requirements Engineering is a process use to discover, analyze and validate system requirements. This process is needed in reducing software errors at the early stage of the development of software. The importance of each of the steps in Requirements Engineering is clearly explained in the context of using detailed problem statement from client/customer to get an overview of an existing system along with expectations from the new system. This paper elicits inadequate Requirements Engineering principle as the major cause of poor software development in developing nations using a case study of final year computer science students of a tertiary-education institution in Nigeria.

Keywords: Client/Customer, Problem Statement, Requirements Engineering, Software Developers.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2448
651 Development of Fake News Model Using Machine Learning through Natural Language Processing

Authors: Sajjad Ahmed, Knut Hinkelmann, Flavio Corradini

Abstract:

Fake news detection research is still in the early stage as this is a relatively new phenomenon in the interest raised by society. Machine learning helps to solve complex problems and to build AI systems nowadays and especially in those cases where we have tacit knowledge or the knowledge that is not known. We used machine learning algorithms and for identification of fake news; we applied three classifiers; Passive Aggressive, Naïve Bayes, and Support Vector Machine. Simple classification is not completely correct in fake news detection because classification methods are not specialized for fake news. With the integration of machine learning and text-based processing, we can detect fake news and build classifiers that can classify the news data. Text classification mainly focuses on extracting various features of text and after that incorporating those features into classification. The big challenge in this area is the lack of an efficient way to differentiate between fake and non-fake due to the unavailability of corpora. We applied three different machine learning classifiers on two publicly available datasets. Experimental analysis based on the existing dataset indicates a very encouraging and improved performance.

Keywords: Fake news detection, types of fake news, machine learning, natural language processing, classification techniques.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1495
650 The Effect of Drying Conditions on the Presence of Volatile Compounds in Cranberries

Authors: Karina Ruse, Martins Sabovics, Tatjana Rakcejeva, Lija Dukalska, Ruta Galoburda, Laima Berzina

Abstract:

the research was accomplished on fresh in Latvia wild growing cranberries and cranberry cultivars. The aim of the study was to evaluate effect of pretreatment method and drying conditions on the volatile compounds composition in cranberries. Berries pre-treatment methods were: perforation, halving and steam-blanching. The berries before drying in a cabinet drier were pre-treated using all three methods, in microwave vacuum drier – using a steam-blanching and halving. Volatile compounds in cranberries were analysed using GC-MS of extracts obtained by SPME. During present research 21 various volatile compounds were detected in fresh cranberries: the cultivar 'Steven' - 15, 'Bergman' and 'Early black' – 13, 'Ben Lear' and 'Pilgrim' – 11 and wild cranberries – 14 volatile compounds. In dried cranberries 20 volatile compounds were detected. Mathematical data processing allows drawing a conclusion that there exists the significant influence of cranberry cultivar, pre-treatment method and drying condition on volatile compounds in berries and new volatile compound formation.

Keywords: volatile compounds, cranberries, convective drier, microwave-vacuum drier

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2580
649 Efficient Detection Using Sequential Probability Ratio Test in Mobile Cognitive Radio Systems

Authors: Yeon-Jea Cho, Sang-Uk Park, Won-Chul Choi, Dong-Jo Park

Abstract:

This paper proposes a smart design strategy for a sequential detector to reliably detect the primary user-s signal, especially in fast fading environments. We study the computation of the log-likelihood ratio for coping with a fast changing received signal and noise sample variances, which are considered random variables. First, we analyze the detectability of the conventional generalized log-likelihood ratio (GLLR) scheme when considering fast changing statistics of unknown parameters caused by fast fading effects. Secondly, we propose an efficient sensing algorithm for performing the sequential probability ratio test in a robust and efficient manner when the channel statistics are unknown. Finally, the proposed scheme is compared to the conventional method with simulation results with respect to the average number of samples required to reach a detection decision.

Keywords: Cognitive radio, fast fading, sequential detection, spectrum sensing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1736