Search results for: fuzzy multi-objective combinatorial programming problem
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8278

Search results for: fuzzy multi-objective combinatorial programming problem

7078 Learning-Oriented School Education: Indicator Construction and Taiwan's Implementation Performance

Authors: Meiju Chen, Chaoyu Guo, Chia Wei Tang

Abstract:

The present study's purpose is twofold: first, to construct indicators for learning-oriented school education and, second, to conduct a survey to examine how learning-oriented education has been implemented in junior high schools after the launch of the 12-year compulsory curriculum. For indicator system construction, we compiled relevant literature to develop a preliminary indicator list model and then conducted two rounds of a questionnaire survey to gain comprehensive feedback from experts to finalize our indicator model. In the survey's first round, 12 experts were invited to evaluate the indicators' appropriateness. Based on the experts' consensus, we determined our final indicator list and used it to develop the Fuzzy Delphi questionnaire to finalize the indicator system and each indicator's relative value. For the fact-finding survey, we collected 454 valid samples to examine how the concept of learning-oriented education is adopted and implemented in the junior high school context. We also used this data in our importance-performance analysis to explore the strengths and weaknesses of school education in Taiwan. The results suggest that the indicator system for learning-oriented school education must consist of seven dimensions and 34 indicators. Among the seven dimensions, 'student learning' and 'curriculum planning and implementation' are the most important yet underperforming dimensions that need immediate improvement. We anticipate that the indicator system will be a useful tool for other countries' evaluation of schools' performance in learning-oriented education.

Keywords: learning-oriented education, school education, fuzzy Delphi method, importance-performance analysis

Procedia PDF Downloads 128
7077 Evaluation of the Socio-Economic Impact of Marine Debris in Coastal Nigeria

Authors: Chibuzo Okoye Daniels, Gillian Glegg, Lynda Rodwell

Abstract:

Marine debris from fishing nets to medical equipment to food packaging that play major roles in boosting the economy and protecting human health is now more than an environmental problem that can be solved by legislation, law enforcement and technical solutions. It has also been identified as a cultural problem that can only be addressed by identifying instruments that can be used to change human attitudes and behaviors. This may be through management approaches, education and involvement of all sectors/interests, including the public. To contribute to the sustainable development of coastal Nigeria, two case study areas (Ikoyi and Victoria Islands of Lagos State) were used to evaluate the socio-economic impacts of marine debris problem in coastal Nigeria. The following methods were used: (1) semi-structured interviews with key stakeholders and businesses on beaches, waterfronts and waterways within the study areas and (2) observational study of beaches, waterfronts and waterways within the study areas. The results of the study have shown that marine debris is a cultural and multi-sectoral problem that poses great threat not only to the environmental sustainability of the study areas but also to the wellbeing of its citizens and the economy of coastal Nigeria. Current solid waste and marine debris management practices are inefficient due to inadequate knowledge of how to tackle the problem. To ensure environmental sustainability in coastal Nigeria and avoid waste of scarce financial resources, adequate, appropriate and cost effective solutions to the marine debris problem need to be identified and effectively transferred for implementation in the study areas.

Keywords: sustainability, coastal Nigeria, study areas, aquaculture

Procedia PDF Downloads 538
7076 Analysis of NFC and Biometrics in the Retail Industry

Authors: Ziwei Xu

Abstract:

The increasing emphasis on mobility has driven the application of innovative communication technologies across various industries. In the retail sector, Near Field Communication (NFC) has emerged as a significant and transformative technology, particularly in the payment and retail supermarket sectors. NFC enables new payment methods, such as electronic wallets, and enhances information management in supermarkets, contributing to the growth of the trade. This report presents a comprehensive analysis of NFC technology, focusing on five key aspects. Firstly, it provides an overview of NFC, including its application methods and development history. Additionally, it incorporates Arthur's work on combinatorial evolution to elucidate the emergence and impact of NFC technology, while acknowledging the limitations of the model in analyzing NFC. The report then summarizes the positive influence of NFC on the retail industry along with its associated constraints. Furthermore, it explores the adoption of NFC from both organizational and individual perspectives, employing the Best Predictors of organizational IT adoption and UTAUT2 models, respectively. Finally, the report discusses the potential future replacement of NFC with biometrics technology, highlighting its advantages over NFC and leveraging Arthur's model to investigate its future development prospects.

Keywords: innovation, NFC, industry, biometrics

Procedia PDF Downloads 52
7075 Development of Automatic Laser Scanning Measurement Instrument

Authors: Chien-Hung Liu, Yu-Fen Chen

Abstract:

This study used triangular laser probe and three-axial direction mobile platform for surface measurement, programmed it and applied it to real-time analytic statistics of different measured data. This structure was used to design a system integration program: using triangular laser probe for scattering or reflection non-contact measurement, transferring the captured signals to the computer through RS-232, and using RS-485 to control the three-axis platform for a wide range of measurement. The data captured by the laser probe are formed into a 3D surface. This study constructed an optical measurement application program in the concept of visual programming language. First, the signals are transmitted to the computer through RS-232/RS-485, and then the signals are stored and recorded in graphic interface timely. This programming concept analyzes various messages, and makes proper presentation graphs and data processing to provide the users with friendly graphic interfaces and data processing state monitoring, and identifies whether the present data are normal in graphic concept. The major functions of the measurement system developed by this study are thickness measurement, SPC, surface smoothness analysis, and analytical calculation of trend line. A result report can be made and printed promptly. This study measured different heights and surfaces successfully, performed on-line data analysis and processing effectively, and developed a man-machine interface for users to operate.

Keywords: laser probe, non-contact measurement, triangulation measurement principle, statistical process control, labVIEW

Procedia PDF Downloads 351
7074 Improving School Design through Diverse Stakeholder Participation in the Programming Phase

Authors: Doris C. C. K. Kowaltowski, Marcella S. Deliberador

Abstract:

The architectural design process, in general, is becoming more complex, as new technical, social, environmental, and economical requirements are imposed. For school buildings, this scenario is also valid. The quality of a school building depends on known design criteria and professional knowledge, as well as feedback from building performance assessments. To attain high-performance school buildings, a design process should add a multidisciplinary team, through an integrated process, to ensure that the various specialists contribute at an early stage to design solutions. The participation of stakeholders is of special importance at the programming phase when the search for the most appropriate design solutions is underway. The composition of a multidisciplinary team should comprise specialists in education, design professionals, and consultants in various fields such as environmental comfort and psychology, sustainability, safety and security, as well as administrators, public officials and neighbourhood representatives. Users, or potential users (teachers, parents, students, school officials, and staff), should be involved. User expectations must be guided, however, toward a proper understanding of a response of design to needs to avoid disappointment. In this context, appropriate tools should be introduced to organize such diverse participants and ensure a rich and focused response to needs and a productive outcome of programming sessions. In this paper, different stakeholder in a school design process are discussed in relation to their specific contributions and a tool in the form of a card game is described to structure the design debates and ensure a comprehensive decision-making process. The game is based on design patterns for school architecture as found in the literature and is adapted to a specific reality: State-run public schools in São Paulo, Brazil. In this State, school buildings are managed by a foundation called Fundação para o Desenvolvimento da Educação (FDE). FDE supervises new designs and is responsible for the maintenance of ~ 5000 schools. The design process of this context was characterised with a recommendation to improve the programming phase. Card games can create a common environment, to which all participants can relate and, therefore, can contribute to briefing debates on an equal footing. The cards of the game described here represent essential school design themes as found in the literature. The tool was tested with stakeholder groups and with architecture students. In both situations, the game proved to be an efficient tool to stimulate school design discussions and to aid in the elaboration of a rich, focused and thoughtful architectural program for a given demand. The game organizes the debates and all participants are shown to spontaneously contribute each in his own field of expertise to the decision-making process. Although the game was specifically based on a local school design process it shows potential for other contexts because the content is based on known facts, needs and concepts of school design, which are global. A structured briefing phase with diverse stakeholder participation can enrich the design process and consequently improve the quality of school buildings.

Keywords: architectural program, design process, school building design, stakeholder

Procedia PDF Downloads 390
7073 A Cognitive Approach to the Optimization of Power Distribution across an Educational Campus

Authors: Mrinmoy Majumder, Apu Kumar Saha

Abstract:

The ever-increasing human population and its demand for energy is placing stress upon conventional energy sources; and as demand for power continues to outstrip supply, the need to optimize energy distribution and utilization is emerging as an important focus for various stakeholders. The distribution of available energy must be achieved in such a way that the needs of the consumer are satisfied. However, if the availability of resources is not sufficient to satisfy consumer demand, it is necessary to find a method to select consumers based on factors such as their socio-economic or environmental impacts. Weighting consumer types in this way can help separate them based on their relative importance, and cognitive optimization of the allocation process can then be carried out so that, even on days of particularly scarce supply, the socio-economic impacts of not satisfying the needs of consumers can be minimized. In this context, the present study utilized fuzzy logic to assign weightage to different types of consumers based at an educational campus in India, and then established optimal allocation by applying the non-linear mapping capability of neuro-genetic algorithms. The outputs of the algorithms were compared with similar outputs from particle swarm optimization and differential evolution algorithms. The results of the study demonstrate an option for the optimal utilization of available energy based on the socio-economic importance of consumers.

Keywords: power allocation, optimization problem, neural networks, environmental and ecological engineering

Procedia PDF Downloads 457
7072 Impact of the Electricity Market Prices during the COVID-19 Pandemic on Energy Storage Operation

Authors: Marin Mandić, Elis Sutlović, Tonći Modrić, Luka Stanić

Abstract:

With the restructuring and deregulation of the power system, storage owners, generation companies or private producers can offer their multiple services on various power markets and earn income in different types of markets, such as the day-ahead, real-time, ancillary services market, etc. During the COVID-19 pandemic, electricity prices, as well as ancillary services prices, increased significantly. The optimization of the energy storage operation was performed using a suitable model for simulating the operation of a pumped storage hydropower plant under market conditions. The objective function maximizes the income earned through energy arbitration, regulation-up, regulation-down and spinning reserve services. The optimization technique used for solving the objective function is mixed integer linear programming (MILP). In numerical examples, the pumped storage hydropower plant operation has been optimized considering the already achieved hourly electricity market prices from Nord Pool for the pre-pandemic (2019) and the pandemic (2020 and 2021) years. The impact of the electricity market prices during the COVID-19 pandemic on energy storage operation is shown through the analysis of income, operating hours, reserved capacity and consumed energy for each service. The results indicate the role of energy storage during a significant fluctuation in electricity and services prices.

Keywords: electrical market prices, electricity market, energy storage optimization, mixed integer linear programming (MILP) optimization

Procedia PDF Downloads 156
7071 Parallel Fuzzy Rough Support Vector Machine for Data Classification in Cloud Environment

Authors: Arindam Chaudhuri

Abstract:

Classification of data has been actively used for most effective and efficient means of conveying knowledge and information to users. The prima face has always been upon techniques for extracting useful knowledge from data such that returns are maximized. With emergence of huge datasets the existing classification techniques often fail to produce desirable results. The challenge lies in analyzing and understanding characteristics of massive data sets by retrieving useful geometric and statistical patterns. We propose a supervised parallel fuzzy rough support vector machine (PFRSVM) for data classification in cloud environment. The classification is performed by PFRSVM using hyperbolic tangent kernel. The fuzzy rough set model takes care of sensitiveness of noisy samples and handles impreciseness in training samples bringing robustness to results. The membership function is function of center and radius of each class in feature space and is represented with kernel. It plays an important role towards sampling the decision surface. The success of PFRSVM is governed by choosing appropriate parameter values. The training samples are either linear or nonlinear separable. The different input points make unique contributions to decision surface. The algorithm is parallelized with a view to reduce training times. The system is built on support vector machine library using Hadoop implementation of MapReduce. The algorithm is tested on large data sets to check its feasibility and convergence. The performance of classifier is also assessed in terms of number of support vectors. The challenges encountered towards implementing big data classification in machine learning frameworks are also discussed. The experiments are done on the cloud environment available at University of Technology and Management, India. The results are illustrated for Gaussian RBF and Bayesian kernels. The effect of variability in prediction and generalization of PFRSVM is examined with respect to values of parameter C. It effectively resolves outliers’ effects, imbalance and overlapping class problems, normalizes to unseen data and relaxes dependency between features and labels. The average classification accuracy for PFRSVM is better than other classifiers for both Gaussian RBF and Bayesian kernels. The experimental results on both synthetic and real data sets clearly demonstrate the superiority of the proposed technique.

Keywords: FRSVM, Hadoop, MapReduce, PFRSVM

Procedia PDF Downloads 475
7070 Logistics and Supply Chain Management Using Smart Contracts on Blockchain

Authors: Armen Grigoryan, Milena Arakelyan

Abstract:

The idea of smart logistics is still quite a complicated one. It can be used to market products to a large number of customers or to acquire raw materials of the highest quality at the lowest cost in geographically dispersed areas. The use of smart contracts in logistics and supply chain management has the potential to revolutionize the way that goods are tracked, transported, and managed. Smart contracts are simply computer programs written in one of the blockchain programming languages (Solidity, Rust, Vyper), which are capable of self-execution once the predetermined conditions are met. They can be used to automate and streamline many of the traditional manual processes that are currently used in logistics and supply chain management, including the tracking and movement of goods, the management of inventory, and the facilitation of payments and settlements between different parties in the supply chain. Currently, logistics is a core area for companies which is concerned with transporting products between parties. Still, the problem of this sector is that its scale may lead to detainments and defaults in the delivery of goods, as well as other issues. Moreover, large distributors require a large number of workers to meet all the needs of their stores. All this may contribute to big detainments in order processing and increases the potentiality of losing orders. In an attempt to break this problem, companies have automated all their procedures, contributing to a significant augmentation in the number of businesses and distributors in the logistics sector. Hence, blockchain technology and smart contracted legal agreements seem to be suitable concepts to redesign and optimize collaborative business processes and supply chains. The main purpose of this paper is to examine the scope of blockchain technology and smart contracts in the field of logistics and supply chain management. This study discusses the research question of how and to which extent smart contracts and blockchain technology can facilitate and improve the implementation of collaborative business structures for sustainable entrepreneurial activities in smart supply chains. The intention is to provide a comprehensive overview of the existing research on the use of smart contracts in logistics and supply chain management and to identify any gaps or limitations in the current knowledge on this topic. This review aims to provide a summary and evaluation of the key findings and themes that emerge from the research, as well as to suggest potential directions for future research on the use of smart contracts in logistics and supply chain management.

Keywords: smart contracts, smart logistics, smart supply chain management, blockchain and smart contracts in logistics, smart contracts for controlling supply chain management

Procedia PDF Downloads 75
7069 A Survey of Discrete Facility Location Problems

Authors: Z. Ulukan, E. Demircioğlu,

Abstract:

Facility location is a complex real-world problem which needs a strategic management decision. This paper provides a general review on studies, efforts and developments in Facility Location Problems which are classical optimization problems having a wide-spread applications in various areas such as transportation, distribution, production, supply chain decisions and telecommunication. Our goal is not to review all variants of different studies in FLPs or to describe very detailed computational techniques and solution approaches, but rather to provide a broad overview of major location problems that have been studied, indicating how they are formulated and what are proposed by researchers to tackle the problem. A brief, elucidative table based on a grouping according to “General Problem Type” and “Methods Proposed” used in the studies is also presented at the end of the work.

Keywords: discrete location problems, exact methods, heuristic algorithms, single source capacitated facility location problems

Procedia PDF Downloads 446
7068 Semantic Indexing Improvement for Textual Documents: Contribution of Classification by Fuzzy Association Rules

Authors: Mohsen Maraoui

Abstract:

In the aim of natural language processing applications improvement, such as information retrieval, machine translation, lexical disambiguation, we focus on statistical approach to semantic indexing for multilingual text documents based on conceptual network formalism. We propose to use this formalism as an indexing language to represent the descriptive concepts and their weighting. These concepts represent the content of the document. Our contribution is based on two steps. In the first step, we propose the extraction of index terms using the multilingual lexical resource Euro WordNet (EWN). In the second step, we pass from the representation of index terms to the representation of index concepts through conceptual network formalism. This network is generated using the EWN resource and pass by a classification step based on association rules model (in attempt to discover the non-taxonomic relations or contextual relations between the concepts of a document). These relations are latent relations buried in the text and carried by the semantic context of the co-occurrence of concepts in the document. Our proposed indexing approach can be applied to text documents in various languages because it is based on a linguistic method adapted to the language through a multilingual thesaurus. Next, we apply the same statistical process regardless of the language in order to extract the significant concepts and their associated weights. We prove that the proposed indexing approach provides encouraging results.

Keywords: concept extraction, conceptual network formalism, fuzzy association rules, multilingual thesaurus, semantic indexing

Procedia PDF Downloads 125
7067 Mixed Integer Programming-Based One-Class Classification Method for Process Monitoring

Authors: Younghoon Kim, Seoung Bum Kim

Abstract:

One-class classification plays an important role in detecting outlier and abnormality from normal observations. In the previous research, several attempts were made to extend the scope of application of the one-class classification techniques to statistical process control problems. For most previous approaches, such as support vector data description (SVDD) control chart, the design of the control limits is commonly based on the assumption that the proportion of abnormal observations is approximately equal to an expected Type I error rate in Phase I process. Because of the limitation of the one-class classification techniques based on convex optimization, we cannot make the proportion of abnormal observations exactly equal to expected Type I error rate: controlling Type I error rate requires to optimize constraints with integer decision variables, but convex optimization cannot satisfy the requirement. This limitation would be undesirable in theoretical and practical perspective to construct effective control charts. In this work, to address the limitation of previous approaches, we propose the one-class classification algorithm based on the mixed integer programming technique, which can solve problems formulated with continuous and integer decision variables. The proposed method minimizes the radius of a spherically shaped boundary subject to the number of normal data to be equal to a constant value specified by users. By modifying this constant value, users can exactly control the proportion of normal data described by the spherically shaped boundary. Thus, the proportion of abnormal observations can be made theoretically equal to an expected Type I error rate in Phase I process. Moreover, analogous to SVDD, the boundary can be made to describe complex structures by using some kernel functions. New multivariate control chart applying the effectiveness of the algorithm is proposed. This chart uses a monitoring statistic to characterize the degree of being an abnormal point as obtained through the proposed one-class classification. The control limit of the proposed chart is established by the radius of the boundary. The usefulness of the proposed method was demonstrated through experiments with simulated and real process data from a thin film transistor-liquid crystal display.

Keywords: control chart, mixed integer programming, one-class classification, support vector data description

Procedia PDF Downloads 161
7066 Machine Learning Facing Behavioral Noise Problem in an Imbalanced Data Using One Side Behavioral Noise Reduction: Application to a Fraud Detection

Authors: Salma El Hajjami, Jamal Malki, Alain Bouju, Mohammed Berrada

Abstract:

With the expansion of machine learning and data mining in the context of Big Data analytics, the common problem that affects data is class imbalance. It refers to an imbalanced distribution of instances belonging to each class. This problem is present in many real world applications such as fraud detection, network intrusion detection, medical diagnostics, etc. In these cases, data instances labeled negatively are significantly more numerous than the instances labeled positively. When this difference is too large, the learning system may face difficulty when tackling this problem, since it is initially designed to work in relatively balanced class distribution scenarios. Another important problem, which usually accompanies these imbalanced data, is the overlapping instances between the two classes. It is commonly referred to as noise or overlapping data. In this article, we propose an approach called: One Side Behavioral Noise Reduction (OSBNR). This approach presents a way to deal with the problem of class imbalance in the presence of a high noise level. OSBNR is based on two steps. Firstly, a cluster analysis is applied to groups similar instances from the minority class into several behavior clusters. Secondly, we select and eliminate the instances of the majority class, considered as behavioral noise, which overlap with behavior clusters of the minority class. The results of experiments carried out on a representative public dataset confirm that the proposed approach is efficient for the treatment of class imbalances in the presence of noise.

Keywords: machine learning, imbalanced data, data mining, big data

Procedia PDF Downloads 114
7065 Some Codes for Variants in Graphs

Authors: Sofia Ait Bouazza

Abstract:

We consider the problem of finding a minimum identifying code in a graph. This problem was initially introduced in 1998 and has been since fundamentally connected to a wide range of applications (fault diagnosis, location detection …). Suppose we have a building into which we need to place fire alarms. Suppose each alarm is designed so that it can detect any fire that starts either in the room in which it is located or in any room that shares a doorway with the room. We want to detect any fire that may occur or use the alarms which are sounding to not only to not only detect any fire but be able to tell exactly where the fire is located in the building. For reasons of cost, we want to use as few alarms as necessary. The first problem involves finding a minimum domination set of a graph. If the alarms are three state alarms capable of distinguishing between a fire in the same room as the alarm and a fire in an adjacent room, we are trying to find a minimum locating domination set. If the alarms are two state alarms that can only sound if there is a fire somewhere nearby, we are looking for a differentiating domination set of a graph. These three areas are the subject of much active research; we primarily focus on the third problem. An identifying code of a graph G is a dominating set C such that every vertex x of G is distinguished from other vertices by the set of vertices in C that are at distance at most r≥1 from x. When only vertices out of the code are asked to be identified, we get the related concept of a locating dominating set. The problem of finding an identifying code (resp a locating dominating code) of minimum size is a NP-hard problem, even when the input graph belongs to a number of specific graph classes. Therefore, we study this problem in some restricted classes of undirected graphs like split graph, line graph and path in a directed graph. Then we present some results on the identifying code by giving an exact value of upper total locating domination and a total 2-identifying code in directed and undirected graph. Moreover we determine exact values of locating dominating code and edge identifying code of thin headless spider and locating dominating code of complete suns.

Keywords: identiying codes, locating dominating set, split graphs, thin headless spider

Procedia PDF Downloads 461
7064 Image Classification with Localization Using Convolutional Neural Networks

Authors: Bhuyain Mobarok Hossain

Abstract:

Image classification and localization research is currently an important strategy in the field of computer vision. The evolution and advancement of deep learning and convolutional neural networks (CNN) have greatly improved the capabilities of object detection and image-based classification. Target detection is important to research in the field of computer vision, especially in video surveillance systems. To solve this problem, we will be applying a convolutional neural network of multiple scales at multiple locations in the image in one sliding window. Most translation networks move away from the bounding box around the area of interest. In contrast to this architecture, we consider the problem to be a classification problem where each pixel of the image is a separate section. Image classification is the method of predicting an individual category or specifying by a shoal of data points. Image classification is a part of the classification problem, including any labels throughout the image. The image can be classified as a day or night shot. Or, likewise, images of cars and motorbikes will be automatically placed in their collection. The deep learning of image classification generally includes convolutional layers; the invention of it is referred to as a convolutional neural network (CNN).

Keywords: image classification, object detection, localization, particle filter

Procedia PDF Downloads 283
7063 The Effect of Problem-Based Mobile-Assisted Tasks on Spoken Intelligibility of English as a Foreign Language Learners

Authors: Loghman Ansarian, Teoh Mei Lin

Abstract:

In an attempt to increase oral proficiency of Iranian EFL learners, the researchers compared the effect of problem-based mobile-assisted language learning with the conventional language learning approach (Communicative Language Teaching) in Iran. The experimental group (n=37) went through PBL instruction and the control group (n=33) went through conventional instruction. The results of quantitative data analysis after 26 sessions of treatment revealed that PBL could positively affect participants' knowledge of grammar, vocabulary, spoken fluency, and pronunciation; however, in terms of task achievement, no significant effect was found. This study can have pedagogical implications for language teachers, and material developers.

Keywords: problem-based learning, spoken intelligibility, Iranian EFL context, cognitive learning

Procedia PDF Downloads 163
7062 Optimum Performance of the Gas Turbine Power Plant Using Adaptive Neuro-Fuzzy Inference System and Statistical Analysis

Authors: Thamir K. Ibrahim, M. M. Rahman, Marwah Noori Mohammed

Abstract:

This study deals with modeling and performance enhancements of a gas-turbine combined cycle power plant. A clean and safe energy is the greatest challenges to meet the requirements of the green environment. These requirements have given way the long-time governing authority of steam turbine (ST) in the world power generation, and the gas turbine (GT) will replace it. Therefore, it is necessary to predict the characteristics of the GT system and optimize its operating strategy by developing a simulation system. The integrated model and simulation code for exploiting the performance of gas turbine power plant are developed utilizing MATLAB code. The performance code for heavy-duty GT and CCGT power plants are validated with the real power plant of Baiji GT and MARAFIQ CCGT plants the results have been satisfactory. A new technology of correlation was considered for all types of simulation data; whose coefficient of determination (R2) was calculated as 0.9825. Some of the latest launched correlations were checked on the Baiji GT plant and apply error analysis. The GT performance was judged by particular parameters opted from the simulation model and also utilized Adaptive Neuro-Fuzzy System (ANFIS) an advanced new optimization technology. The best thermal efficiency and power output attained were about 56% and 345MW respectively. Thus, the operation conditions and ambient temperature are strongly influenced on the overall performance of the GT. The optimum efficiency and power are found at higher turbine inlet temperatures. It can be comprehended that the developed models are powerful tools for estimating the overall performance of the GT plants.

Keywords: gas turbine, optimization, ANFIS, performance, operating conditions

Procedia PDF Downloads 408
7061 Optimization of Technical and Technological Solutions for the Development of Offshore Hydrocarbon Fields in the Kaliningrad Region

Authors: Pavel Shcherban, Viktoria Ivanova, Alexander Neprokin, Vladislav Golovanov

Abstract:

Currently, LLC «Lukoil-Kaliningradmorneft» is implementing a comprehensive program for the development of offshore fields of the Kaliningrad region. This is largely associated with the depletion of the resource base of land in the region, as well as the positive results of geological investigation surrounding the Baltic Sea area and the data on the volume of hydrocarbon recovery from a single offshore field are working on the Kaliningrad region – D-6 «Kravtsovskoye».The article analyzes the main stages of the LLC «Lukoil-Kaliningradmorneft»’s development program for the development of the hydrocarbon resources of the region's shelf and suggests an optimization algorithm that allows managing a multi-criteria process of development of shelf deposits. The algorithm is formed on the basis of the problem of sequential decision making, which is a section of dynamic programming. Application of the algorithm during the consolidation of the initial data, the elaboration of project documentation, the further exploration and development of offshore fields will allow to optimize the complex of technical and technological solutions and increase the economic efficiency of the field development project implemented by LLC «Lukoil-Kaliningradmorneft».

Keywords: offshore fields of hydrocarbons of the Baltic Sea, development of offshore oil and gas fields, optimization of the field development scheme, solution of multicriteria tasks in oil and gas complex, quality management in oil and gas complex

Procedia PDF Downloads 185
7060 AHP and TOPSIS Methods for Supplier Selection Problem in Medical Devices Company

Authors: Sevde D. Karayel, Ediz Atmaca

Abstract:

Supplier selection subject is vital because of development competitiveness and performance of firms which have right, rapid and with low cost procurement. Considering the fact that competition between firms is no longer on their supply chains, hence it is very clear that performance of the firms’ not only depend on their own success but also success of all departments in supply chain. For this purpose, firms want to work with suppliers which are cost effective, flexible in terms of demand and high quality level for customer satisfaction. However, diversification and redundancy of their expectations from suppliers, supplier selection problems need to be solved as a hard problem. In this study, supplier selection problem is discussed for critical piece, which is using almost all production of products in and has troubles with lead time from supplier, in a firm that produces medical devices. Analyzing policy in the current situation of the firm in the supplier selection indicates that supplier selection is made based on the purchasing department experience and other authorized persons’ general judgments. Because selection do not make based on the analytical methods, it is caused disruptions in production, lateness and extra cost. To solve the problem, AHP and TOPSIS which are multi-criteria decision making techniques, which are effective, easy to implement and can analyze many criteria simultaneously, are used to make a selection among alternative suppliers.

Keywords: AHP-TOPSIS methods, multi-criteria decision making, supplier selection problem, supply chain management

Procedia PDF Downloads 247
7059 Finite Eigenstrains in Nonlinear Elastic Solid Wedges

Authors: Ashkan Golgoon, Souhayl Sadik, Arash Yavari

Abstract:

Eigenstrains in nonlinear solids are created due to anelastic effects such as non-uniform temperature distributions, growth, remodeling, and defects. Eigenstrains understanding is indispensable, as they can generate residual stresses and strongly affect the overall response of solids. Here, we study the residual stress and deformation fields of an incompressible isotropic infinite wedge with a circumferentially-symmetric distribution of finite eigenstrains. We construct a material manifold, whose Riemannian metric explicitly depends on the eigenstrain distribution, thereby we turn the problem into a classical nonlinear elasticity problem, where we find an embedding of the Riemannian material manifold into the ambient Euclidean space. In particular, we find exact solutions for the residual stress and deformation fields of a neo-Hookean wedge having a symmetric inclusion with finite radial and circumferential eigenstrains. Moreover, we numerically solve a similar problem when a symmetric Mooney-Rivlin inhomogeneity with finite eigenstrains is placed in a neo-Hookean wedge. Generalization of the eigenstrain problem to other geometries are also discussed.

Keywords: finite eigenstrains, geometric mechanics, inclusion, inhomogeneity, nonlinear elasticity

Procedia PDF Downloads 240
7058 Secure Bio Semantic Computing Scheme

Authors: Hiroshi Yamaguchi, Phillip C. Y. Sheu, Ryo Fujita, Shigeo Tsujii

Abstract:

In this paper, the secure BioSemantic Scheme is presented to bridge biological/biomedical research problems and computational solutions via semantic computing. Due to the diversity of problems in various research fields, the semantic capability description language (SCDL) plays and important role as a common language and generic form for problem formalization. SCDL is expected the essential for future semantic and logical computing in Biosemantic field. We show several example to Biomedical problems in this paper. Moreover, in the coming age of cloud computing, the security problem is considered to be crucial issue and we presented a practical scheme to cope with this problem.

Keywords: biomedical applications, private information retrieval (PIR), semantic capability description language (SCDL), semantic computing

Procedia PDF Downloads 377
7057 Elastohydrodynamic Lubrication Study Using Discontinuous Finite Volume Method

Authors: Prawal Sinha, Peeyush Singh, Pravir Dutt

Abstract:

Problems in elastohydrodynamic lubrication have attracted a lot of attention in the last few decades. Solving a two-dimensional problem has always been a big challenge. In this paper, a new discontinuous finite volume method (DVM) for two-dimensional point contact Elastohydrodynamic Lubrication (EHL) problem has been developed and analyzed. A complete algorithm has been presented for solving such a problem. The method presented is robust and easily parallelized in MPI architecture. GMRES technique is implemented to solve the matrix obtained after the formulation. A new approach is followed in which discontinuous piecewise polynomials are used for the trail functions. It is natural to assume that the advantages of using discontinuous functions in finite element methods should also apply to finite volume methods. The nature of the discontinuity of the trail function is such that the elements in the corresponding dual partition have the smallest support as compared with the Classical finite volume methods. Film thickness calculation is done using singular quadrature approach. Results obtained have been presented graphically and discussed. This method is well suited for solving EHL point contact problem and can probably be used as commercial software.

Keywords: elastohydrodynamic, lubrication, discontinuous finite volume method, GMRES technique

Procedia PDF Downloads 243
7056 Pinch Technology for Minimization of Water Consumption at a Refinery

Authors: W. Mughees, M. Alahmad

Abstract:

Water is the most significant entity that controls local and global development. For the Gulf region, especially Saudi Arabia, with its limited potable water resources, the potential of the fresh water problem is highly considerable. In this research, the study involves the design and analysis of pinch-based water/wastewater networks. Multiple water/wastewater networks were developed using pinch analysis involving direct recycle/material recycle method. Property-integration technique was adopted to carry out direct recycle method. Particularly, a petroleum refinery was considered as a case study. In direct recycle methodology, minimum water discharge and minimum fresh water resource targets were estimated. Re-design (or retrofitting) of water allocation in the networks was undertaken. Chemical Oxygen Demand (COD) and hardness properties were taken as pollutants. This research was based on single and double contaminant approach for COD and hardness and the amount of fresh water was reduced from 340.0 m3/h to 149.0 m3/h (43.8%), 208.0 m3/h (61.18%) respectively. While regarding double contaminant approach, reduction in fresh water demand was 132.0 m3/h (38.8%). The required analysis was also carried out using mathematical programming technique. Operating software such as LINGO was used for these studies which have verified the graphical method results in a valuable and accurate way. Among the multiple water networks, the one possible water allocation network was developed based on mass exchange.

Keywords: minimization, water pinch, water management, pollution prevention

Procedia PDF Downloads 459
7055 Fuzzy Nail Cream Formula Treatment with Basic Iranian Traditional Medicine

Authors: Elahe Najafizade, Ahmad Mohammad Alkhateeb, Seyed Ali Hossein Zahraei, Iman Dianat

Abstract:

Introduction: Hangnails are short, torn, down parts of the skin surrounding the nails. At times they are very painful. The usual treatment advised is cutting the excess skin with clippers or scissors. To provide instant relief to the patients, we describe a simpler and more effective way to use surgical glue to paste them back into their original position. Method: The cream should not be on the heat; it is on the bain-marie. To achieve the desired emulsifier, 1 gram of borax was mixed in 10 grams of distilled water in a bain-marie until it melted, then stirred oserin, beeswax, and oil in the bain-marie until it melted. After that, 32 grams of distilled water was added little by little. We add and stir and gradually add the borax dissolved in 10 grams of distilled water. The bowl of cream was placed in a bowl of cold water and stirred until the cream was smooth. After that, we add gasoline, alcohol, or methylparaben preservatives. It should be noted that this amount of ingredients is enough for a 350-gram can (when we prepare the cream, we also add the extract). Result: The patient was a 40-year-old female with a hangnail problem that had been used several different creams and Vaseline, but the treatment was not useful, but after this cream was applied for treatment; the hangnail started to cure within one week, and complete treatment achieved after two weeks. Conclusion: Traditional methods with modification without using chemical substances somehow work better and safer, so research programs on them will be useful for less risky treatment procedures.

Keywords: nail, cream, formula, traditional medicine

Procedia PDF Downloads 88
7054 Building 1-Well-Covered Graphs by Corona, Join, and Rooted Product of Graphs

Authors: Vadim E. Levit, Eugen Mandrescu

Abstract:

A graph is well-covered if all its maximal independent sets are of the same size. A well-covered graph is 1-well-covered if deletion of every vertex of the graph leaves it well-covered. It is known that a graph without isolated vertices is 1-well-covered if and only if every two disjoint independent sets are included in two disjoint maximum independent sets. Well-covered graphs are related to combinatorial commutative algebra (e.g., every Cohen-Macaulay graph is well-covered, while each Gorenstein graph without isolated vertices is 1-well-covered). Our intent is to construct several infinite families of 1-well-covered graphs using the following known graph operations: corona, join, and rooted product of graphs. Adopting some known techniques used to advantage for well-covered graphs, one can prove that: if the graph G has no isolated vertices, then the corona of G and H is 1-well-covered if and only if H is a complete graph of order two at least; the join of the graphs G and H is 1-well-covered if and only if G and H have the same independence number and both are 1-well-covered; if H satisfies the property that every three pairwise disjoint independent sets are included in three pairwise disjoint maximum independent sets, then the rooted product of G and H is 1-well-covered, for every graph G. These findings show not only how to generate some more families of 1-well-covered graphs, but also that, to this aim, sometimes, one may use graphs that are not necessarily 1-well-covered.

Keywords: maximum independent set, corona, concatenation, join, well-covered graph

Procedia PDF Downloads 187
7053 Asymptotic Expansion of Double Oscillatory Integrals: Contribution of Non Stationary Critical Points of the Second Kind

Authors: Abdallah Benaissa

Abstract:

In this paper, we consider the problem of asymptotics of double oscillatory integrals in the case of critical points of the second kind, the order of contact between the boundary and a level curve of the phase being even, the situation when the order of contact is odd will be studied in other occasions. Complete asymptotic expansions will be derived and the coefficient of the leading term will be computed in terms of the original data of the problem. A multitude of people have studied this problem using a variety of methods, but only in a special case when the order of contact is minimal: the more cited papers are a paper of Jones and Kline and an other one of Chako. These integrals are encountered in many areas of science, especially in problems of diffraction of optics.

Keywords: asymptotic expansion, double oscillatory integral, critical point of the second kind, optics diffraction

Procedia PDF Downloads 335
7052 Agile Smartphone Porting and App Integration of Signal Processing Algorithms Obtained through Rapid Development

Authors: Marvin Chibuzo Offiah, Susanne Rosenthal, Markus Borschbach

Abstract:

Certain research projects in Computer Science often involve research on existing signal processing algorithms and developing improvements on them. Research budgets are usually limited, hence there is limited time for implementing the algorithms from scratch. It is therefore common practice, to use implementations provided by other researchers as a template. These are most commonly provided in a rapid development, i.e. 4th generation, programming language, usually Matlab. Rapid development is a common method in Computer Science research for quickly implementing and testing new developed algorithms, which is also a common task within agile project organization. The growing relevance of mobile devices in the computer market also gives rise to the need to demonstrate the successful executability and performance measurement of these algorithms on a mobile device operating system and processor, particularly on a smartphone. Open mobile systems such as Android, are most suitable for this task, which is to be performed most efficiently. Furthermore, efficiently implementing an interaction between the algorithm and a graphical user interface (GUI) that runs exclusively on the mobile device is necessary in cases where the project’s goal statement also includes such a task. This paper examines different proposed solutions for porting computer algorithms obtained through rapid development into a GUI-based smartphone Android app and evaluates their feasibilities. Accordingly, the feasible methods are tested and a short success report is given for each tested method.

Keywords: SMARTNAVI, Smartphone, App, Programming languages, Rapid Development, MATLAB, Octave, C/C++, Java, Android, NDK, SDK, Linux, Ubuntu, Emulation, GUI

Procedia PDF Downloads 466
7051 Solving Nonconvex Economic Load Dispatch Problem Using Particle Swarm Optimization with Time Varying Acceleration Coefficients

Authors: Alireza Alizadeh, Hossein Ghadimi, Oveis Abedinia, Noradin Ghadimi

Abstract:

A Particle Swarm Optimization with Time Varying Acceleration Coefficients (PSO-TVAC) is proposed to determine optimal economic load dispatch (ELD) problem in this paper. The proposed methodology easily takes care of solving non-convex economic load dispatch problems along with different constraints like transmission losses, dynamic operation constraints and prohibited operating zones. The proposed approach has been implemented on the 3-machines 6-bus, IEEE 5-machines 14-bus, IEEE 6-machines 30-bus systems and 13 thermal units power system. The proposed technique is compared to solve the ELD problem with hybrid approach by using the valve-point effect. The comparison results prove the capability of the proposed method giving significant improvements in the generation cost for the economic load dispatch problem.

Keywords: PSO-TVAC, economic load dispatch, non-convex cost function, prohibited operating zone, transmission losses

Procedia PDF Downloads 372
7050 Solving of Types Mathematical Routine and Non-Routine Problems in Algebra

Authors: Verónica Díaz Quezada

Abstract:

The importance given to the development of the problem solving skill and the requirement to solve problems framed in mathematical or real life contexts, in practice, they are not evidence in relation to the teaching of proportional variations. This qualitative and descriptive study aims to (1) to improve problem solving ability of high school students in Chile, (ii) to elaborate and describe a didactic intervention strategy based on learning situations in proportional variations, focused on solving types of routine problems of various contexts and non-routine problems. For this purpose, participant observation was conducted, test of mathematics problems and an opinion questionnaire to thirty-six high school students. Through the results, the highest academic performance is evidenced in the routine problems of purely mathematical context, realistic, fantasy context, and non-routine problems, except in the routine problems of real context and compound proportionality problems. The results highlight the need to consider in the curriculum different types of problems in the teaching of mathematics that relate the discipline to everyday life situations

Keywords: algebra, high school, proportion variations, nonroutine problem solving, routine problem solving

Procedia PDF Downloads 119
7049 How Geant4 Hadronic Models Handle Tracking of Pion Particles Resulting from Antiproton Annihilation

Authors: M. B. Tavakoli, R. Reiazi, M. M. Mohammadi, K. Jabbari

Abstract:

From 2003, AD4/ACE experiment in CERN tried to investigate different aspects of antiproton as a new modality in particle therapy. Because of lack of reliable absolute dose measurements attempts to find out the radiobiological characteristics of antiproton have not reached to a reasonable result yet. From the other side, application of Geant4 in medical approaches is increased followed by Geant4-DNA project which focuses on using this code to predict radiation effects in the cellular scale. This way we can exploit Geant4-DNA results for antiproton. Unfortunately, previous studies showed there are serious problem in simulating an antiproton beam using Geant4. Since most of the problem was in the Bragg peak region which antiproton annihilates there, in this work we tried to understand if the problem came from the way in which Geant4 handles annihilation products especially pion particles. This way, we can predict the source of the dose discrepancies between Geant4 simulations and dose measurements done in CERN.

Keywords: Geant4, antiproton, annihilation, pion plus, pion minus

Procedia PDF Downloads 642