Search results for: optimization for learning and data analysis
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 15606

Search results for: optimization for learning and data analysis

13236 Modified PSO Based Optimal Control for Maximizing Benefits of Distributed Generation System

Authors: Priyanka Sen, Kaibalya Prasad Panda, Soumyakanta Samantaray, Sreyasee Rout, Bishnupriya Biswal

Abstract:

Deregulation in the power system industry and the invention of new technologies for producing electrical energy has led to innovations in power system planning. Distributed generation (DG) is one of the most attractive technologies that bring different kinds of advantages to a lot of entities, engaged in power systems. In this paper, a model for considering DGs in the power system planning problem is presented. Dynamic power system planning for reduction of maintenance and operational cost is presented in this paper. In addition to that, a modified particle swarm optimization (PSO) is used to find the optimal topology solution. Voltage Profile Improvement Index (VPII) and Line Loss Reduction Index (LLRI) are taken as benefit index of employing DG. The effectiveness of this method is demonstrated through examination of IEEE 30 bus test system.

Keywords: Distributed generation, line loss reduction index, particle swarm optimization, power system, voltage profile improvement index.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 892
13235 Quantifying and Adjusting the Effects of Publication Bias in Continuous Meta-Analysis

Authors: N.R.N. Idris

Abstract:

This study uses simulated meta-analysis to assess the effects of publication bias on meta-analysis estimates and to evaluate the efficacy of the trim and fill method in adjusting for these biases. The estimated effect sizes and the standard error were evaluated in terms of the statistical bias and the coverage probability. The results demonstrate that if publication bias is not adjusted it could lead to up to 40% bias in the treatment effect estimates. Utilization of the trim and fill method could reduce the bias in the overall estimate by more than half. The method is optimum in presence of moderate underlying bias but has minimal effects in presence of low and severe publication bias. Additionally, the trim and fill method improves the coverage probability by more than half when subjected to the same level of publication bias as those of the unadjusted data. The method however tends to produce false positive results and will incorrectly adjust the data for publication bias up to 45 % of the time. Nonetheless, the bias introduced into the estimates due to this adjustment is minimal

Keywords: Publication bias, Trim and Fill method, percentage relative bias, coverage probability

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1539
13234 A Case Study of Mobile Game Based Learning Design for Gender Responsive STEM Education

Authors: Raluca Ionela Maxim

Abstract:

Designing a gender responsive Science, Technology, Engineering and Mathematics (STEM) mobile game based learning solution (mGBL) is a challenge in terms of content, gamification level and equal engagement of girls and boys. The goal of this case study was to research and create a high-fidelity prototype design of a mobile game that contains role-models as avatars that guide and expose girls and boys to STEM learning content. For this research purpose it was applied the methodology of design sprint with five-phase process that combines design thinking principles. The technique of this methodology comprises smart interviews with STEM experts, mind-map creation, sketching, prototyping and usability testing of the interactive prototype of the gender responsive STEM mGBL. The results have shown that the effect of the avatar/role model had a positive impact. Therefore, by exposing students (boys and girls) to STEM role models in an mGBL tool is helpful for the decreasing of the gender inequalities in STEM fields.

Keywords: Design thinking, design sprint, gender-responsive STEM education, mobile game based learning, role-models.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1325
13233 An Effort at Improving Reliability of Laboratory Data in Titrimetric Analysis for Zinc Sulphate Tablets Using Validated Spreadsheet Calculators

Authors: M. A. Okezue, K. L. Clase, S. R. Byrn

Abstract:

The requirement for maintaining data integrity in laboratory operations is critical for regulatory compliance. Automation of procedures reduces incidence of human errors. Quality control laboratories located in low-income economies may face some barriers in attempts to automate their processes. Since data from quality control tests on pharmaceutical products are used in making regulatory decisions, it is important that laboratory reports are accurate and reliable. Zinc Sulphate (ZnSO4) tablets is used in treatment of diarrhea in pediatric population, and as an adjunct therapy for COVID-19 regimen. Unfortunately, zinc content in these formulations is determined titrimetrically; a manual analytical procedure. The assay for ZnSO4 tablets involves time-consuming steps that contain mathematical formulae prone to calculation errors. To achieve consistency, save costs, and improve data integrity, validated spreadsheets were developed to simplify the two critical steps in the analysis of ZnSO4 tablets: standardization of 0.1M Sodium Edetate (EDTA) solution, and the complexometric titration assay procedure. The assay method in the United States Pharmacopoeia was used to create a process flow for ZnSO4 tablets. For each step in the process, different formulae were input into two spreadsheets to automate calculations. Further checks were created within the automated system to ensure validity of replicate analysis in titrimetric procedures. Validations were conducted using five data sets of manually computed assay results. The acceptance criteria set for the protocol were met. Significant p-values (p < 0.05, α = 0.05, at 95% Confidence Interval) were obtained from students’ t-test evaluation of the mean values for manual-calculated and spreadsheet results at all levels of the analysis flow. Right-first-time analysis and principles of data integrity were enhanced by use of the validated spreadsheet calculators in titrimetric evaluations of ZnSO4 tablets. Human errors were minimized in calculations when procedures were automated in quality control laboratories. The assay procedure for the formulation was achieved in a time-efficient manner with greater level of accuracy. This project is expected to promote cost savings for laboratory business models.

Keywords: Data integrity, spreadsheets, titrimetry, validation, zinc sulphate tablets.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 483
13232 Efficiency Evaluation of E-Commerce Websites

Authors: A. K. Abd El-Aleem, W. F. Abd El-wahed, N. A. Ismail, F. A. Torkey

Abstract:

This study suggests a model of a new set of evaluation criteria that will be used to measure the efficiency of real-world E-commerce websites. Evaluation criteria include design, usability and performance for websites, the Data Envelopment Analysis (DEA) technique has been used to measure the websites efficiency. An efficient Web site is defined as a site that generates the most outputs, using the smallest amount of inputs. Inputs refer to measurements representing the amount of effort required to build, maintain and perform the site. Output is amount of traffic the site generates. These outputs are measured as the average number of daily hits and the average number of daily unique visitors.

Keywords: Data Envelopment Analysis, E-commerce, Efficiency.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4077
13231 An Analysis of Dynamic Economic Dispatch Using Search Space Reduction Based Gravitational Search Algorithm

Authors: K. C. Meher, R. K. Swain, C. K. Chanda

Abstract:

This paper presents the performance analysis of dynamic search space reduction (DSR) based gravitational search algorithm (GSA) to solve dynamic economic dispatch of thermal generating units with valve point effects. Dynamic economic dispatch basically dictates the best setting of generator units with anticipated load demand over a definite period of time. In this paper, the presented technique is considered that deals an inequality constraints treatment mechanism known as DSR strategy to accelerate the optimization process. The presented method is demonstrated through five-unit test systems to verify its effectiveness and robustness. The simulation results are compared with other existing evolutionary methods reported in the literature. It is intuited from the comparison that the fuel cost and other performances of the presented approach yield fruitful results with marginal value of simulation time.

Keywords: Dynamic economic dispatch, dynamic search space reduction strategy, gravitational search algorithm, ramp rate limits, valve-point effects.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1472
13230 Developing New Algorithm and Its Application on Optimal Control of Pumps in Water Distribution Network

Authors: R. Rajabpour, N. Talebbeydokhti, M. H. Ahmadi

Abstract:

In recent years, new techniques for solving complex problems in engineering are proposed. One of these techniques is JPSO algorithm. With innovative changes in the nature of the jump algorithm JPSO, it is possible to construct a graph-based solution with a new algorithm called G-JPSO. In this paper, a new algorithm to solve the optimal control problem Fletcher-Powell and optimal control of pumps in water distribution network was evaluated. Optimal control of pumps comprise of optimum timetable operation (status on and off) for each of the pumps at the desired time interval. Maximum number of status on and off for each pumps imposed to the objective function as another constraint. To determine the optimal operation of pumps, a model-based optimization-simulation algorithm was developed based on G-JPSO and JPSO algorithms. The proposed algorithm results were compared well with the ant colony algorithm, genetic and JPSO results. This shows the robustness of proposed algorithm in finding near optimum solutions with reasonable computational cost.

Keywords: G-JPSO, operation, optimization, pumping station, water distribution networks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1620
13229 Spatially Random Sampling for Retail Food Risk Factors Study

Authors: Guilan Huang

Abstract:

In 2013 and 2014, the U.S. Food and Drug Administration (FDA) collected data from selected fast food restaurants and full service restaurants for tracking changes in the occurrence of foodborne illness risk factors. This paper discussed how we customized spatial random sampling method by considering financial position and availability of FDA resources, and how we enriched restaurants data with location. Location information of restaurants provides opportunity for quantitatively determining random sampling within non-government units (e.g.: 240 kilometers around each data-collector). Spatial analysis also could optimize data-collectors’ work plans and resource allocation. Spatial analytic and processing platform helped us handling the spatial random sampling challenges. Our method fits in FDA’s ability to pinpoint features of foodservice establishments, and reduced both time and expense on data collection.

Keywords: Geospatial technology, restaurant, retail food risk factors study, spatial random sampling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1439
13228 Feature Analysis of Predictive Maintenance Models

Authors: Zhaoan Wang

Abstract:

Research in predictive maintenance modeling has improved in the recent years to predict failures and needed maintenance with high accuracy, saving cost and improving manufacturing efficiency. However, classic prediction models provide little valuable insight towards the most important features contributing to the failure. By analyzing and quantifying feature importance in predictive maintenance models, cost saving can be optimized based on business goals. First, multiple classifiers are evaluated with cross-validation to predict the multi-class of failures. Second, predictive performance with features provided by different feature selection algorithms are further analyzed. Third, features selected by different algorithms are ranked and combined based on their predictive power. Finally, linear explainer SHAP (SHapley Additive exPlanations) is applied to interpret classifier behavior and provide further insight towards the specific roles of features in both local predictions and global model behavior. The results of the experiments suggest that certain features play dominant roles in predictive models while others have significantly less impact on the overall performance. Moreover, for multi-class prediction of machine failures, the most important features vary with type of machine failures. The results may lead to improved productivity and cost saving by prioritizing sensor deployment, data collection, and data processing of more important features over less importance features.

Keywords: Automated supply chain, intelligent manufacturing, predictive maintenance machine learning, feature engineering, model interpretation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1967
13227 Computational Aspects of Regression Analysis of Interval Data

Authors: Michal Cerny

Abstract:

We consider linear regression models where both input data (the values of independent variables) and output data (the observations of the dependent variable) are interval-censored. We introduce a possibilistic generalization of the least squares estimator, so called OLS-set for the interval model. This set captures the impact of the loss of information on the OLS estimator caused by interval censoring and provides a tool for quantification of this effect. We study complexity-theoretic properties of the OLS-set. We also deal with restricted versions of the general interval linear regression model, in particular the crisp input – interval output model. We give an argument that natural descriptions of the OLS-set in the crisp input – interval output cannot be computed in polynomial time. Then we derive easily computable approximations for the OLS-set which can be used instead of the exact description. We illustrate the approach by an example.

Keywords: Linear regression, interval-censored data, computational complexity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1449
13226 Analytical Modelling of Surface Roughness during Compacted Graphite Iron Milling Using Ceramic Inserts

Authors: S. Karabulut, A. Güllü, A. Güldas, R. Gürbüz

Abstract:

This study investigates the effects of the lead angle and chip thickness variation on surface roughness during the machining of compacted graphite iron using ceramic cutting tools under dry cutting conditions. Analytical models were developed for predicting the surface roughness values of the specimens after the face milling process. Experimental data was collected and imported to the artificial neural network model. A multilayer perceptron model was used with the back propagation algorithm employing the input parameters of lead angle, cutting speed and feed rate in connection with chip thickness. Furthermore, analysis of variance was employed to determine the effects of the cutting parameters on surface roughness. Artificial neural network and regression analysis were used to predict surface roughness. The values thus predicted were compared with the collected experimental data, and the corresponding percentage error was computed. Analysis results revealed that the lead angle is the dominant factor affecting surface roughness. Experimental results indicated an improvement in the surface roughness value with decreasing lead angle value from 88° to 45°.

Keywords: CGI, milling, surface roughness, ANN, regression, modeling, analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1951
13225 A Study of the Problems and Demands of Community Leaders- Training in the Upper Northeastern Region

Authors: Teerawach Khamkorn, Laongtip Mathurasa, Savittree Rochanasmita Arnold, Witthaya Mekhum

Abstract:

This research is aimed at studying the nature of problems and demands of the training for community leaders in the upper northeastern region of Thailand. Population and group samplings are based on 360 community leaders in the region who have experienced prior training from the Udonthani Rajabhat University. Stratified random samplings have been drawn upon 186 participants. The research tools is questionnaires. The frequency, percentage and standard deviation are employed in data analysis. The findings indicate that most of community leaders are males and senior adults. The problems in training are associated with the inconveniences of long-distance travelling to training locations, inadequacy of learning centers and training sites and high training costs. The demand of training is basically motivated by a desire for self-development in modern knowledge in keeping up-to-date with the changing world and the need for technological application and facilitation in shortening the distance to training locations and in limiting expensive training costs.

Keywords: Community leaders, Distance Training, Management, Technology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1340
13224 Modeling Language for Machine Learning

Authors: Tsuyoshi Okita, Tatsuya Niwa

Abstract:

For a given specific problem an efficient algorithm has been the matter of study. However, an alternative approach orthogonal to this approach comes out, which is called a reduction. In general for a given specific problem this reduction approach studies how to convert an original problem into subproblems. This paper proposes a formal modeling language to support this reduction approach. We show three examples from the wide area of learning problems. The benefit is a fast prototyping of algorithms for a given new problem.

Keywords: Formal language, statistical inference problem, reduction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1594
13223 Reliability Analysis of Press Unit using Vague Set

Authors: S. P. Sharma, Monica Rani

Abstract:

In conventional reliability assessment, the reliability data of system components are treated as crisp values. The collected data have some uncertainties due to errors by human beings/machines or any other sources. These uncertainty factors will limit the understanding of system component failure due to the reason of incomplete data. In these situations, we need to generalize classical methods to fuzzy environment for studying and analyzing the systems of interest. Fuzzy set theory has been proposed to handle such vagueness by generalizing the notion of membership in a set. Essentially, in a Fuzzy Set (FS) each element is associated with a point-value selected from the unit interval [0, 1], which is termed as the grade of membership in the set. A Vague Set (VS), as well as an Intuitionistic Fuzzy Set (IFS), is a further generalization of an FS. Instead of using point-based membership as in FS, interval-based membership is used in VS. The interval-based membership in VS is more expressive in capturing vagueness of data. In the present paper, vague set theory coupled with conventional Lambda-Tau method is presented for reliability analysis of repairable systems. The methodology uses Petri nets (PN) to model the system instead of fault tree because it allows efficient simultaneous generation of minimal cuts and path sets. The presented method is illustrated with the press unit of the paper mill.

Keywords: Lambda -Tau methodology, Petri nets, repairable system, vague fuzzy set.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1512
13222 Reducing Test Vectors Count Using Fault Based Optimization Schemes in VLSI Testing

Authors: Vinod Kumar Khera, R. K. Sharma, A. K. Gupta

Abstract:

Power dissipation increases exponentially during test mode as compared to normal operation of the circuit. In extreme cases, test power is more than twice the power consumed during normal operation mode. Test vector generation scheme is key component in deciding the power hungriness of a circuit during testing. Test vector count and consequent leakage current are functions of test vector generation scheme. Fault based test vector count optimization has been presented in this work. It helps in reducing test vector count and the leakage current. In the presented scheme, test vectors have been reduced by extracting essential child vectors. The scheme has been tested experimentally using stuck at fault models and results ensure the reduction in test vector count.

Keywords: Low power VLSI testing, independent fault, essential faults, test vector reduction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1401
13221 Optimization of Enzymatic Hydrolysis of Manihot Esculenta Root Starch by Immobilizeda-Amylase Using Response Surface Methodology

Authors: G. Baskar, C. Muthukumaran, S. Renganathan

Abstract:

Enzymatic hydrolysis of starch from natural sources finds potential application in commercial production of alcoholic beverage and bioethanol. In this study the effect of starch concentration, temperature, time and enzyme concentration were studied and optimized for hydrolysis of cassava (Manihot esculenta) starch powder (of mesh 80/120) into glucose syrup by immobilized (using Polyacrylamide gel) a-amylase using central composite design. The experimental result on enzymatic hydrolysis of cassava starch was subjected to multiple linear regression analysis using MINITAB 14 software. Positive linear effect of starch concentration, enzyme concentration and time was observed on hydrolysis of cassava starch by a-amylase. The statistical significance of the model was validated by F-test for analysis of variance (p < 0.01). The optimum value of starch concentration temperature, time and enzyme concentration were found to be 4.5% (w/v), 45oC, 150 min, and 1% (w/v) enzyme. The maximum glucose yield at optimum condition was 5.17 mg/mL.

Keywords: Enzymatic hydrolysis, Alcoholic beverage, Centralcomposite design, Polynomial model, glucose yield.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2211
13220 Implementing Education 4.0 Trends in Language Learning

Authors: Luz Janeth Ospina M.

Abstract:

The fourth industrial revolution is changing the role of education substantially and, therefore, the role of instructors and learners at all levels. Education 4.0 is an imminent response to the needs of a globalized world where humans and technology are being aligned to enable endless possibilities, among them the need for students, as digital natives, to communicate effectively in at least one language besides their mother tongue, and also the requirement of developing theirs. This is an exploratory study in which a control group (N = 21), all of the students of Spanish as a foreign language at the university level, after taking a Spanish class, responded to an online questionnaire about the engagement, atmosphere, and environment in which their course was delivered. These aspects considered in the survey were relative to the instructor’s teaching style, including: (a) active, hands-on learning; (b) flexibility for in-class activities, easily switching between small group work, individual work, and whole-class discussion; and (c) integrating technology into the classroom. Strongly believing in these principles, the instructor deliberately taught the course in a SCALE-UP room, as it could facilitate such a positive and encouraging learning environment. These aspects are trends related to Education 4.0 and have become integral to the instructor’s pedagogical stance that calls for a constructive-affective role, instead of a transmissive one. As expected, with a learning environment that (a) fosters student engagement and (b) improves student outcomes, the subjects were highly engaged, which was partially due to the learning environment. An overwhelming majority (all but one) of students agreed or strongly agreed that the atmosphere and the environment were ideal. Outcomes of this study are relevant and indicate that it is about time for teachers to build up a meaningful correlation between humans and technology. We should see the trends of Education 4.0 not as a threat but as practices that should be in the hands of critical and creative instructors whose pedagogical stance responds to the needs of the learners in the 21st century.

Keywords: Active learning, education 4.0, higher education, pedagogical stance.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 673
13219 Professional Development of Pre-Service Teachers: The Case of Practicum Experience

Authors: G. Lingam, N. Lingam, K. Raghuwaiya

Abstract:

The reported study focuses on pre-service teachers’ professional development during the teaching practice. The cohort studied comprised participants in their final year in the Bachelor of Arts and Bachelor of Science with Graduate Certificate in Education programmes of a university in Fiji. Analysis of the data obtained using a survey questionnaire indicates that overall, the pre-service teachers were satisfied with the practicum experience. This is assumed to demonstrate that the practicum experience contributed well towards their professional preparation for work expected of them in Fiji secondary schools. Participants also identified some concerns as needing attention. To conclude, the paper provides suggestions for improving the preparation of teachers by strengthening the identified areas of the practicum offered by the university. The study has implications for other teacher education providers in small developing island states and even beyond for the purpose of enhancing learning in student teachers’ for future work.

Keywords: Pre-service, teacher education, practicum, teachers’ world of work, student teachers.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3809
13218 Thermal Characterization of Smart and Large-Scale Building Envelope System in a Subtropical Climate

Authors: Andrey A. Chernousov, Ben Y. B. Chan

Abstract:

The thermal behavior of a large-scale, phase change material (PCM) enhanced building envelope system was studied in regard to the need for pre-fabricated construction in subtropical regions. The proposed large-scale envelope consists of a reinforced aluminum skin, insulation core, phase change material and reinforced gypsum board. The PCM impact on an energy efficiency of an enveloped room was resolved by validation of the EnergyPlus numerical scheme and optimization of a smart material location in the core. The PCM location was optimized by a minimization method of a cooling energy demand. It has been shown that there is good agreement between the test and simulation results. The optimal location of the PCM layer in Hong Kong summer conditions has been then recomputed for core thicknesses of 40, 60 and 80 mm. A non-dimensional value of the optimal PCM location was obtained to be same for all the studied cases and the considered external and internal conditions.

Keywords: Thermal performance, phase change material, energy efficiency, PCM optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2210
13217 Optimization of Fuel Consumption of a Bus used in City Line with Regulation of Driving Characteristics

Authors: Muammer Ozkan, Orkun Ozener, Irfan Yavasliol

Abstract:

The fuel cost of the motor vehicle operating on its common route is an important part of the operating cost. Therefore, the importance of the fuel saving is increasing day by day. One of the parameters which improve fuel saving is the regulation of driving characteristics. The number and duration of stop is increased by the heavy traffic load. It is possible to improve the fuel saving with regulation of traffic flow and driving characteristics. The researches show that the regulation of the traffic flow decreases fuel consumption, but it is not enough to improve fuel saving without the regulation of driving characteristics. This study analyses the fuel consumption of two trips of city bus operating on its common route and determines the effect of traffic density and driving characteristics on fuel consumption. Finally it offers some suggestions about regulation of driving characteristics to improve the fuel saving. Fuel saving is determined according to the results obtained from simulation program. When experimental and simulation results are compared, it has been found that the fuel saving was reached up the to 40 percent ratios.

Keywords: Fuel Consumption, Fuel Economy, Driving Characteristics, Optimization

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1815
13216 Lifetime Maximization in Wireless Ad Hoc Networks with Network Coding and Matrix Game

Authors: Jain-Shing Liu

Abstract:

In this paper, we present a matrix game-theoretic cross-layer optimization formulation to maximize the network lifetime in wireless ad hoc networks with network coding. To this end, we introduce a cross-layer formulation of general NUM (network utility maximization) that accommodates routing, scheduling, and stream control from different layers in the coded networks. Specifically, for the scheduling problem and then the objective function involved, we develop a matrix game with the strategy sets of the players corresponding to hyperlinks and transmission modes, and design the payoffs specific to the lifetime. In particular, with the inherit merit that matrix game can be solved with linear programming, our cross-layer programming formulation can benefit from both game-based and NUM-based approaches at the same time by cooperating the programming model for the matrix game with that for the other layers in a consistent framework. Finally, our numerical example demonstrates its performance results on a well-known wireless butterfly network to verify the cross-layer optimization scheme.

Keywords: Cross-layer design, Lifetime maximization, Matrix game, Network coding

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1681
13215 Context Detection in Spreadsheets Based on Automatically Inferred Table Schema

Authors: Alexander Wachtel, Michael T. Franzen, Walter F. Tichy

Abstract:

Programming requires years of training. With natural language and end user development methods, programming could become available to everyone. It enables end users to program their own devices and extend the functionality of the existing system without any knowledge of programming languages. In this paper, we describe an Interactive Spreadsheet Processing Module (ISPM), a natural language interface to spreadsheets that allows users to address ranges within the spreadsheet based on inferred table schema. Using the ISPM, end users are able to search for values in the schema of the table and to address the data in spreadsheets implicitly. Furthermore, it enables them to select and sort the spreadsheet data by using natural language. ISPM uses a machine learning technique to automatically infer areas within a spreadsheet, including different kinds of headers and data ranges. Since ranges can be identified from natural language queries, the end users can query the data using natural language. During the evaluation 12 undergraduate students were asked to perform operations (sum, sort, group and select) using the system and also Excel without ISPM interface, and the time taken for task completion was compared across the two systems. Only for the selection task did users take less time in Excel (since they directly selected the cells using the mouse) than in ISPM, by using natural language for end user software engineering, to overcome the present bottleneck of professional developers.

Keywords: Natural language processing, end user development; natural language interfaces, human computer interaction, data recognition, dialog systems, spreadsheet.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1103
13214 Predicting the Minimum Free Energy RNA Secondary Structures using Harmony Search Algorithm

Authors: Abdulqader M. Mohsen, Ahamad Tajudin Khader, Dhanesh Ramachandram, Abdullatif Ghallab

Abstract:

The physical methods for RNA secondary structure prediction are time consuming and expensive, thus methods for computational prediction will be a proper alternative. Various algorithms have been used for RNA structure prediction including dynamic programming and metaheuristic algorithms. Musician's behaviorinspired harmony search is a recently developed metaheuristic algorithm which has been successful in a wide variety of complex optimization problems. This paper proposes a harmony search algorithm (HSRNAFold) to find RNA secondary structure with minimum free energy and similar to the native structure. HSRNAFold is compared with dynamic programming benchmark mfold and metaheuristic algorithms (RnaPredict, SetPSO and HelixPSO). The results showed that HSRNAFold is comparable to mfold and better than metaheuristics in finding the minimum free energies and the number of correct base pairs.

Keywords: Metaheuristic algorithms, dynamic programming algorithms, harmony search optimization, RNA folding, Minimum free energy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2315
13213 Optimization of Supersonic Ejector via Sequence-Adapted Micro-Genetic Algorithm

Authors: Kolar Jan, Dvorak Vaclav

Abstract:

In this study, an optimization of supersonic air-to-air ejector is carried out by a recently developed single-objective genetic algorithm based on adaption of sequence of individuals. Adaptation of sequence is based on Shape-based distance of individuals and embedded micro-genetic algorithm. The optimal sequence found defines the succession of CFD-aimed objective calculation within each generation of regular micro-genetic algorithm. A spring-based deformation mutates the computational grid starting the initial individualvia adapted population in the optimized sequence. Selection of a generation initial individual is knowledge-based. A direct comparison of the newly defined and standard micro-genetic algorithm is carried out for supersonic air-to-air ejector. The only objective is to minimize the loose of total stagnation pressure in the ejector. The result is that sequence-adopted micro-genetic algorithm can provide comparative results to standard algorithm but in significantly lower number of overall CFD iteration steps.

Keywords: Grid deformation, Micro-genetic algorithm, shapebased sequence, supersonic ejector.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1539
13212 Exploiting Machine Learning Techniques for the Enhancement of Acceptance Sampling

Authors: Aikaterini Fountoulaki, Nikos Karacapilidis, Manolis Manatakis

Abstract:

This paper proposes an innovative methodology for Acceptance Sampling by Variables, which is a particular category of Statistical Quality Control dealing with the assurance of products quality. Our contribution lies in the exploitation of machine learning techniques to address the complexity and remedy the drawbacks of existing approaches. More specifically, the proposed methodology exploits Artificial Neural Networks (ANNs) to aid decision making about the acceptance or rejection of an inspected sample. For any type of inspection, ANNs are trained by data from corresponding tables of a standard-s sampling plan schemes. Once trained, ANNs can give closed-form solutions for any acceptance quality level and sample size, thus leading to an automation of the reading of the sampling plan tables, without any need of compromise with the values of the specific standard chosen each time. The proposed methodology provides enough flexibility to quality control engineers during the inspection of their samples, allowing the consideration of specific needs, while it also reduces the time and the cost required for these inspections. Its applicability and advantages are demonstrated through two numerical examples.

Keywords: Acceptance Sampling, Neural Networks, Statistical Quality Control.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1674
13211 Optimization of R507A-R23 Cascade Refrigeration System using Genetic Algorithm

Authors: A. D. Parekh, P. R. Tailor, H.R Jivanramajiwala

Abstract:

The present work deals with optimization of cascade refrigeration system using eco friendly refrigerants pair R507A and R23. R507A is azeotropic mixture composed of HFC refrigerants R125/R143a (50%/50% by wt.). R23 is a single component HFC refrigerant used as replacement to CFC refrigerant R13 in low temperature applications. These refrigerants have zero ozone depletion potential and are non-flammable. Optimization of R507AR23 cascade refrigeration system performance parameters such as minimum work required, refrigeration effect, coefficient of performance and exergetic efficiency was carried out in terms of eight operating parameters- combinations using Genetic Algorithm tool. The eight operating parameters include (1) low side evaporator temperature (2) high side condenser temperature (3) temperature difference in the cascade heat exchanger (4) low side condenser temperature (5) low side degree of subcooling (6) high side degree of subcooling (7) low side degree of superheating (8) high side degree of superheating. Results show that for minimum work system should operate at high temperature in low side evaporator, low temperature in high side condenser, low temperature difference in cascade condenser, high temperature in low side condenser and low degree of subcooling and superheating in both side. For maximum refrigeration effect system should operate at high temperature in low side evaporator, high temperature in high side condenser, high temperature difference in cascade condenser, low temperature in low side condenser and higher degree of subcooling in LT and HT side. For maximum coefficient of performance and exergetic efficiency, system should operate at high temperature in low side evaporator, low temperature in high side condenser, low temperature difference in cascade condenser, high temperature in low side condenser and higher degree of subcooling and superheating in low side of the system.

Keywords: Cascade refrigeration system, Genetic Algorithm, R507A, R23,

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2119
13210 Cloud Computing Support for Diagnosing Researches

Authors: A. Amirov, O. Gerget, V. Kochegurov

Abstract:

One of the main biomedical problem lies in detecting dependencies in semi structured data. Solution includes biomedical portal and algorithms (integral rating health criteria, multidimensional data visualization methods). Biomedical portal allows to process diagnostic and research data in parallel mode using Microsoft System Center 2012, Windows HPC Server cloud technologies. Service does not allow user to see internal calculations instead it provides practical interface. When data is sent for processing user may track status of task and will achieve results as soon as computation is completed. Service includes own algorithms and allows diagnosing and predicating medical cases. Approved methods are based on complex system entropy methods, algorithms for determining the energy patterns of development and trajectory models of biological systems and logical–probabilistic approach with the blurring of images.

Keywords: Biomedical portal, cloud computing, diagnostic and prognostic research, mathematical data analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1607
13209 Removal of Malachite Green from Aqueous Solution using Hydrilla verticillata -Optimization, Equilibrium and Kinetic Studies

Authors: R. Rajeshkannan, M. Rajasimman, N. Rajamohan

Abstract:

In this study, the sorption of Malachite green (MG) on Hydrilla verticillata biomass, a submerged aquatic plant, was investigated in a batch system. The effects of operating parameters such as temperature, adsorbent dosage, contact time, adsorbent size, and agitation speed on the sorption of Malachite green were analyzed using response surface methodology (RSM). The proposed quadratic model for central composite design (CCD) fitted very well to the experimental data that it could be used to navigate the design space according to ANOVA results. The optimum sorption conditions were determined as temperature - 43.5oC, adsorbent dosage - 0.26g, contact time - 200min, adsorbent size - 0.205mm (65mesh), and agitation speed - 230rpm. The Langmuir and Freundlich isotherm models were applied to the equilibrium data. The maximum monolayer coverage capacity of Hydrilla verticillata biomass for MG was found to be 91.97 mg/g at an initial pH 8.0 indicating that the optimum sorption initial pH. The external and intra particle diffusion models were also applied to sorption data of Hydrilla verticillata biomass with MG, and it was found that both the external diffusion as well as intra particle diffusion contributes to the actual sorption process. The pseudo-second order kinetic model described the MG sorption process with a good fitting.

Keywords: Response surface methodology, Hydrilla verticillata, malachite green, adsorption, central composite design

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1970
13208 Optimizing Performance of Tablet's Direct Compression Process Using Fuzzy Goal Programming

Authors: Abbas Al-Refaie

Abstract:

This paper aims at improving the performance of the tableting process using statistical quality control and fuzzy goal programming. The tableting process was studied. Statistical control tools were used to characterize the existing process for three critical responses including the averages of a tablet’s weight, hardness, and thickness. At initial process factor settings, the estimated process capability index values for the tablet’s averages of weight, hardness, and thickness were 0.58, 3.36, and 0.88, respectively. The L9 array was utilized to provide experimentation design. Fuzzy goal programming was then employed to find the combination of optimal factor settings. Optimization results showed that the process capability index values for a tablet’s averages of weight, hardness, and thickness were improved to 1.03, 4.42, and 1.42, respectively. Such improvements resulted in significant savings in quality and production costs.

Keywords: Fuzzy goal programming, control charts, process capability, tablet optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 989
13207 Research Trend Analysis – A Sample in the Field of Information Systems

Authors: Hei-Chia Wang, Wei-Pin Chiu

Abstract:

As research performance in academia is treated as one of indices for national competency, many countries devote much attention and resources to increasing their research performance. Understand the research trend is the basic step to improve the research performance. The goal of this research is to design an analysis system to evaluate research trends from analyzing data from different countries. In this paper, information system researches in Taiwan and other countries, including Asian countries and prominent countries represented by the Group of Eight (G8) is used as example. Our research found the trends are varied in different countries. Our research suggested that Taiwan-s scholars can pay more attention to interdisciplinary applications and try to increase their collaboration with other countries, in order to increase Taiwan's competency in the area of information science.

Keywords: Bibliometric analysis, research trend, scientometric analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1661