Search results for: information model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 10344

Search results for: information model

10194 Role-Governed Categorization and Category Learning as a Result from Structural Alignment: The RoleMap Model

Authors: Yolina A. Petrova, Georgi I. Petkov

Abstract:

The paper presents a symbolic model for category learning and categorization (called RoleMap). Unlike the other models which implement learning in a separate working mode, role-governed category learning and categorization emerge in RoleMap while it does its usual reasoning. The model is based on several basic mechanisms known as reflecting the sub-processes of analogy-making. It steps on the assumption that in their everyday life people constantly compare what they experience and what they know. Various commonalities between the incoming information (current experience) and the stored one (long-term memory) emerge from those comparisons. Some of those commonalities are considered to be highly important, and they are transformed into concepts for further use. This process denotes the category learning. When there is missing knowledge in the incoming information (i.e. the perceived object is still not recognized), the model makes anticipations about what is missing, based on the similar episodes from its long-term memory. Various such anticipations may emerge for different reasons. However, with time only one of them wins and is transformed into a category member. This process denotes the act of categorization.

Keywords: Categorization, category learning, role-governed category, analogy-making, cognitive modeling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 651
10193 Further Thoughtson a Sequential Life Testing Approach Using an Inverse Weibull Model

Authors: D. I. De Souza, G. P. Azevedo, D. R. Fonseca

Abstract:

In this paper we will develop further the sequential life test approach presented in a previous article by [1] using an underlying two parameter Inverse Weibull sampling distribution. The location parameter or minimum life will be considered equal to zero. Once again we will provide rules for making one of the three possible decisions as each observation becomes available; that is: accept the null hypothesis H0; reject the null hypothesis H0; or obtain additional information by making another observation. The product being analyzed is a new electronic component. There is little information available about the possible values the parameters of the corresponding Inverse Weibull underlying sampling distribution could have.To estimate the shape and the scale parameters of the underlying Inverse Weibull model we will use a maximum likelihood approach for censored failure data. A new example will further develop the proposed sequential life testing approach.

Keywords: Sequential Life Testing, Inverse Weibull Model, Maximum Likelihood Approach, Hypothesis Testing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1408
10192 Case-Based Reasoning: A Hybrid Classification Model Improved with an Expert's Knowledge for High-Dimensional Problems

Authors: Bruno Trstenjak, Dzenana Donko

Abstract:

Data mining and classification of objects is the process of data analysis, using various machine learning techniques, which is used today in various fields of research. This paper presents a concept of hybrid classification model improved with the expert knowledge. The hybrid model in its algorithm has integrated several machine learning techniques (Information Gain, K-means, and Case-Based Reasoning) and the expert’s knowledge into one. The knowledge of experts is used to determine the importance of features. The paper presents the model algorithm and the results of the case study in which the emphasis was put on achieving the maximum classification accuracy without reducing the number of features.

Keywords: Case based reasoning, classification, expert's knowledge, hybrid model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1407
10191 Degeneracy of MIS under the Conditions of Instability: A Mathematical Formulation

Authors: Nazar Younis, Raied Salman

Abstract:

It has been always observed that the effectiveness of MIS as a support tool for management decisions degenerate after time of implementation, despite the substantial investments being made. This is true for organizations at the initial stages of MIS implementations, manual or computerized. A survey of a sample of middle to top managers in business and government institutions was made. A large ratio indicates that the MIS has lost its impact on the day-to-day operations, and even the response lag time expands sometimes indefinitely. The data indicates an infant mortality phenomenon of the bathtub model. Reasons may be monotonous nature of MIS delivery, irrelevance, irreverence, timeliness, and lack of adequate detail. All those reasons collaborate to create a degree of degeneracy. We investigate and model as a bathtub model the phenomenon of MIS degeneracy that inflicts the MIS systems and renders it ineffective. A degeneracy index is developed to identify the status of the MIS system and possible remedies to prevent the onset of total collapse of the system to the point of being useless.

Keywords: MIS, management theory, information technology, information systems, IS, organizational environment, organizations, degeneracy, organizational change.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1571
10190 The Content Based Objective Metrics for Video Quality Evaluation

Authors: Michal Mardiak, Jaroslav Polec

Abstract:

In this paper we proposed comparison of four content based objective metrics with results of subjective tests from 80 video sequences. We also include two objective metrics VQM and SSIM to our comparison to serve as “reference” objective metrics because their pros and cons have already been published. Each of the video sequence was preprocessed by the region recognition algorithm and then the particular objective video quality metric were calculated i.e. mutual information, angular distance, moment of angle and normalized cross-correlation measure. The Pearson coefficient was calculated to express metrics relationship to accuracy of the model and the Spearman rank order correlation coefficient to represent the metrics relationship to monotonicity. The results show that model with the mutual information as objective metric provides best result and it is suitable for evaluating quality of video sequences.

Keywords: Objective quality metrics, mutual information, region recognition, content based metrics

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1493
10189 An Empirical Study of the Expectation- Perception Gap of I.S. Development

Authors: Linda, Sau-ling Lai

Abstract:

This paper adopts a notion of expectation-perception gap of systems users as information systems (IS) failure. Problems leading to the expectation-perception gap are identified and modelled as five interrelated discrepancies or gaps throughout the process of information systems development (ISD). It describes an empirical study on how systems developers and users perceive the size of each gap and the extent to which each problematic issue contributes to the gap. The key to achieving success in ISD is to keep the expectationperception gap closed by closing all 5 pertaining gaps. The gap model suggests that most factors in IS failure are related to organizational, cognitive and social aspects of information systems design. Organization requirement analysis, being the weakest link of IS development, is particularly worthy of investigation.

Keywords: Information Systems Development, Expectation- Perception Gap, Gap Analysis, Organization Analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1447
10188 A Model Predictive Control and Time Series Forecasting Framework for Supply Chain Management

Authors: Philip Doganis, Eleni Aggelogiannaki, Haralambos Sarimveis

Abstract:

Model Predictive Control has been previously applied to supply chain problems with promising results; however hitherto proposed systems possessed no information on future demand. A forecasting methodology will surely promote the efficiency of control actions by providing insight on the future. A complete supply chain management framework that is based on Model Predictive Control (MPC) and Time Series Forecasting will be presented in this paper. The proposed framework will be tested on industrial data in order to assess the efficiency of the method and the impact of forecast accuracy on overall control performance of the supply chain. To this end, forecasting methodologies with different characteristics will be implemented on test data to generate forecasts that will serve as input to the Model Predictive Control module.

Keywords: Forecasting, Model predictive control, production planning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1961
10187 A Mixed Integer Programming for Port Anzali Development Plan

Authors: Mahdieh Allahviranloo

Abstract:

This paper introduces a mixed integer programming model to find the optimum development plan for port Anzali. The model minimizes total system costs taking into account both port infrastructure costs and shipping costs. Due to the multipurpose function of the port, the model consists of 1020 decision variables and 2490 constraints. Results of the model determine the optimum number of berths that should be constructed in each period and for each type of cargo. In addition to, the results of sensitivity analysis on port operation quantity provide useful information for managers to choose the best scenario for port planning with the lowest investment risks. Despite all limitations-due to data availability-the model offers a straightforward decision tools to port planners aspiring to achieve optimum port planning steps.

Keywords: MILP, Multipurpose Terminal, Port Operation Optimization, Port Anzali.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1637
10186 Context Modeling and Reasoning Approach in Context-Aware Middleware for URC System

Authors: Chung-Seong Hong, Hyung-Sun Kim, Joonmyun Cho, Hyun Kyu Cho, Hyun-Chan Lee

Abstract:

To realize the vision of ubiquitous computing, it is important to develop a context-aware infrastructure which can help ubiquitous agents, services, and devices become aware of their contexts because such computational entities need to adapt themselves to changing situations. A context-aware infrastructure manages the context model representing contextual information and provides appropriate information. In this paper, we introduce Context-Aware Middleware for URC System (hereafter CAMUS) as a context-aware infrastructure for a network-based intelligent robot system and discuss the ontology-based context modeling and reasoning approach which is used in that infrastructure.

Keywords: CAMUS, Context-Aware, Context Model, Ontology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1904
10185 COVID_ICU_BERT: A Fine-tuned Language Model for COVID-19 Intensive Care Unit Clinical Notes

Authors: Shahad Nagoor, Lucy Hederman, Kevin Koidl, Annalina Caputo

Abstract:

Doctors’ notes reflect their impressions, attitudes, clinical sense, and opinions about patients’ conditions and progress, and other information that is essential for doctors’ daily clinical decisions. Despite their value, clinical notes are insufficiently researched within the language processing community. Automatically extracting information from unstructured text data is known to be a difficult task as opposed to dealing with structured information such as physiological vital signs, images and laboratory results. The aim of this research is to investigate how Natural Language Processing (NLP) techniques and machine learning techniques applied to clinician notes can assist in doctors’ decision making in Intensive Care Unit (ICU) for coronavirus disease 2019 (COVID-19) patients. The hypothesis is that clinical outcomes like survival or mortality can be useful to influence the judgement of clinical sentiment in ICU clinical notes. This paper presents two contributions: first, we introduce COVID_ICU_BERT, a fine-tuned version of a clinical transformer model that can reliably predict clinical sentiment for notes of COVID patients in ICU. We train the model on clinical notes for COVID-19 patients, ones not previously seen by Bio_ClinicalBERT or Bio_Discharge_Summary_BERT. The model which was based on Bio_ClinicalBERT achieves higher predictive accuracy than the one based on Bio_Discharge_Summary_BERT (Acc 93.33%, AUC 0.98, and Precision 0.96). Second, we perform data augmentation using clinical contextual word embedding that is based on a pre-trained clinical model to balance the samples in each class in the data (survived vs. deceased patients). Data augmentation improves the accuracy of prediction slightly (Acc 96.67%, AUC 0.98, and Precision 0.92).

Keywords: BERT fine-tuning, clinical sentiment, COVID-19, data augmentation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 230
10184 Using Perspective Schemata to Model the ETL Process

Authors: Valeria M. Pequeno, Joao Carlos G. M. Pires

Abstract:

Data Warehouses (DWs) are repositories which contain the unified history of an enterprise for decision support. The data must be Extracted from information sources, Transformed and integrated to be Loaded (ETL) into the DW, using ETL tools. These tools focus on data movement, where the models are only used as a means to this aim. Under a conceptual viewpoint, the authors want to innovate the ETL process in two ways: 1) to make clear compatibility between models in a declarative fashion, using correspondence assertions and 2) to identify the instances of different sources that represent the same entity in the real-world. This paper presents the overview of the proposed framework to model the ETL process, which is based on the use of a reference model and perspective schemata. This approach provides the designer with a better understanding of the semantic associated with the ETL process.

Keywords: conceptual data model, correspondence assertions, data warehouse, data integration, ETL process, object relational database.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1498
10183 A Novel Web Metric for the Evaluation of Internet Trends

Authors: Radek Malinský, Ivan Jelínek

Abstract:

Web 2.0 (social networking, blogging and online forums) can serve as a data source for social science research because it contains vast amount of information from many different users. The volume of that information has been growing at a very high rate and becoming a network of heterogeneous data; this makes things difficult to find and is therefore not almost useful. We have proposed a novel theoretical model for gathering and processing data from Web 2.0, which would reflect semantic content of web pages in better way. This article deals with the analysis part of the model and its usage for content analysis of blogs. The introductory part of the article describes methodology for the gathering and processing data from blogs. The next part of the article is focused on the evaluation and content analysis of blogs, which write about specific trend.

Keywords: Blog, Sentiment Analysis, Web 2.0, Webometrics

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3529
10182 Application of Company Financial Crisis Early Warning Model- Use of “Financial Reference Database“

Authors: Chiung-ying Lee, Chia-hua Chang

Abstract:

In July 1, 2007, Taiwan Stock Exchange (TWSE) on market observation post system (MOPS) adds a new "Financial reference database" for investors to do investment reference. This database as a warning to public offering companies listed on the public financial information and it original within eight targets. In this paper, this database provided by the indicators for the application of company financial crisis early warning model verify that the database provided by the indicator forecast for the financial crisis, whether or not companies have a high accuracy rate as opposed to domestic and foreign scholars have positive results. There is use of Logistic Regression Model application of the financial early warning model, in which no joined back-conditions is the first model, joined it in is the second model, has been taken occurred in the financial crisis of companies to research samples and then business took place before the financial crisis point with T-1 and T-2 sample data to do positive analysis. The results show that this database provided the debt ratio and net per share for the best forecast variables.

Keywords: Financial reference database, Financial early warning model, Logistic Regression.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1414
10181 Organizational Management Model based on Knowledge Management, Talent Management and Technology Management Framework “Gomak“

Authors: Nieto Bernal W., Luna Amaya C.

Abstract:

This paper aims to present a framework for the organizational knowledge management, which seeks to deploy a standardized structure for the integrated management of knowledge is a common language based on domains, processes and global indicators inspired by the COBIT framework 5 (ISACA, 2012), which supports the integration of three technologies, enterprise information architecture (EIA), the business process modeling (BPM) and service-oriented architecture (SOA). The Gomak Framework is a management platform that seeks to integrate the information technology infrastructure, the structure of applications, information infrastructure, and business logic and business model to support a sound strategy of organizational knowledge management, low process-based approach and concurrent engineering. Concurrent engineering (CE) is a systematic approach to integrated product development that respond to customer expectations, involving all perspectives in parallel, from the beginning of the product life cycle. (European Space Agency, 2000).

Keywords: Business Process Modeling, Enterprise Information Architecture, Government and Knowledge Management, Service Oriented Architecture, Process Management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1833
10180 Evaluation of Risk Attributes Driven by Periodically Changing System Functionality

Authors: Dariusz Dymek, Leszek Kotulski

Abstract:

Modeling of the distributed systems allows us to represent the whole its functionality. The working system instance rarely fulfils the whole functionality represented by model; usually some parts of this functionality should be accessible periodically. The reporting system based on the Data Warehouse concept seams to be an intuitive example of the system that some of its functionality is required only from time to time. Analyzing an enterprise risk associated with the periodical change of the system functionality, we should consider not only the inaccessibility of the components (object) but also their functions (methods), and the impact of such a situation on the system functionality from the business point of view. In the paper we suggest that the risk attributes should be estimated from risk attributes specified at the requirements level (Use Case in the UML model) on the base of the information about the structure of the model (presented at other levels of the UML model). We argue that it is desirable to consider the influence of periodical changes in requirements on the enterprise risk estimation. Finally, the proposition of such a solution basing on the UML system model is presented.

Keywords: Risk assessing, software maintenance, UML, graph grammars.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1372
10179 Performance Appraisal System using Multifactorial Evaluation Model

Authors: C. C. Yee, Y.Y.Chen

Abstract:

Performance appraisal of employee is important in managing the human resource of an organization. With the change towards knowledge-based capitalism, maintaining talented knowledge workers is critical. However, management classification of “outstanding", “poor" and “average" performance may not be an easy decision. Besides that, superior might also tend to judge the work performance of their subordinates informally and arbitrarily especially without the existence of a system of appraisal. In this paper, we propose a performance appraisal system using multifactorial evaluation model in dealing with appraisal grades which are often express vaguely in linguistic terms. The proposed model is for evaluating staff performance based on specific performance appraisal criteria. The project was collaboration with one of the Information and Communication Technology company in Malaysia with reference to its performance appraisal process.

Keywords: Multifactorial Evaluation Model, performance appraisal system, decision support system.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4250
10178 A Data Hiding Model with High Security Features Combining Finite State Machines and PMM method

Authors: Souvik Bhattacharyya, Gautam Sanyal

Abstract:

Recent years have witnessed the rapid development of the Internet and telecommunication techniques. Information security is becoming more and more important. Applications such as covert communication, copyright protection, etc, stimulate the research of information hiding techniques. Traditionally, encryption is used to realize the communication security. However, important information is not protected once decoded. Steganography is the art and science of communicating in a way which hides the existence of the communication. Important information is firstly hidden in a host data, such as digital image, video or audio, etc, and then transmitted secretly to the receiver.In this paper a data hiding model with high security features combining both cryptography using finite state sequential machine and image based steganography technique for communicating information more securely between two locations is proposed. The authors incorporated the idea of secret key for authentication at both ends in order to achieve high level of security. Before the embedding operation the secret information has been encrypted with the help of finite-state sequential machine and segmented in different parts. The cover image is also segmented in different objects through normalized cut.Each part of the encoded secret information has been embedded with the help of a novel image steganographic method (PMM) on different cuts of the cover image to form different stego objects. Finally stego image is formed by combining different stego objects and transmit to the receiver side. At the receiving end different opposite processes should run to get the back the original secret message.

Keywords: Cover Image, Finite state sequential machine, Melaymachine, Pixel Mapping Method (PMM), Stego Image, NCUT.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2248
10177 Philosophy of Education: The Challenges of Globalization and Innovation in the Information Society

Authors: Shattyk Aliyev, Zhakypbek Altayev, Zuchra Ismagambetova, Yerkin Massanov

Abstract:

Information society is an absolutely new public formation at which the infrastructure and the social relations correspond to the socialized essence of «information genotype» mankind. Information society is a natural social environment which allows the person to open completely the information nature, to use intelligence for joint creation with other people of new information on the basis of knowledge earlier saved up by previous generations.

Keywords: Information society, Philosophy, Education, Globalization and innovation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1944
10176 3D Model Retrieval based on Normal Vector Interpolation Method

Authors: Ami Kim, Oubong Gwun, Juwhan Song

Abstract:

In this paper, we proposed the distribution of mesh normal vector direction as a feature descriptor of a 3D model. A normal vector shows the entire shape of a model well. The distribution of normal vectors was sampled in proportion to each polygon's area so that the information on the surface with less surface area may be less reflected on composing a feature descriptor in order to enhance retrieval performance. At the analysis result of ANMRR, the enhancement of approx. 12.4%~34.7% compared to the existing method has also been indicated.

Keywords: Interpolated Normal Vector, Feature Descriptor, 3DModel Retrieval.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1463
10175 Hippocampus Segmentation using a Local Prior Model on its Boundary

Authors: Dimitrios Zarpalas, Anastasios Zafeiropoulos, Petros Daras, Nicos Maglaveras

Abstract:

Segmentation techniques based on Active Contour Models have been strongly benefited from the use of prior information during their evolution. Shape prior information is captured from a training set and is introduced in the optimization procedure to restrict the evolution into allowable shapes. In this way, the evolution converges onto regions even with weak boundaries. Although significant effort has been devoted on different ways of capturing and analyzing prior information, very little thought has been devoted on the way of combining image information with prior information. This paper focuses on a more natural way of incorporating the prior information in the level set framework. For proof of concept the method is applied on hippocampus segmentation in T1-MR images. Hippocampus segmentation is a very challenging task, due to the multivariate surrounding region and the missing boundary with the neighboring amygdala, whose intensities are identical. The proposed method, mimics the human segmentation way and thus shows enhancements in the segmentation accuracy.

Keywords: Medical imaging & processing, Brain MRI segmentation, hippocampus segmentation, hippocampus-amygdala missingboundary, weak boundary segmentation, region based segmentation, prior information, local weighting scheme in level sets, spatialdistribution of labels, gradient distribution on boundary.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1739
10174 Adaptive Integral Backstepping Motion Control for Inverted Pendulum

Authors: Ö. Tolga Altınöz

Abstract:

The adaptive backstepping controller for inverted pendulum is designed by using the general motion control model. Backstepping is a novel nonlinear control technique based on the Lyapunov design approach, used when higher derivatives of parameter estimation appear. For easy parameter adaptation, the mathematical model of the inverted pendulum converted into the motion control model. This conversion is performed by taking functions of unknown parameters and dynamics of the system. By using motion control model equations, inverted pendulum is simulated without any information about not only parameters but also measurable dynamics. Also these results are compare with the adaptive backstepping controller which extended with integral action that given from [1].

Keywords: Adaptive backstepping, inverted pendulum, nonlinear adaptive control.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3475
10173 Application of Generalized Stochastic Petri Nets(GSPN) in Modeling and Evaluating a Resource Sharing Flexible Manufacturing System

Authors: Aryanejad Mir Bahador Goli, Zahra Honarmand Shah Zileh

Abstract:

In most study fields, a phenomenon may not be studied directly but it will be examined indirectly by phenomenon model. Making an accurate model of system, there is attained new information from modeled phenomenon without any charge, danger, etc... there have been developed more solutions for describing and analyzing the recent complicated systems but few of them have analyzed the performance in the range of system description. Petri nets are of limited solutions which may make such union. Petri nets are being applied in problems related to modeling and designing the systems. Theory of Petri nets allow a system to model mathematically by a Petri net and analyzing the Petri net can then determine main information of modeled system-s structure and dynamic. This information can be used for assessing the performance of systems and suggesting corrections in the system. In this paper, beside the introduction of Petri nets, a real case study will be studied in order to show the application of generalized stochastic Petri nets in modeling a resource sharing production system and evaluating the efficiency of its machines and robots. The modeling tool used here is SHARP software which calculates specific indicators helping to make decision.

Keywords: Flexible manufacturing system, generalizedstochastic Petri nets, Markov chain, performance evaluation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1889
10172 Features for Measuring Credibility on Facebook Information

Authors: Kanda Runapongsa Saikaew, Chaluemwut Noyunsan

Abstract:

Nowadays social media information, such as news, links, images, or VDOs, is shared extensively. However, the effectiveness of disseminating information through social media lacks in quality: less fact checking, more biases, and several rumors. Many researchers have investigated about credibility on Twitter, but there is no the research report about credibility information on Facebook. This paper proposes features for measuring credibility on Facebook information. We developed the system for credibility on Facebook. First, we have developed FB credibility evaluator for measuring credibility of each post by manual human’s labelling. We then collected the training data for creating a model using Support Vector Machine (SVM). Secondly, we developed a chrome extension of FB credibility for Facebook users to evaluate the credibility of each post. Based on the usage analysis of our FB credibility chrome extension, about 81% of users’ responses agree with suggested credibility automatically computed by the proposed system.

Keywords: Facebook, social media, credibility measurement.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3655
10171 Thermomechanical Damage Modeling of F114 Carbon Steel

Authors: A. El Amri, M. El Yakhloufi Haddou, A. Khamlichi

Abstract:

The numerical simulation based on the Finite Element Method (FEM) is widely used in academic institutes and in the industry. It is a useful tool to predict many phenomena present in the classical manufacturing forming processes such as fracture. But, the results of such numerical model depend strongly on the parameters of the constitutive behavior model. The influences of thermal and mechanical loads cause damage. The temperature and strain rate dependent materials’ properties and their modelling are discussed. A Johnson-Cook Model of damage has been selected for the numerical simulations. Virtual software called the ABAQUS 6.11 is used for finite element analysis. This model was introduced in order to give information concerning crack initiation during thermal and mechanical loads.

Keywords: Thermomechanical fatigue, failure, numerical simulation, fracture, damages.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1482
10170 A New Distribution and Application on the Lifetime Data

Authors: Gamze Ozel, Selen Cakmakyapan

Abstract:

We introduce a new model called the Marshall-Olkin Rayleigh distribution which extends the Rayleigh distribution using Marshall-Olkin transformation and has increasing and decreasing shapes for the hazard rate function. Various structural properties of the new distribution are derived including explicit expressions for the moments, generating and quantile function, some entropy measures, and order statistics are presented. The model parameters are estimated by the method of maximum likelihood and the observed information matrix is determined. The potentiality of the new model is illustrated by means of a simulation study. 

Keywords: Marshall-Olkin distribution, Rayleigh distribution, estimation, maximum likelihood.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1377
10169 Classifying Students for E-Learning in Information Technology Course Using ANN

Authors: S. Areerachakul, N. Ployong, S. Na Songkla

Abstract:

This research’s objective is to select the model with most accurate value by using Neural Network Technique as a way to filter potential students who enroll in IT course by Electronic learning at Suan Suanadha Rajabhat University. It is designed to help students selecting the appropriate courses by themselves. The result showed that the most accurate model was 100 Folds Cross-validation which had 73.58% points of accuracy.

Keywords: Artificial neural network, classification, students.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1489
10168 Methodology for Obtaining Static Alignment Model

Authors: Lely A. Luengas, Pedro R. Vizcaya, Giovanni Sánchez

Abstract:

In this paper, a methodology is presented to obtain the Static Alignment Model for any transtibial amputee person. The proposed methodology starts from experimental data collected on the Hospital Militar Central, Bogotá, Colombia. The effects of transtibial prosthesis malalignment on amputees were measured in terms of joint angles, center of pressure (COP) and weight distribution. Some statistical tools are used to obtain the model parameters. Mathematical predictive models of prosthetic alignment were created. The proposed models are validated in amputees and finding promising results for the prosthesis Static Alignment. Static alignment process is unique to each subject; nevertheless the proposed methodology can be used in each transtibial amputee.

Keywords: Information theory, prediction model, prosthetic alignment, transtibial prosthesis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 920
10167 Modelling of Designing a Conceptual Schema for Multimodal Freight Transportation Information System

Authors: Gia Surguladze, Lily Petriashvili, Nino Topuria, Giorgi Surguladze

Abstract:

Modelling of building processes of a multimodal freight transportation support information system is discussed based on modern CASE technologies. Functional efficiencies of ports in the eastern part of the Black Sea are analyzed taking into account their ecological, seasonal, resource usage parameters. By resources, we mean capacities of berths, cranes, automotive transport, as well as work crews and neighbouring airports. For the purpose of designing database of computer support system for Managerial (Logistics) function, using Object-Role Modeling (ORM) tool (NORMA–Natural ORM Architecture) is proposed, after which Entity Relationship Model (ERM) is generated in automated process. Software is developed based on Process-Oriented and Service-Oriented architecture, in Visual Studio.NET environment.

Keywords: Seaport resources, business-processes, multimodal transportation, CASE technology, object-role model, entity relationship model, SOA.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1975
10166 A Generalised Relational Data Model

Authors: Georgia Garani

Abstract:

A generalised relational data model is formalised for the representation of data with nested structure of arbitrary depth. A recursive algebra for the proposed model is presented. All the operations are formally defined. The proposed model is proved to be a superset of the conventional relational model (CRM). The functionality and validity of the model is shown by a prototype implementation that has been undertaken in the functional programming language Miranda.

Keywords: nested relations, recursive algebra, recursive nested operations, relational data model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1546
10165 Process-Oriented Learning Requirements for Employees and for Organizations

Authors: Richard Pircher, Lukas Zenk, Hanna Risku

Abstract:

Using activity theory, organisational theory and didactics as theoretical foundations, a comprehensive model of the organisational dimensions relevant for learning and knowledge transfer will be developed. In a second step, a Learning Assessment Guideline will be elaborated. This guideline will be designed to permit a targeted analysis of organisations to identify the status quo in those areas crucial to the implementation of learning and knowledge transfer. In addition, this self-analysis tool will enable learning managers to select adequate didactic models for e- and blended learning. As part of the European Integrated Project "Process-oriented Learning and Information Exchange" (PROLIX), this model of organisational prerequisites for learning and knowledge transfer will be empirically tested in four profit and non-profit organisations in Great Britain, Germany and France (to be finalized in autumn 2006). The findings concern not only the capability of the model of organisational dimensions, but also the predominant perceptions of and obstacles to learning in organisations.

Keywords: Activity theory, knowledge management organisational theory, "Process-oriented Learning and Information Exchange" (PROLIX).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1732