Search results for: panel data regression models
7759 The Relationship between Conceptual Organizational Culture and the Level of Tolerance in Employees
Authors: M. Sadoughi, R. Ehsani
Abstract:
The aim of the present study is examining the relationship between conceptual organizational culture and the level of tolerance in employees of Islamic Azad University of Shahre Ghods. This research is a correlational and analytic-descriptive one. The samples included 144 individuals. A 24-item standard questionnaire of organizational culture by Cameron and Queen was used in this study. This questionnaire has six criteria and each criterion includes four items that each item indicates one cultural dimension. Reliability coefficient of this questionnaire was normed using Cronbach's alpha of 0.91. Also, the 25-item questionnaire of tolerance by Conor and Davidson was used. This questionnaire is in a five-degree Likert scale form. It has seven criteria and is designed to measure the power of coping with pressure and threat. It has the needed content reliability and its reliability coefficient is normed using Cronbach's alpha of 0.87. Data were analyzed using Pearson correlation coefficient and multivariable regression. The results showed among various dimensions of organizational culture, there is a positive significant relationship between three dimensions (family, adhocracy, bureaucracy) and tolerance, there is a negative significant relationship between dimension of market and tolerance and components of organizational culture have the power of prediction and explaining the tolerance. In this explanation, the component of family is the most effective and the best predictor of tolerance.
Keywords: Adhocracy, bureaucracy, organizational culture, tolerance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10327758 Are XBRL-based Financial Reports Better than Non-XBRL Reports? A Quality Assessment
Authors: Zhenkun Wang, Simon S. Gao
Abstract:
Using a scoring system, this paper provides a comparative assessment of the quality of data between XBRL formatted financial reports and non-XBRL financial reports. It shows a major improvement in the quality of data of XBRL formatted financial reports. Although XBRL formatted financial reports do not show much advantage in the quality at the beginning, XBRL financial reports lately display a large improvement in the quality of data in almost all aspects. With the improved XBRL web data managing, presentation and analysis applications, XBRL formatted financial reports have a much better accessibility, are more accurate and better in timeliness.Keywords: Data Quality; Financial Report; Information; XBRL
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25667757 Dynamic Analysis of Composite Doubly Curved Panels with Variable Thickness
Authors: I. Algul, G. Akgun, H. Kurtaran
Abstract:
Dynamic analysis of composite doubly curved panels with variable thickness subjected to different pulse types using Generalized Differential Quadrature method (GDQ) is presented in this study. Panels with variable thickness are used in the construction of aerospace and marine industry. Giving variable thickness to panels can allow the designer to get optimum structural efficiency. For this reason, estimating the response of variable thickness panels is very important to design more reliable structures under dynamic loads. Dynamic equations for composite panels with variable thickness are obtained using virtual work principle. Partial derivatives in the equation of motion are expressed with GDQ and Newmark average acceleration scheme is used for temporal discretization. Several examples are used to highlight the effectiveness of the proposed method. Results are compared with finite element method. Effects of taper ratios, boundary conditions and loading type on the response of composite panel are investigated.
Keywords: Generalized differential quadrature method, doubly curved panels, laminated composite materials, small displacement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9417756 Video Data Mining based on Information Fusion for Tamper Detection
Authors: Girija Chetty, Renuka Biswas
Abstract:
In this paper, we propose novel algorithmic models based on information fusion and feature transformation in crossmodal subspace for different types of residue features extracted from several intra-frame and inter-frame pixel sub-blocks in video sequences for detecting digital video tampering or forgery. An evaluation of proposed residue features – the noise residue features and the quantization features, their transformation in cross-modal subspace, and their multimodal fusion, for emulated copy-move tamper scenario shows a significant improvement in tamper detection accuracy as compared to single mode features without transformation in cross-modal subspace.Keywords: image tamper detection, digital forensics, correlation features image fusion
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19007755 Wireless Transmission of Big Data Using Novel Secure Algorithm
Authors: K. Thiagarajan, K. Saranya, A. Veeraiah, B. Sudha
Abstract:
This paper presents a novel algorithm for secure, reliable and flexible transmission of big data in two hop wireless networks using cooperative jamming scheme. Two hop wireless networks consist of source, relay and destination nodes. Big data has to transmit from source to relay and from relay to destination by deploying security in physical layer. Cooperative jamming scheme determines transmission of big data in more secure manner by protecting it from eavesdroppers and malicious nodes of unknown location. The novel algorithm that ensures secure and energy balance transmission of big data, includes selection of data transmitting region, segmenting the selected region, determining probability ratio for each node (capture node, non-capture and eavesdropper node) in every segment, evaluating the probability using binary based evaluation. If it is secure transmission resume with the two- hop transmission of big data, otherwise prevent the attackers by cooperative jamming scheme and transmit the data in two-hop transmission.Keywords: Big data, cooperative jamming, energy balance, physical layer, two-hop transmission, wireless security.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21807754 Using Radial Basis Function Neural Networks to Calibrate Water Quality Model
Authors: Lihui Ma, Kunlun Xin, Suiqing Liu
Abstract:
Modern managements of water distribution system (WDS) need water quality models that are able to accurately predict the dynamics of water quality variations within the distribution system environment. Before water quality models can be applied to solve system problems, they should be calibrated. Although former researchers use GA solver to calibrate relative parameters, it is difficult to apply on the large-scale or medium-scale real system for long computational time. In this paper a new method is designed which combines both macro and detailed model to optimize the water quality parameters. This new combinational algorithm uses radial basis function (RBF) metamodeling as a surrogate to be optimized for the purpose of decreasing the times of time-consuming water quality simulation and can realize rapidly the calibration of pipe wall reaction coefficients of chlorine model of large-scaled WDS. After two cases study this method is testified to be more efficient and promising, and deserve to generalize in the future.Keywords: Metamodeling, model calibration, radial basisfunction, water distribution system, water quality model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20227753 Logistics Model for Improving Quality in Railway Transport
Authors: Eva Nedeliakova, Juraj Camaj, Jaroslav Masek
Abstract:
This contribution is focused on the methodology for identifying levels of quality and improving quality through new logistics model in railway transport. It is oriented on the application of dynamic quality models, which represent an innovative method of evaluation quality services. Through this conception, time factor, expected, and perceived quality in each moment of the transportation process within logistics chain can be taken into account. Various models describe the improvement of the quality which emphases the time factor throughout the whole transportation logistics chain. Quality of services in railway transport can be determined by the existing level of service quality, by detecting the causes of dissatisfaction employees but also customers, to uncover strengths and weaknesses. This new logistics model is able to recognize critical processes in logistic chain. It includes service quality rating that must respect its specific properties, which are unrepeatability, impalpability, their use right at the time they are provided and particularly changeability, which is significant factor in the conditions of rail transport as well. These peculiarities influence the quality of service regarding the constantly increasing requirements and that result in new ways of finding progressive attitudes towards the service quality rating.Keywords: Logistics model, quality, railway transport.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18287752 Numerical Simulation of Axially Loaded to Failure Large Diameter Bored Pile
Authors: M. Ezzat, Y. Zaghloul, T. Sorour, A. Hefny, M. Eid
Abstract:
Ultimate capacity of large diameter bored piles is usually determined from pile loading tests as recommended by several international codes and foundation design standards. However, loading of this type of piles till achieving apparent failure is practically seldom. In this paper, numerical analyses are carried out to simulate load test of a large diameter bored pile performed at the location of Alzey highway bridge project (Germany). Test results of pile load settlement relationship till failure as well as results of the base and shaft resistances are available. Apparent failure was indicated in this test by the significant increase of the induced settlement during the last load increment applied on the pile head. Measurements of this pile load test are used to assess the quality of the numerical models investigated. Three different material soil models are implemented in the analyses: Mohr coulomb (MC), Soft soil (SS), and Modified Mohr coulomb (MMC). Very good agreement is obtained between the field measured settlement and the calculated settlement using the MMC model. Results of analysis showed also that the MMC constitutive model is superior to MC, and SS models in predicting the ultimate base and shaft resistances of the large diameter bored pile. After calibrating the numerical model, behavior of large diameter bored piles under axial loads is discussed and the formation of the plastic zone around the pile is explored. Results obtained showed that the plastic zone below the base of the pile at failure extended laterally to about four times the pile diameter and vertically to about three times the pile diameter.
Keywords: Ultimate capacity, large diameter bored piles, plastic zone, failure, pile load test.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9197751 Study of Efficiency and Capability LZW++ Technique in Data Compression
Authors: Yusof. Mohd Kamir, Mat Deris. Mohd Sufian, Abidin. Ahmad Faisal Amri
Abstract:
The purpose of this paper is to show efficiency and capability LZWµ in data compression. The LZWµ technique is enhancement from existing LZW technique. The modification the existing LZW is needed to produce LZWµ technique. LZW read one by one character at one time. Differ with LZWµ technique, where the LZWµ read three characters at one time. This paper focuses on data compression and tested efficiency and capability LZWµ by different data format such as doc type, pdf type and text type. Several experiments have been done by different types of data format. The results shows LZWµ technique is better compared to existing LZW technique in term of file size.
Keywords: Data Compression, Huffman Encoding, LZW, LZWµ, RLL, Size.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20897750 When Explanations “Cause“ Error: A Look at Representations and Compressions
Authors: Michael Lissack
Abstract:
We depend upon explanation in order to “make sense" out of our world. And, making sense is all the more important when dealing with change. But, what happens if our explanations are wrong? This question is examined with respect to two types of explanatory model. Models based on labels and categories we shall refer to as “representations." More complex models involving stories, multiple algorithms, rules of thumb, questions, ambiguity we shall refer to as “compressions." Both compressions and representations are reductions. But representations are far more reductive than compressions. Representations can be treated as a set of defined meanings – coherence with regard to a representation is the degree of fidelity between the item in question and the definition of the representation, of the label. By contrast, compressions contain enough degrees of freedom and ambiguity to allow us to make internal predictions so that we may determine our potential actions in the possibility space. Compressions are explanatory via mechanism. Representations are explanatory via category. Managers are often confusing their evocation of a representation (category inclusion) as the creation of a context of compression (description of mechanism). When this type of explanatory error occurs, more errors follow. In the drive for efficiency such substitutions are all too often proclaimed – at the manager-s peril..Keywords: Coherence, Emergence, Reduction, Model
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12397749 Impact of Stack Caches: Locality Awareness and Cost Effectiveness
Authors: Abdulrahman K. Alshegaifi, Chun-Hsi Huang
Abstract:
Treating data based on its location in memory has received much attention in recent years due to its different properties, which offer important aspects for cache utilization. Stack data and non-stack data may interfere with each other’s locality in the data cache. One of the important aspects of stack data is that it has high spatial and temporal locality. In this work, we simulate non-unified cache design that split data cache into stack and non-stack caches in order to maintain stack data and non-stack data separate in different caches. We observe that the overall hit rate of non-unified cache design is sensitive to the size of non-stack cache. Then, we investigate the appropriate size and associativity for stack cache to achieve high hit ratio especially when over 99% of accesses are directed to stack cache. The result shows that on average more than 99% of stack cache accuracy is achieved by using 2KB of capacity and 1-way associativity. Further, we analyze the improvement in hit rate when adding small, fixed, size of stack cache at level1 to unified cache architecture. The result shows that the overall hit rate of unified cache design with adding 1KB of stack cache is improved by approximately, on average, 3.9% for Rijndael benchmark. The stack cache is simulated by using SimpleScalar toolset.
Keywords: Hit rate, Locality of program, Stack cache, and Stack data.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15087748 Measurement and Prediction of Speed of Sound in Petroleum Fluids
Authors: S. Ghafoori, A. Al-Harbi, B. Al-Ajmi, A. Al-Shaalan, A. Al-Ajmi, M. Ali Juma
Abstract:
Seismic methods play an important role in the exploration for hydrocarbon reservoirs. However, the success of the method depends strongly on the reliability of the measured or predicted information regarding the velocity of sound in the media. Speed of sound has been used to study the thermodynamic properties of fluids. In this study, experimental data are reported and analyzed on the speed of sound in toluene and octane binary mixture. Three-factor three-level Box-Benhkam design is used to determine the significance of each factor, the synergetic effects of the factors, and the most significant factors on speed of sound. The developed mathematical model and statistical analysis provided a critical analysis of the simultaneous interactive effects of the independent variables indicating that the developed quadratic models were highly accurate and predictive.
Keywords: Experimental design, octane, speed of sound, toluene.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14137747 Factors Affecting the Wages of Native Workers in Thailand's Construction Industry
Authors: C. Noknoi, W. Boripunt, K. Boomid, S. Suwitphanwong
Abstract:
This research studies the factors influencing the wages of native workers in Thailand's construction industry. The sample used comprised some 156 native construction workers from Songkhla Province, Thailand. The utilized research instrument was a questionnaire, with the data being analyzed according to frequency, percentage, and regression analysis. The results revealed that in general, native Thai construction workers are generally married males aged between 26 and 37 years old. They typically have four to six years of education, are employed as laborers with an average salary of 4,000–9,200 baht per month, and have fewer than five years of work experience. Most Thai workers work five days a week. Each establishment typically has 10–30 employees, with fewer than 10 of these being migrant workers in general. Most Thai workers are at a 20% to 40% risk from work, and they have never changed employer. The average wage of Thai workers was found to be 10,843.03 baht per month with a standard deviation of 4,898.31 baht per month. Hypothesis testing revealed that position, work experience, and the number of times they had switched employer were the factors most affecting the wages of native Thai construction workers. These three factors alone explain the salaries of Thai construction workers at 51.9%.
Keywords: Construction industry, native workers, Thailand, wages.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9187746 Towards a New Methodology for Developing Web-Based Systems
Authors: Omer Ishag Eldai, Ahmed Hassan M. H. Ali, S. Raviraja
Abstract:
Web-based systems have become increasingly important due to the fact that the Internet and the World Wide Web have become ubiquitous, surpassing all other technological developments in our history. The Internet and especially companies websites has rapidly evolved in their scope and extent of use, from being a little more than fixed advertising material, i.e. a "web presences", which had no particular influence for the company's business, to being one of the most essential parts of the company's core business. Traditional software engineering approaches with process models such as, for example, CMM and Waterfall models, do not work very well since web system development differs from traditional development. The development differs in several ways, for example, there is a large gap between traditional software engineering designs and concepts and the low-level implementation model, many of the web based system development activities are business oriented (for example web application are sales-oriented, web application and intranets are content-oriented) and not engineering-oriented. This paper aims to introduce Increment Iterative extreme Programming (IIXP) methodology for developing web based systems. In difference to the other existence methodologies, this methodology is combination of different traditional and modern software engineering and web engineering principles.Keywords: Web based systems, Web engineering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18527745 Effect of Blanching on the Quality of Microwave Vacuum Dried Dill (Anethum graveolens L.)
Authors: Evita Straumite, Zanda Kruma, Ruta Galoburda, Kaiva Saulite
Abstract:
Dill (Anethum graveolens L.) is a popular herb used in many regions, including Baltic countries. Dill is widely used for flavoring foods and beverages due to its pleasant spicy aroma. The aim of this work was to determine the best blanching method for processing of dill prior to microwave vacuum drying based on sensory properties, color and volatile compounds in dried product. Two blanching mediums were used – water and steam, and for part of samples microwave pretreatment was additionally used. Evaluation of dried dill volatile aroma compounds, color changes and sensory attributes was performed. Results showed that blanching significantly influences the quality of dried dill. After evaluation of volatile aroma compounds, color and sensory properties of microwave vacuum dried dill, as the best method for dill pretreatment was established blanching at 90 °C for 30 s.Keywords: dried dill, sensory panel, sensory properties, aroma compounds, color
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21787744 A Kinetic Study on the Adsorption of Cd(II) and Zn(II) Ions from Aqueous Solutions on Zeolite NaA
Authors: Dimitar Georgiev, Bogdan Bogdanov, Irena Markovska, Yancho Hristov, Dencho Stanev
Abstract:
The present paper reports the removal of Cd(II) and Zn(II) ions using synthetic Zeolit NaA. The adsorption capacity of the sorbent (Zeolite NaA) strongly depends on simultaneous or not simultaneous (concurrent) presence of Cd(II) and Zn(II) in the sorbate. When Cd(II) and Zn(II) are present simultaneously (concurrently) in the sorbate, Zn(II) ions were sorbed at higher rate. Equilibrium data fitted Langmuir, Freundlich and Tempkin isotherms well. The applicability of the isotherm equation to describe the adsorption process was judged by the correlation coefficients R2. The Langmuir model yielded the best fit with R2 values equal to or higher than 0.970, as compared to the Freundlich and Tempkin models. The fact that 1/n values range from 0.322 to 0.755 indicates that the adsorption of Cd(II) and Zn(II) ions from aqueous solutions also favored by the Freundlich model.Keywords: Adsorption, adsorption capacity, kinetic sorption, Zeolite NaA
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14727743 Effect of Non Uniformity Factors and Assignment Factors on Errors in Charge Simulation Method with Point Charge Model
Authors: Gururaj S Punekar, N K Kishore Senior, H S Y Shastry
Abstract:
Charge Simulation Method (CSM) is one of the very widely used numerical field computation technique in High Voltage (HV) engineering. The high voltage fields of varying non uniformities are encountered in practice. CSM programs being case specific, the simulation accuracies heavily depend on the user (programmers) experience. Here is an effort to understand CSM errors and evolve some guidelines to setup accurate CSM models, relating non uniformities with assignment factors. The results are for the six-point-charge model of sphere-plane gap geometry. Using genetic algorithm (GA) as tool, optimum assignment factors at different non uniformity factors for this model have been evaluated and analyzed. It is shown that the symmetrically placed six-point-charge models can be good enough to set up CSM programs with potential errors less than 0.1% when the field non uniformity factor is greater than 2.64 (field utilization factor less than 52.76%).
Keywords: Assignment factor, Charge Simulation Method, High Voltage, Numerical field computation, Non uniformity factor, Simulation errors.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20527742 Removal of Tartrazine Dye form Aqueous Solutions by Adsorption on the Surface of Polyaniline/Iron Oxide Composite
Authors: Salem Ali Jebreil
Abstract:
In this work, a polyaniline/Iron oxide (PANI/Fe2O3) composite was chemically prepared by oxidative polymerization of aniline in acid medium, in presence of ammonium persulphate as an oxidant and amount of Fe2O3. The composite was characterized by a scanning electron microscopy (SEM). The prepared composite has been used as adsorbent to remove Tartrazine dye form aqueous solutions. The effects of initial dye concentration and temperature on the adsorption capacity of PANI/Fe2O3 for Tartrazine dye have been studied in this paper. The Langmuir and Freundlich adsorption models have been used for the mathematical description of adsorption equilibrium data. The best fit is obtained using the Freundlich isotherm with an R2 value of 0.998. The change of Gibbs energy, enthalpy, and entropy of adsorption has been also evaluated for the adsorption of Tartrazine onto PANI/ Fe2O3. It has been proved according the results that the adsorption process is endothermic in nature.
Keywords: Adsorption, Composite, dye, Polyaniline, Tartrazine.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24727741 Probabilistic Electrical Power Generation Modeling Using Decimal to Binary Conversion
Authors: Ahmed S. Al-Abdulwahab
Abstract:
Generation system reliability assessment is an important task which can be performed using deterministic or probabilistic techniques. The probabilistic approaches have significant advantages over the deterministic methods. However, more complicated modeling is required by the probabilistic approaches. Power generation model is a basic requirement for this assessment. One form of the generation models is the well known capacity outage probability table (COPT). Different analytical techniques have been used to construct the COPT. These approaches require considerable mathematical modeling of the generating units. The unit-s models are combined to build the COPT which will add more burdens on the process of creating the COPT. Decimal to Binary Conversion (DBC) technique is widely and commonly applied in electronic systems and computing This paper proposes a novel utilization of the DBC to create the COPT without engaging in analytical modeling or time consuming simulations. The simple binary representation , “0 " and “1 " is used to model the states o f generating units. The proposed technique is proven to be an effective approach to build the generation model.Keywords: Decimal to Binary, generation, reliability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20417740 Adaptive Helmholtz Resonator in a Hydraulic System
Authors: Lari Kela
Abstract:
An adaptive Helmholtz resonator was designed and adapted to hydraulics. The resonator was controlled by open- and closed-loop controls so that 20 dB attenuation of the peak-to-peak value of the pulsating pressure was maintained. The closed-loop control was noted to be better, albeit it was slower because of its low pressure and temperature variation, which caused variation in the effective bulk modulus of the hydraulic system. Low-pressure hydraulics contains air, which affects the stiffness of the hydraulics, and temperature variation changes the viscosity of the oil. Thus, an open-loop control loses its efficiency if a condition such as temperature or the amount of air changes after calibration. The instability of the low-pressure hydraulic system reduced the operational frequency range of the Helmholtz resonator when compared with the results of an analytical model. Different dampers for hydraulics are presented. Then analytical models of a hydraulic pipe and a hydraulic pipe with a Helmholtz resonator are presented. The analytical models are based on the wave equation of sound pressure. Finally, control methods and the results of experiments are presented.Keywords: adaptive, damper, hydraulics, pressure, pulsating
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 43437739 An Ant-based Clustering System for Knowledge Discovery in DNA Chip Analysis Data
Authors: Minsoo Lee, Yun-mi Kim, Yearn Jeong Kim, Yoon-kyung Lee, Hyejung Yoon
Abstract:
Biological data has several characteristics that strongly differentiate it from typical business data. It is much more complex, usually large in size, and continuously changes. Until recently business data has been the main target for discovering trends, patterns or future expectations. However, with the recent rise in biotechnology, the powerful technology that was used for analyzing business data is now being applied to biological data. With the advanced technology at hand, the main trend in biological research is rapidly changing from structural DNA analysis to understanding cellular functions of the DNA sequences. DNA chips are now being used to perform experiments and DNA analysis processes are being used by researchers. Clustering is one of the important processes used for grouping together similar entities. There are many clustering algorithms such as hierarchical clustering, self-organizing maps, K-means clustering and so on. In this paper, we propose a clustering algorithm that imitates the ecosystem taking into account the features of biological data. We implemented the system using an Ant-Colony clustering algorithm. The system decides the number of clusters automatically. The system processes the input biological data, runs the Ant-Colony algorithm, draws the Topic Map, assigns clusters to the genes and displays the output. We tested the algorithm with a test data of 100 to1000 genes and 24 samples and show promising results for applying this algorithm to clustering DNA chip data.
Keywords: Ant colony system, biological data, clustering, DNA chip.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19747738 The Resource Description Framework (RDF) as a Modern Structure for Medical Data
Authors: Gabriela Lindemann, Danilo Schmidt, Thomas Schrader, Dietmar Keune
Abstract:
The amount and heterogeneity of data in biomedical research, notably in interdisciplinary fields, requires new methods for the collection, presentation and analysis of information. Important data from laboratory experiments as well as patient trials are available but come out of distributed resources. The Charité - University Hospital Berlin has established together with the German Research Foundation (DFG) a new information service centre for kidney diseases and transplantation (Open European Nephrology Science Centre - OpEN.SC). Beside a collaborative aspect to create new research groups every single partner or institution of this science information centre making his own data available is allowed to search the whole data pool of the various involved centres. A core task is the implementation of a non-restricting open data structure for the various different data sources. We decided to use a modern RDF model and in a first phase transformed original data coming from the web-based Electronic Patient Record database TBase©.
Keywords: Medical databases, Resource Description Framework (RDF), metadata repository.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20327737 Detecting the Capacity Reserve in an Overhead Line
Authors: S. Berjozkina, A. Sauhats, V. Bargels, E. Vanzovichs
Abstract:
There are various solutions for improving existing overhead line systems with the general purpose of increasing their limited capacity. The capacity reserve of the existing overhead lines is an important problem that must be considered from different aspects. The paper contains a comparative analysis of the mechanical and thermal limitations of an existing overhead line based on certain calculation conditions characterizing the examined variants. The methodology of the proposed estimation of the permissible conductor temperature and maximum load current is described in detail. The transmission line model consists of specific information of an existing overhead line of the Latvian power network. The main purpose of the simulation tasks is to find an additional capacity reserve by using accurate mathematical models. The results of the obtained data are presented.
Keywords: capacity of an overhead line, mechanical conditions, permissible conductor temperature, thermal conditions.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20037736 The Theory and Practice of the State Model of Corporate Governance
Authors: Asaiel Alohaly
Abstract:
A theoretical framework for corporate governance is needed to bridge the gap between the corporate governance of private companies and State-Owned Enterprises (SOEs). The two dominant models, being shareholder and stakeholder, do not always address the specific requirements and challenges posed by ‘hybrid’ companies; namely, previously national bodies that have been privatised while the government retains significant control or holds a majority of shares. Thus, an exploratory theoretical study is needed to identify how ‘hybrid’ companies should be defined and why the state model should be acknowledged since it is the less conspicuous model in comparison with the shareholder and stakeholder models. This research focuses on the state model of corporate governance to understand the complex ownership, control pattern, goals, and corporate governance of these hybrid companies. The significance of this research lies in the fact that there is a limited available publication on the state model. This research argues for the state model, which proceeds from an understanding of the institutionally embedded characteristics of hybrid companies, where the government as a shareholder, is either a majority of the total shares, or has been granted power based on the rule of law; the company bylaws.
Keywords: Corporate governance, control, shareholders, state model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13027735 A Generic Approach to Reuse Unified Modeling Language Components Following an Agile Process
Authors: Rim Bouhaouel, Naoufel Kraïem, Zuhoor Al Khanjari
Abstract:
Unified Modeling Language (UML) is considered as one of the widespread modeling language standardized by the Object Management Group (OMG). Therefore, the model driving engineering (MDE) community attempts to provide reuse of UML diagrams, and do not construct it from scratch. The UML model appears according to a specific software development process. The existing method generation models focused on the different techniques of transformation without considering the development process. Our work aims to construct an UML component from fragments of UML diagram basing on an agile method. We define UML fragment as a portion of a UML diagram, which express a business target. To guide the generation of fragments of UML models using an agile process, we need a flexible approach, which adapts to the agile changes and covers all its activities. We use the software product line (SPL) to derive a fragment of process agile method. This paper explains our approach, named RECUP, to generate UML fragments following an agile process, and overviews the different aspects. In this paper, we present the approach and we define the different phases and artifacts.Keywords: UML, component, fragment, agile, SPL.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9177734 Kinetic Study of Gluconic Acid Batch Fermentation by Aspergillus niger
Authors: Akbarningrum Fatmawati, Rudy Agustriyanto, Lindawati
Abstract:
Gluconic acid is one of interesting chemical products in industries such as detergents, leather, photographic, textile, and especially in food and pharmaceutical industries. Fermentation is an advantageous process to produce gluconic acid. Mathematical modeling is important in the design and operation of fermentation process. In fact, kinetic data must be available for modeling. The kinetic parameters of gluconic acid production by Aspergillus niger in batch culture was studied in this research at initial substrate concentration of 150, 200 and 250 g/l. The kinetic models used were logistic equation for growth, Luedeking-Piret equation for gluconic acid formation, and Luedeking-Piret-like equation for glucose consumption. The Kinetic parameters in the model were obtained by minimizing non linear least squares curve fitting.Keywords: Aspergillus niger, fermentation, gluconic acid, kinetic.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26937733 Female Labor Force Participation in Third World Countries: An Empirical Analysis
Authors: Anam Azam, Muhammad Rafiq
Abstract:
The study identified the socio-economic and demographic factors of both married and unmarried females in third world countries. Almost all the countries have same problems but we have selected Pakistan as a sample country. The main purpose of this study was to examine which factors forced women to participate in labor market. So the best technique of data collection was survey of both married and unmarried females between the ages of 20 to 49. Two models (probit and logit) were used to analyze the factors which effect on FLFP. The result showed that some factors e.g. age; education and marital status have significant effect on FLFP. The findings showed that educated women and those who belong to joint families are more participate because of financial pressure.
Keywords: Education, Financial status, Family pressure Labor Market participation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20297732 XML Data Management in Compressed Relational Database
Authors: Hongzhi Wang, Jianzhong Li, Hong Gao
Abstract:
XML is an important standard of data exchange and representation. As a mature database system, using relational database to support XML data may bring some advantages. But storing XML in relational database has obvious redundancy that wastes disk space, bandwidth and disk I/O when querying XML data. For the efficiency of storage and query XML, it is necessary to use compressed XML data in relational database. In this paper, a compressed relational database technology supporting XML data is presented. Original relational storage structure is adaptive to XPath query process. The compression method keeps this feature. Besides traditional relational database techniques, additional query process technologies on compressed relations and for special structure for XML are presented. In this paper, technologies for XQuery process in compressed relational database are presented..Keywords: XML, compression, query processing
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18067731 A System for Analyzing and Eliciting Public Grievances Using Cache Enabled Big Data
Authors: P. Kaladevi, N. Giridharan
Abstract:
The system for analyzing and eliciting public grievances serves its main purpose to receive and process all sorts of complaints from the public and respond to users. Due to the more number of complaint data becomes big data which is difficult to store and process. The proposed system uses HDFS to store the big data and uses MapReduce to process the big data. The concept of cache was applied in the system to provide immediate response and timely action using big data analytics. Cache enabled big data increases the response time of the system. The unstructured data provided by the users are efficiently handled through map reduce algorithm. The processing of complaints takes place in the order of the hierarchy of the authority. The drawbacks of the traditional database system used in the existing system are set forth by our system by using Cache enabled Hadoop Distributed File System. MapReduce framework codes have the possible to leak the sensitive data through computation process. We propose a system that add noise to the output of the reduce phase to avoid signaling the presence of sensitive data. If the complaints are not processed in the ample time, then automatically it is forwarded to the higher authority. Hence it ensures assurance in processing. A copy of the filed complaint is sent as a digitally signed PDF document to the user mail id which serves as a proof. The system report serves to be an essential data while making important decisions based on legislation.Keywords: Big Data, Hadoop, HDFS, Caching, MapReduce, web personalization, e-governance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15927730 Implementation of MPPT Algorithm for Grid Connected PV Module with IC and P&O Method
Authors: Arvind Kumar, Manoj Kumar, Dattatraya H. Nagaraj, Amanpreet Singh, Jayanthi Prattapati
Abstract:
In recent years, the use of renewable energy resources instead of pollutant fossil fuels and other forms has increased. Photovoltaic generation is becoming increasingly important as a renewable resource since it does not cause in fuel costs, pollution, maintenance, and emitting noise compared with other alternatives used in power applications. In this paper, Perturb and Observe and Incremental Conductance methods are used to improve energy conversion efficiency under different environmental conditions. PI controllers are used to control easily DC-link voltage, active and reactive currents. The whole system is simulated under standard climatic conditions (1000 W/m2, 250C) in MATLAB and the irradiance is varied from 1000 W/m2 to 300 W/m2. The use of PI controller makes it easy to directly control the power of the grid connected PV system. Finally the validity of the system will be verified through the simulations in MATLAB/Simulink environment.Keywords: Incremental conductance algorithm, modeling of PV panel, perturb and observe algorithm, photovoltaic system and simulation results.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1863