Search results for: Data privacy
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7460

Search results for: Data privacy

4880 Research on Transformer Condition-based Maintenance System using the Method of Fuzzy Comprehensive Evaluation

Authors: Po-Chun Lin, Jyh-Cherng Gu

Abstract:

This study adopted previous fault patterns, results of detection analysis, historical records and data, and experts- experiences to establish fuzzy principles and estimate the failure probability index of components of a power transformer. Considering that actual parameters and limiting conditions of parameters may differ, this study used the standard data of IEC, IEEE, and CIGRE as condition parameters. According to the characteristics of each condition parameter, relative degradation was introduced to reflect the degree of influence of the factors on the transformer condition. The method of fuzzy mathematics was adopted to determine the subordinate function of the transformer condition. The calculation used the Matlab Fuzzy Tool Box to select the condition parameters of coil winding, iron core, bushing, OLTC, insulating oil and other auxiliary components and factors (e.g., load records, performance history, and maintenance records) of the transformer to establish the fuzzy principles. Examples were presented to support the rationality and effectiveness of the evaluation method of power transformer performance conditions, as based on fuzzy comprehensive evaluation.

Keywords: Fuzzy, relative degradation degree, condition-basedmaintenance, power transformer

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2443
4879 Social Semantic Web-Based Analytics Approach to Support Lifelong Learning

Authors: Khaled Halimi, Hassina Seridi-Bouchelaghem

Abstract:

The purpose of this paper is to describe how learning analytics approaches based on social semantic web techniques can be applied to enhance the lifelong learning experiences in a connectivist perspective. For this reason, a prototype of a system called SoLearn (Social Learning Environment) that supports this approach. We observed and studied literature related to lifelong learning systems, social semantic web and ontologies, connectivism theory, learning analytics approaches and reviewed implemented systems based on these fields to extract and draw conclusions about necessary features for enhancing the lifelong learning process. The semantic analytics of learning can be used for viewing, studying and analysing the massive data generated by learners, which helps them to understand through recommendations, charts and figures their learning and behaviour, and to detect where they have weaknesses or limitations. This paper emphasises that implementing a learning analytics approach based on social semantic web representations can enhance the learning process. From one hand, the analysis process leverages the meaning expressed by semantics presented in the ontology (relationships between concepts). From the other hand, the analysis process exploits the discovery of new knowledge by means of inferring mechanism of the semantic web.

Keywords: Connectivism, data visualization, informal learning, learning analytics, semantic web, social web.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 785
4878 3D Shape Modelling of Left Ventricle: Towards Correlation of Myocardial Scintigraphy Data and Coronarography Result

Authors: A. Ben Abdallah, H. Essabbah, M. H. Bedoui

Abstract:

The myocardial sintigraphy is an imaging modality which provides functional informations. Whereas, coronarography modality gives useful informations about coronary arteries anatomy. In case of coronary artery disease (CAD), the coronarography can not determine precisely which moderate lesions (artery reduction between 50% and 70%), known as the “gray zone", are haemodynamicaly significant. In this paper, we aim to define the relationship between the location and the degree of the stenosis in coronary arteries and the observed perfusion on the myocardial scintigraphy. This allows us to model the impact evolution of these stenoses in order to justify a coronarography or to avoid it for patients suspected being in the gray zone. Our approach is decomposed in two steps. The first step consists in modelling a coronary artery bed and stenoses of different location and degree. The second step consists in modelling the left ventricle at stress and at rest using the sphercical harmonics model and myocardial scintigraphic data. We use the spherical harmonics descriptors to analyse left ventricle model deformation between stress and rest which permits us to conclude if ever an ischemia exists and to quantify it.

Keywords: Spherical harmonics model, vascular bed, 3D reconstruction, left ventricle, myocardial scintigraphy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1768
4877 Perturbed-Chain Statistical Association Fluid Theory (PC-SAFT) Parameters for Propane, Ethylene, and Hydrogen under Supercritical Conditions

Authors: Ilke Senol

Abstract:

Perturbed-Chain Statistical Association Fluid Theory (PC-SAFT) equation of state (EOS) is a modified SAFT EOS with three pure component specific parameters: segment number (m), diameter (σ) and energy (ε). These PC-SAFT parameters need to be determined for each component under the conditions of interest by fitting experimental data, such as vapor pressure, density or heat capacity. PC-SAFT parameters for propane, ethylene and hydrogen in supercritical region were successfully estimated by fitting experimental density data available in literature. The regressed PCSAFT parameters were compared with the literature values by means of estimating pure component density and calculating average absolute deviation between the estimated and experimental density values. PC-SAFT parameters available in literature especially for ethylene and hydrogen estimated density in supercritical region reasonably well. However, the regressed PC-SAFT parameters performed better in supercritical region than the PC-SAFT parameters from literature.

Keywords: Equation of state, perturbed-chain, PC-SAFT, super critical.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6948
4876 Thermodynamic Modeling of the High Temperature Shift Converter Reactor Using Minimization of Gibbs Free Energy

Authors: H. Zare Aliabadi

Abstract:

The equilibrium chemical reactions taken place in a converter reactor of the Khorasan Petrochemical Ammonia plant was studied using the minimization of Gibbs free energy method. In the minimization of the Gibbs free energy function the Davidon– Fletcher–Powell (DFP) optimization procedure using the penalty terms in the well-defined objective function was used. It should be noted that in the DFP procedure along with the corresponding penalty terms the Hessian matrices for the composition of constituents in the Converter reactor can be excluded. This, in fact, can be considered as the main advantage of the DFP optimization procedure. Also the effect of temperature and pressure on the equilibrium composition of the constituents was investigated. The results obtained in this work were compared with the data collected from the converter reactor of the Khorasan Petrochemical Ammonia plant. It was concluded that the results obtained from the method used in this work are in good agreement with the industrial data. Notably, the algorithm developed in this work, in spite of its simplicity, takes the advantage of short computation and convergence time.

Keywords: Gibbs free energy, converter reactors, Chemical equilibrium

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2537
4875 Survey of Access Controls in Cloud Computing

Authors: Monirah Alkathiry, Hanan Aljarwan

Abstract:

Cloud computing is one of the most significant technologies that the world deals with, in different sectors with different purposes and capabilities. The cloud faces various challenges in securing data from unauthorized access or modification. Consequently, security risks and levels have greatly increased. Therefore, cloud service providers (CSPs) and users need secure mechanisms that ensure that data are kept secret and safe from any disclosures or exploits. For this reason, CSPs need a number of techniques and technologies to manage and secure access to the cloud services to achieve security goals, such as confidentiality, integrity, identity access management (IAM), etc. Therefore, this paper will review and explore various access controls implemented in a cloud environment that achieve different security purposes. The methodology followed in this survey was conducting an assessment, evaluation, and comparison between those access controls mechanisms and technologies based on different factors, such as the security goals it achieves, usability, and cost-effectiveness. This assessment resulted in the fact that the technology used in an access control affects the security goals it achieves as well as there is no one access control method that achieves all security goals. Consequently, such a comparison would help decision-makers to choose properly the access controls that meet their requirements.

Keywords: Access controls, cloud computing, confidentiality, identity and access management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 697
4874 Nonlinear Analysis of Postural Sway in Multiple Sclerosis

Authors: Hua Cao, Laurent Peyrodie, Olivier Agnani, Cécile Donzé

Abstract:

Multiple Sclerosis (MS) is a disease which affects the central nervous system and causes balance problem. In clinical, this disorder is usually evaluated using static posturography. Some linear or nonlinear measures, extracted from the posturographic data (i.e. center of pressure, COP) recorded during a balance test, has been used to analyze postural control of MS patients. In this study, the trend (TREND) and the sample entropy (SampEn), two nonlinear parameters were chosen to investigate their relationships with the expanded disability status scale (EDSS) score. 40 volunteers with different EDSS scores participated in our experiments with eyes open (EO) and closed (EC). TREND and 2 types of SampEn (SampEn1 and SampEn2) were calculated for each combined COP’s position signal. The results have shown that TREND had a weak negative correlation to EDSS while SampEn2 had a strong positive correlation to EDSS. Compared to TREND and SampEn1, SampEn2 showed a better significant correlation to EDSS and an ability to discriminate the MS patients in the EC case. In addition, the outcome of the study suggests that the multi-dimensional nonlinear analysis could provide some information about the impact of disability progression in MS on dynamics of the COP data.

Keywords: Balance, multiple sclerosis, nonlinear analysis, postural sway.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1953
4873 Applying Multiple Kinect on the Development of a Rapid 3D Mannequin Scan Platform

Authors: Shih-Wen Hsiao, Yi-Cheng Tsao

Abstract:

In the field of reverse engineering and creative industries, applying 3D scanning process to obtain geometric forms of the objects is a mature and common technique. For instance, organic objects such as faces and nonorganic objects such as products could be scanned to acquire the geometric information for further application. However, although the data resolution of 3D scanning device is increasing and there are more and more abundant complementary applications, the penetration rate of 3D scanning for the public is still limited by the relative high price of the devices. On the other hand, Kinect, released by Microsoft, is known for its powerful functions, considerably low price, and complete technology and database support. Therefore, related studies can be done with the applying of Kinect under acceptable cost and data precision. Due to the fact that Kinect utilizes optical mechanism to extracting depth information, limitations are found due to the reason of the straight path of the light. Thus, various angles are required sequentially to obtain the complete 3D information of the object when applying a single Kinect for 3D scanning. The integration process which combines the 3D data from different angles by certain algorithms is also required. This sequential scanning process costs much time and the complex integration process often encounter some technical problems. Therefore, this paper aimed to apply multiple Kinects simultaneously on the field of developing a rapid 3D mannequin scan platform and proposed suggestions on the number and angles of Kinects. In the content, a method of establishing the coordination based on the relation between mannequin and the specifications of Kinect is proposed, and a suggestion of angles and number of Kinects is also described. An experiment of applying multiple Kinect on the scanning of 3D mannequin is constructed by Microsoft API, and the results show that the time required for scanning and technical threshold can be reduced in the industries of fashion and garment design.

Keywords: 3D scan, depth sensor, fashion and garment design, mannequin, multiple kinect sensor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2252
4872 A Multigrid Approach for Three-Dimensional Inverse Heat Conduction Problems

Authors: Jianhua Zhou, Yuwen Zhang

Abstract:

A two-step multigrid approach is proposed to solve the inverse heat conduction problem in a 3-D object under laser irradiation. In the first step, the location of the laser center is estimated using a coarse and uniform grid system. In the second step, the front-surface temperature is recovered in good accuracy using a multiple grid system in which fine mesh is used at laser spot center to capture the drastic temperature rise in this region but coarse mesh is employed in the peripheral region to reduce the total number of sensors required. The effectiveness of the two-step approach and the multiple grid system are demonstrated by the illustrative inverse solutions. If the measurement data for the temperature and heat flux on the back surface do not contain random error, the proposed multigrid approach can yield more accurate inverse solutions. When the back-surface measurement data contain random noise, accurate inverse solutions cannot be obtained if both temperature and heat flux are measured on the back surface.

Keywords: Conduction, inverse problems, conjugated gradient method, laser.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 819
4871 Variability of Metal Composition and Concentrations in Road Dust in the Urban Environment

Authors: Sandya Mummullage, Prasanna Egodawatta, Ashantha Goonetilleke, Godwin A. Ayoko

Abstract:

Urban road dust comprises of a range of potentially  toxic metal elements and plays a critical role in degrading urban  receiving water quality. Hence, assessing the metal composition and  concentration in urban road dust is a high priority. This study  investigated the variability of metal composition and concentrations  in road dust in 4 different urban land uses in Gold Coast, Australia.  Samples from 16 road sites were collected and tested for selected 12  metal species. The data set was analyzed using both univariate and  multivariate techniques. Outcomes of the data analysis revealed that  the metal concentrations inroad dust differs considerably within and  between different land uses. Iron, aluminum, magnesium and zinc are  the most abundant in urban land uses. It was also noted that metal  species such as titanium, nickel, copper and zinc have the highest  concentrations in industrial land use. The study outcomes revealed  that soil and traffic related sources as key sources of metals deposited  on road surfaces.

 

Keywords: Metals build-up, Pollutant accumulation, Stormwater quality, Urban road dust.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2328
4870 Cooperative Energy Efficient Routing for Wireless Sensor Networks in Smart Grid Communications

Authors: Ghazi AL-Sukkar, Iyad Jafar, Khalid Darabkh, Raed Al-Zubi, Mohammed Hawa

Abstract:

Smart Grids employ wireless sensor networks for their control and monitoring. Sensors are characterized by limitations in the processing power, energy supply and memory spaces, which require a particular attention on the design of routing and data management algorithms. Since most routing algorithms for sensor networks, focus on finding energy efficient paths to prolong the lifetime of sensor networks, the power of sensors on efficient paths depletes quickly, and consequently sensor networks become incapable of monitoring events from some parts of their target areas. In consequence, the design of routing protocols should consider not only energy efficiency paths, but also energy efficient algorithms in general. In this paper we propose an energy efficient routing protocol for wireless sensor networks without the support of any location information system. The reliability and the efficiency of this protocol have been demonstrated by simulation studies where we compare them to the legacy protocols. Our simulation results show that these algorithms scale well with network size and density.

Keywords: Data-centric storage, Dynamic Address Allocation, Sensor networks, Smart Grid Communications.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1833
4869 Investigation of Hydraulic and Thermal Performances of Fin Array at Different Shield Positions without By-Pass

Authors: Ramy H. Mohammed

Abstract:

In heat sinks, the flow within the core exhibits separation and hence does not lend itself to simple analytical boundary layer or duct flow analysis of the wall friction. In this paper, we present some findings from an experimental and numerical study aimed to obtain physical insight into the influence of the presence of the shield and its position on the hydraulic and thermal performance of square pin fin heat sink without top by-pass. The variations of the Nusselt number and friction factor are obtained under varied parameters, such as the Reynolds number and the shield position. The numerical code is validated by comparing the numerical results with the available experimental data. It is shown that, there is a good agreement between the temperature predictions based on the model and the experimental data. Results show that, as the presence of the shield, the heat transfer of fin array is enhanced and the flow resistance increased. The surface temperature distribution of the heat sink base is more uniform when the dimensionless shield position equals to 1/3 or 2/3. The comprehensive performance evaluation approach based on identical pumping power criteria is adopted and shows that the optimum shield position is at x/l=0.43.

Keywords: Shield, Fin array, Performance evaluation, Heat transfer, Validation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1818
4868 Fast Adjustable Threshold for Uniform Neural Network Quantization

Authors: Alexander Goncharenko, Andrey Denisov, Sergey Alyamkin, Evgeny Terentev

Abstract:

The neural network quantization is highly desired procedure to perform before running neural networks on mobile devices. Quantization without fine-tuning leads to accuracy drop of the model, whereas commonly used training with quantization is done on the full set of the labeled data and therefore is both time- and resource-consuming. Real life applications require simplification and acceleration of quantization procedure that will maintain accuracy of full-precision neural network, especially for modern mobile neural network architectures like Mobilenet-v1, MobileNet-v2 and MNAS. Here we present a method to significantly optimize training with quantization procedure by introducing the trained scale factors for discretization thresholds that are separate for each filter. Using the proposed technique, we quantize the modern mobile architectures of neural networks with the set of train data of only ∼ 10% of the total ImageNet 2012 sample. Such reduction of train dataset size and small number of trainable parameters allow to fine-tune the network for several hours while maintaining the high accuracy of quantized model (accuracy drop was less than 0.5%). Ready-for-use models and code are available in the GitHub repository.

Keywords: Distillation, machine learning, neural networks, quantization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 707
4867 Person-Environment Fit (PE Fit): Evidence from Brazil

Authors: Jucelia Appio, Danielle Deimling De Carli, Bruno Henrique Rocha Fernandes, Nelson Natalino Frizon

Abstract:

The purpose of this paper is to investigate if there are positive and significant correlations between the dimensions of Person-Environment Fit (Person-Job, Person-Organization, Person-Group and Person-Supervisor) at the “Best Companies to Work for” in Brazil in 2017. For that, a quantitative approach was used with a descriptive method being defined as a research sample the "150 Best Companies to Work for", according to data base collected in 2017 and provided by Fundação Instituto of Administração (FIA) of the University of São Paulo (USP). About the data analysis procedures, asymmetry and kurtosis, factorial analysis, Kaiser-Meyer-Olkin (KMO) tests, Bartlett sphericity and Cronbach's alpha were used for the 69 research variables, and as a statistical technique for the purpose of analyzing the hypothesis, Pearson's correlation analysis was performed. As a main result, we highlight that there was a positive and significant correlation between the dimensions of Person-Environment Fit, corroborating the H1 hypothesis that there is a positive and significant correlation between Person-Job Fit, Person-Organization Fit, Person-Group Fit and Person-Supervisor Fit.

Keywords: Human resource management, person-environment fit, strategic people management, best companies to work for.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 973
4866 Investigating the Vehicle-Bicyclists Conflicts Using LIDAR Sensor Technology at Signalized Intersections

Authors: Alireza Ansariyar, Mansoureh Jeihani

Abstract:

Light Detection and Ranging (LiDAR) sensors is capable of recording traffic data including the number of passing vehicles and bicyclists, the speed of vehicles and bicyclists, and the number of conflicts among both road users. In order to collect real-time traffic data and investigate the safety of different road users, a LiDAR sensor was installed at Cold Spring Ln – Hillen Rd intersection in Baltimore city. The frequency and severity of collected real-time conflicts were analyzed and the results highlighted that 122 conflicts were recorded over a 10-month time interval from May 2022 to February 2023. By employing an image-processing algorithm, a safety Measure of Effectiveness (MOE) aims to identify critical zones for bicyclists upon entering each respective zone at the signalized intersection. Considering the trajectory of conflicts, the results of analysis demonstrated that conflicts in the northern approach (zone N) are more frequent and severe. Additionally, sunny weather is more likely to cause severe vehicle-bike conflicts.

Keywords: LiDAR sensor, Post Encroachment Time threshold, vehicle-bike conflicts, measure of effectiveness, weather condition.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 99
4865 Application of Geographic Information Systems(GIS) in the History of Cartography

Authors: Bangbo Hu

Abstract:

This paper discusses applications of a revolutionary information technology, Geographic Information Systems (GIS), in the field of the history of cartography by examples, including assessing accuracy of early maps, establishing a database of places and historical administrative units in history, integrating early maps in GIS or digital images, and analyzing social, political, and economic information related to production of early maps. GIS provides a new mean to evaluate the accuracy of early maps. Four basic steps using GIS for this type of study are discussed. In addition, several historical geographical information systems are introduced. These include China Historical Geographic Information Systems (CHGIS), the United States National Historical Geographic Information System (NHGIS), and the Great Britain Historical Geographical Information System. GIS also provides digital means to display and analyze the spatial information on the early maps or to layer them with modern spatial data. How GIS relational data structure may be used to analyze social, political, and economic information related to production of early maps is also discussed in this paper. Through discussion on these examples, this paper reveals value of GIS applications in this field.

Keywords: Cartography, GIS, history, maps.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3120
4864 New Features for Specific JPEG Steganalysis

Authors: Johann Barbier, Eric Filiol, Kichenakoumar Mayoura

Abstract:

We present in this paper a new approach for specific JPEG steganalysis and propose studying statistics of the compressed DCT coefficients. Traditionally, steganographic algorithms try to preserve statistics of the DCT and of the spatial domain, but they cannot preserve both and also control the alteration of the compressed data. We have noticed a deviation of the entropy of the compressed data after a first embedding. This deviation is greater when the image is a cover medium than when the image is a stego image. To observe this deviation, we pointed out new statistic features and combined them with the Multiple Embedding Method. This approach is motivated by the Avalanche Criterion of the JPEG lossless compression step. This criterion makes possible the design of detectors whose detection rates are independent of the payload. Finally, we designed a Fisher discriminant based classifier for well known steganographic algorithms, Outguess, F5 and Hide and Seek. The experiemental results we obtained show the efficiency of our classifier for these algorithms. Moreover, it is also designed to work with low embedding rates (< 10-5) and according to the avalanche criterion of RLE and Huffman compression step, its efficiency is independent of the quantity of hidden information.

Keywords: Compressed frequency domain, Fisher discriminant, specific JPEG steganalysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2142
4863 Comparison of Stochastic Point Process Models of Rainfall in Singapore

Authors: Y. Lu, X. S. Qin

Abstract:

Extensive rainfall disaggregation approaches have been developed and applied in climate change impact studies such as flood risk assessment and urban storm water management.In this study, five rainfall models that were capable ofdisaggregating daily rainfall data into hourly one were investigated for the rainfall record in theChangi Airport, Singapore. The objectives of this study were (i) to study the temporal characteristics of hourly rainfall in Singapore, and (ii) to evaluate the performance of variousdisaggregation models. The used models included: (i) Rectangular pulse Poisson model (RPPM), (ii) Bartlett-Lewis Rectangular pulse model (BLRPM), (iii) Bartlett-Lewis model with 2 cell types (BL2C), (iv) Bartlett-Lewis Rectangular with cell depth distribution dependent on duration (BLRD), and (v) Neyman-Scott Rectangular pulse model (NSRPM). All of these models werefitted using hourly rainfall data ranging from 1980 to 2005 (which was obtained from Changimeteorological station).The study results indicated that the weight scheme of inversely proportional variance could deliver more accurateoutputs for fitting rainfall patterns in tropical areas, and BLRPM performedrelatively better than other disaggregation models.

Keywords: Rainfall disaggregation, statistical properties, poisson processed, Bartlett-Lewis model, Neyman-Scott model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2266
4862 Surveying Earthquake Vulnerabilities of District 13 of Kabul City, Afghanistan

Authors: Mohsen Mohammadi, Toshio Fujimi

Abstract:

High population and irregular urban development in Kabul city, Afghanistan's capital, are among factors that increase its vulnerability to earthquake disasters (on top of its location in a high seismic region); this can lead to widespread economic loss and casualties. This study aims to evaluate earthquake risks in Kabul's 13th district based on scientific data. The research data, which include hazard curves of Kabul, vulnerability curves, and a questionnaire survey through sampling in district 13, have been incorporated to develop risk curves. To estimate potential casualties, we used a set of M parameters in a model developed by Coburn and Spence. The results indicate that in the worst case scenario, more than 90% of district 13, which comprises mostly residential buildings, is exposed to high risk; this may lead to nearly 1000 million USD economic loss and 120 thousand casualties (equal to 25.88% of the 13th district's population) for a nighttime earthquake. To reduce risks, we present the reconstruction of the most vulnerable buildings, which are primarily adobe and masonry buildings. A comparison of risk reduction between reconstructing adobe and masonry buildings indicates that rebuilding adobe buildings would be more effective.

Keywords: Earthquake risk evaluation, Kabul, mitigation, vulnerability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1613
4861 Biosynthesis and Metabolism of Anthraquinone Derivatives

Authors: Dmitry Yu. Korulkin, Raissa A. Muzychkina

Abstract:

In review the generalized data about biosynthetic routs formation anthraquinone molecules in natural cells. The basic possibilities of various ways of biosynthesis of different quinoid substances are shown.

Keywords: Anthraquinones, biochemical evolution, biosynthesis, metabolism.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3936
4860 In vitro Cytotoxic and Genotoxic Effects of Arsenic Trioxide on Human Keratinocytes

Authors: H. Bouaziz, M. Sefi, J. de Lapuente, M. Borras, N. Zeghal

Abstract:

Although, arsenic trioxide has been the subject of toxicological research, in vitro cytotoxicity and genotoxicity studies using relevant cell models and uniform methodology are not well elucidated. Hence, the aim of the present study was to evaluate the cytotoxicity and genotoxicity induced by arsenic trioxide in human keratinocytes (HaCaT) using the MTT [3-(4, 5-dimethylthiazol-2-yl)- 2,5-diphenyltetrazolium bromide] and alkaline single cell gel electrophoresis (Comet) assays, respectively. Human keratinocytes were treated with different doses of arsenic trioxide for 4 h prior to cytogenetic assessment. Data obtained from the MTT assay indicated that arsenic trioxide significantly reduced the viability of HaCaT cells in a dose-dependent manner, showing an IC50 value of 34.18 ± 0.6 μM. Data generated from the comet assay also indicated a significant dose-dependent increase in DNA damage in HaCaT cells associated with arsenic trioxide exposure. We observed a significant increase in comet tail length and tail moment, showing an evidence of arsenic trioxide -induced genotoxic damage in HaCaT cells. This study confirms that the comet assay is a sensitive and effective method to detect DNA damage caused by arsenic.

Keywords: Arsenic trioxide, cytotoxixity, genotoxicity, HaCaT.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2208
4859 Designing Information Systems in Education as Prerequisite for Successful Management Results

Authors: Vladimir Simovic, Matija Varga, Tonco Marusic

Abstract:

This research paper shows matrix technology models and examples of information systems in education (in the Republic of Croatia and in the Germany) in support of business, education (when learning and teaching) and e-learning. Here we researched and described the aims and objectives of the main process in education and technology, with main matrix classes of data. In this paper, we have example of matrix technology with detailed description of processes related to specific data classes in the processes of education and an example module that is support for the process: ‘Filling in the directory and the diary of work’ and ‘evaluation’. Also, on the lower level of the processes, we researched and described all activities which take place within the lower process in education. We researched and described the characteristics and functioning of modules: ‘Fill the directory and the diary of work’ and ‘evaluation’. For the analysis of the affinity between the aforementioned processes and/or sub-process we used our application model created in Visual Basic, which was based on the algorithm for analyzing the affinity between the observed processes and/or sub-processes.

Keywords: Designing, education management, information systems, matrix technology, process affinity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1073
4858 STEP Implementation on Turn-mill Manufacturing Environment

Authors: Ahmad Majdi Bin Abdul-Rani, Mesfin Gizaw, Yusri Yusof

Abstract:

Researches related to standard product model and development of neutral manufacturing interfaces for numerical control machines becomes a significant topic since the last 25 years. In this paper, a detail description of STEP implementation on turnmill manufacturing has been discussed. It shows requirements of information contents from ISO14649 data model. It covers to describe the design of STEP-NC framework applicable to turn-mill manufacturing. In the framework, EXPRESS-G and UML modeling tools are used to depict the information contents of the system and established the bases of information model requirement. A product and manufacturing data model applicable for STEP compliant manufacturing. The next generation turn-mill operations requirements have been represented by a UML diagram. An object oriented classes of ISO1449 has been developed on Visual Basic dot NET platform for binding the static information model represented by the UML diagram. An architect of the proposed system implementation has been given on the bases of the design and manufacturing module of STEP-NC interface established. Finally, a part 21 file process plan generated for an illustration of turn-mill components.

Keywords: CAPP, ISO14649, Product modeling, STEP-NC

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1641
4857 Retarding Potential Analyzer Design and Result Analysis for Ion Energy Distribution Measurement of the Thruster Plume in the Laboratory

Authors: Ma Ya-li, Tang Fu-jun, Xue Yu-xiong, Chen Yi-feng, Gao Xin, Wang Yi, Tian Kai, Yan Ze-dong

Abstract:

Plasma plume will be produced and arrive at spacecraft when the electric thruster operates on orbit. It-s important to characterize the thruster plasma parameters because the plume has significant effects or hazards on spacecraft sub-systems and parts. Through the ground test data of the desired parameters, the major characteristics of the thruster plume will be achieved. Also it is very important for optimizing design of Ion thruster. Retarding Potential Analyzer (RPA) is an effective instrument for plasma ion energy per unit charge distribution measurement. Special RPA should be designed according to certain plume plasma parameters range and feature. In this paper, major principles usable for good RPA design are discussed carefully. Conform to these principles, a four-grid planar electrostatic energy analyzer RPA was designed to avoid false data, and details were discussed including construction, materials, aperture diameter and so on. At the same time, it was designed more suitable for credible and long-duration measurements in the laboratory. In the end, RPA measurement results in the laboratory were given and discussed.

Keywords: Thruster plume ion energy distributions, retarding potential analyzer, ground test.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3968
4856 The Relationship between Competency-Based Learning and Learning Efficiency of Media Communication Students at Suan Sunandha Rajabhat University

Authors: Somtop Keawchuer

Abstract:

This research aims to study (1) the relationship between competency-based learning and learning efficiency of new media communication students at Suan Sunandha University (2) the demographic factor effect on learning efficiency of students at Suan Sunandha University. This research method will use quantitative research; data was collected by questionnaires distributed to students from new media communication in management science faculty of Suan Sunandha Rajabhat University for 1340 sample by purposive sampling method. Data was analyzed by descriptive statistic including percentage, mean, standard deviation and inferential statistic including T-test, ANOVA and Pearson correlation for hypothesis testing. The results showed that the competency-based learning in term of ability to communicate, ability to think and solve the problem, life skills and ability to use technology has a significant relationship with learning efficiency in term of the cognitive domain, psychomotor domain and affective domain at the 0.05 level and which is in harmony with the research hypotheses.

Keywords: Competency-based learning, learning efficiency, new media communication students, Suan Sunandha Rajabhat University.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1149
4855 Multi-Agent System for Irrigation Using Fuzzy Logic Algorithm and Open Platform Communication Data Access

Authors: T. Wanyama, B. Far

Abstract:

Automatic irrigation systems usually conveniently protect landscape investment. While conventional irrigation systems are known to be inefficient, automated ones have the potential to optimize water usage. In fact, there is a new generation of irrigation systems that are smart in the sense that they monitor the weather, soil conditions, evaporation and plant water use, and automatically adjust the irrigation schedule. In this paper, we present an agent based smart irrigation system. The agents are built using a mix of commercial off the shelf software, including MATLAB, Microsoft Excel and KEPServer Ex5 OPC server, and custom written code. The Irrigation Scheduler Agent uses fuzzy logic to integrate the information that affect the irrigation schedule. In addition, the Multi-Agent system uses Open Platform Connectivity (OPC) technology to share data. OPC technology enables the Irrigation Scheduler Agent to communicate over the Internet, making the system scalable to a municipal or regional agent based water monitoring, management, and optimization system. Finally, this paper presents simulation and pilot installation test result that show the operational effectiveness of our system.

Keywords: Community water usage, fuzzy logic, irrigation, multi-agent system.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1314
4854 On Adaptive Optimization of Filter Performance Based on Markov Representation for Output Prediction Error

Authors: Hong Son Hoang, Remy Baraille

Abstract:

This paper addresses the problem of how one can improve the performance of a non-optimal filter. First the theoretical question on dynamical representation for a given time correlated random process is studied. It will be demonstrated that for a wide class of random processes, having a canonical form, there exists a dynamical system equivalent in the sense that its output has the same covariance function. It is shown that the dynamical approach is more effective for simulating and estimating a Markov and non- Markovian random processes, computationally is less demanding, especially with increasing of the dimension of simulated processes. Numerical examples and estimation problems in low dimensional systems are given to illustrate the advantages of the approach. A very useful application of the proposed approach is shown for the problem of state estimation in very high dimensional systems. Here a modified filter for data assimilation in an oceanic numerical model is presented which is proved to be very efficient due to introducing a simple Markovian structure for the output prediction error process and adaptive tuning some parameters of the Markov equation.

Keywords: Statistical simulation, canonical form, dynamical system, Markov and non-Markovian processes, data assimilation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1280
4853 Validating Condition-Based Maintenance Algorithms Through Simulation

Authors: Marcel Chevalier, Léo Dupont, Sylvain Marié, Frédérique Roffet, Elena Stolyarova, William Templier, Costin Vasile

Abstract:

Industrial end users are currently facing an increasing need to reduce the risk of unexpected failures and optimize their maintenance. This calls for both short-term analysis and long-term ageing anticipation. At Schneider Electric, we tackle those two issues using both Machine Learning and First Principles models. Machine learning models are incrementally trained from normal data to predict expected values and detect statistically significant short-term deviations. Ageing models are constructed from breaking down physical systems into sub-assemblies, then determining relevant degradation modes and associating each one to the right kinetic law. Validating such anomaly detection and maintenance models is challenging, both because actual incident and ageing data are rare and distorted by human interventions, and incremental learning depends on human feedback. To overcome these difficulties, we propose to simulate physics, systems and humans – including asset maintenance operations – in order to validate the overall approaches in accelerated time and possibly choose between algorithmic alternatives.

Keywords: Degradation models, ageing, anomaly detection, soft sensor, incremental learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 278
4852 The Bright Side of Organizational Politics as a Driver of Firm Competitiveness: The Mediating Role of Corporate Entrepreneurship

Authors: Monika Kulikowska-Pawlak, Katarzyna Bratnicka-Myśliwiec, Tomasz Ingram

Abstract:

This study seeks to contribute to the literature on firm competitiveness by advancing the perspective of organizational politics that views this process as a driver which creates identifiable differences in firm performance. The hypothesized relationships were tested on the basis of data from 355 Polish medium and large-sized enterprises. Data were analyzed using correlation analysis, EFA and robustness tests. The main result of the conducted analyses proved the coexistence, previously examined in the literature, of corporate entrepreneurship and firm performance. The obtained research findings made it possible to add organizational politics to a wide range of elements determining corporate entrepreneurship, followed by competitive advantage, in addition to antecedents such as strategic leadership, corporate culture, opportunity-oriented resource-based management, etc. Also, the empirical results suggest that four dimensions of organizational politics (dominant coalition, influence exertion, making organizational changes, and information openness) are positively related to firm competitiveness. In addition, these findings seem to underline a supposition that corporate entrepreneurship is an important mediator which strengthens the competitive effects of organizational politics.

Keywords: Corporate entrepreneurship, firm competitiveness organizational politics, sensemaking.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 909
4851 Analyzing the Factors that Cause Parallel Performance Degradation in Parallel Graph-Based Computations Using Graph500

Authors: Mustafa Elfituri, Jonathan Cook

Abstract:

Recently, graph-based computations have become more important in large-scale scientific computing as they can provide a methodology to model many types of relations between independent objects. They are being actively used in fields as varied as biology, social networks, cybersecurity, and computer networks. At the same time, graph problems have some properties such as irregularity and poor locality that make their performance different than regular applications performance. Therefore, parallelizing graph algorithms is a hard and challenging task. Initial evidence is that standard computer architectures do not perform very well on graph algorithms. Little is known exactly what causes this. The Graph500 benchmark is a representative application for parallel graph-based computations, which have highly irregular data access and are driven more by traversing connected data than by computation. In this paper, we present results from analyzing the performance of various example implementations of Graph500, including a shared memory (OpenMP) version, a distributed (MPI) version, and a hybrid version. We measured and analyzed all the factors that affect its performance in order to identify possible changes that would improve its performance. Results are discussed in relation to what factors contribute to performance degradation.

Keywords: Graph computation, Graph500 benchmark, parallel architectures, parallel programming, workload characterization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 511