Search results for: operator action validation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4120

Search results for: operator action validation

2890 A Large Ion Collider Experiment (ALICE) Diffractive Detector Control System for RUN-II at the Large Hadron Collider

Authors: J. C. Cabanillas-Noris, M. I. Martínez-Hernández, I. León-Monzón

Abstract:

The selection of diffractive events in the ALICE experiment during the first data taking period (RUN-I) of the Large Hadron Collider (LHC) was limited by the range over which rapidity gaps occur. It would be possible to achieve better measurements by expanding the range in which the production of particles can be detected. For this purpose, the ALICE Diffractive (AD0) detector has been installed and commissioned for the second phase (RUN-II). Any new detector should be able to take the data synchronously with all other detectors and be operated through the ALICE central systems. One of the key elements that must be developed for the AD0 detector is the Detector Control System (DCS). The DCS must be designed to operate safely and correctly this detector. Furthermore, the DCS must also provide optimum operating conditions for the acquisition and storage of physics data and ensure these are of the highest quality. The operation of AD0 implies the configuration of about 200 parameters, from electronics settings and power supply levels to the archiving of operating conditions data and the generation of safety alerts. It also includes the automation of procedures to get the AD0 detector ready for taking data in the appropriate conditions for the different run types in ALICE. The performance of AD0 detector depends on a certain number of parameters such as the nominal voltages for each photomultiplier tube (PMT), their threshold levels to accept or reject the incoming pulses, the definition of triggers, etc. All these parameters define the efficiency of AD0 and they have to be monitored and controlled through AD0 DCS. Finally, AD0 DCS provides the operator with multiple interfaces to execute these tasks. They are realized as operating panels and scripts running in the background. These features are implemented on a SCADA software platform as a distributed control system which integrates to the global control system of the ALICE experiment.

Keywords: AD0, ALICE, DCS, LHC

Procedia PDF Downloads 293
2889 The Study of Thai Millennial Attitude toward End-of-Life Planning, Opportunity of Service Design Development

Authors: Mawong R., Bussracumpakorn C.

Abstract:

Millions of young people around the world have been affected by COVID-19 to their psychological and social effects. Millennials’ stresses have been shaped by a few global issues, including climate change, political instability, and financial crisis. In particular, the spread of COVID-19 has become laying psychological and socioeconomic scars on them. When end-of-life planning turns into more widely discussed, the stigma and taboos around this issue are greatly lessened. End-of-life planning is defined as a future life plan, such as financial, legacy, funeral, and memorial planning. This plan would help millennials to discover the value and meaning of life. This study explores the attitudes of Thai Millennials toward end-of-life planning as a new normal awareness of life in order to initiate an innovative service concept to fit with their value and meaning. The study conducts an in-depth interview with 12 potential participants who have awareness or action on the plan. The framework of the customer journey map is used to analyze the responses to examine trigger points, barriers, beliefs, and expectations. The findings pointed to a service concept that is suggested for a new end-of-life planning service that is suited to Thai Millennials in 4 different groups, which are 1. Social -Conscious as a socially aware who to donate time and riches to make the world and society a better place, their end-of-life planning value is inspired by the social impact of giving something or some action that they will be able to do after life or during life which provides a variety of choice based on their preference to give to society, 2. Life Fulfillment who make a life goal for themselves and attempt to achieve it before the time comes to their value will be to inspire life value with a customized plan and provide guidance to suggest, 3. Prevention of the After-Death Effect who want to plan to avoid the effects of their death as patriarch, head of the family, and anchor of someone, so they want to have a plan that brings confidence and feel relief while they are still alive and they want to find some reliable service that they can leave the death will or asset, and 4. No Guilty Planning who plan for when they wish to be worry-free as a self-responsible they want to have the plan which is easy to understand and easy to access. The overall finding of the study is to understand the new service concept of end-of-life planning which to improve knowledge of significant life worth rather than death planning, encouraging people to reassess their lives in a positive way, leading to higher self-esteem and intrinsic motivation for this generation in this time of global crisis.

Keywords: design management, end-of-life planning, millennial generation, service design solution

Procedia PDF Downloads 174
2888 How to Modernise the European Competition Network (ECN)

Authors: Dorota Galeza

Abstract:

This paper argues that networks, such as the ECN and the American network, are affected by certain small events which are inherent to path dependence and preclude the full evolution towards efficiency. It is advocated that the American network is superior to the ECN in many respects due to its greater flexibility and longer history. This stems in particular from the creation of the American network, which was based on a small number of cases. Such a structure encourages further changes and modifications which are not necessarily radical. The ECN, by contrast, was established by legislative action, which explains its rigid structure and resistance to change. This paper is an attempt to transpose the superiority of the American network on to the ECN. It looks at concepts such as judicial cooperation, harmonisation of procedure, peer review and regulatory impact assessments (RIAs), and dispute resolution procedures.

Keywords: antitrust, competition, networks, path dependence

Procedia PDF Downloads 299
2887 Modelling of Moisture Loss and Oil Uptake during Deep-Fat Frying of Plantain

Authors: James A. Adeyanju, John O. Olajide, Akinbode A. Adedeji

Abstract:

A predictive mathematical model based on the fundamental principles of mass transfer was developed to simulate the moisture content and oil content during Deep-Fat Frying (DFF) process of dodo. The resulting governing equation, that is, partial differential equation that describes rate of moisture loss and oil uptake was solved numerically using explicit Finite Difference Technique (FDT). Computer codes were written in MATLAB environment for the implementation of FDT at different frying conditions and moisture loss as well as oil uptake simulation during DFF of dodo. Plantain samples were sliced into 5 mm thickness and fried at different frying oil temperatures (150, 160 and 170 ⁰C) for periods varying from 2 to 4 min. The comparison between the predicted results and experimental data for the validation of the model showed reasonable agreement. The correlation coefficients between the predicted and experimental values of moisture and oil transfer models ranging from 0.912 to 0.947 and 0.895 to 0.957, respectively. The predicted results could be further used for the design, control and optimization of deep-fat frying process.

Keywords: frying, moisture loss, modelling, oil uptake

Procedia PDF Downloads 428
2886 A Picture is worth a Billion Bits: Real-Time Image Reconstruction from Dense Binary Pixels

Authors: Tal Remez, Or Litany, Alex Bronstein

Abstract:

The pursuit of smaller pixel sizes at ever increasing resolution in digital image sensors is mainly driven by the stringent price and form-factor requirements of sensors and optics in the cellular phone market. Recently, Eric Fossum proposed a novel concept of an image sensor with dense sub-diffraction limit one-bit pixels (jots), which can be considered a digital emulation of silver halide photographic film. This idea has been recently embodied as the EPFL Gigavision camera. A major bottleneck in the design of such sensors is the image reconstruction process, producing a continuous high dynamic range image from oversampled binary measurements. The extreme quantization of the Poisson statistics is incompatible with the assumptions of most standard image processing and enhancement frameworks. The recently proposed maximum-likelihood (ML) approach addresses this difficulty, but suffers from image artifacts and has impractically high computational complexity. In this work, we study a variant of a sensor with binary threshold pixels and propose a reconstruction algorithm combining an ML data fitting term with a sparse synthesis prior. We also show an efficient hardware-friendly real-time approximation of this inverse operator. Promising results are shown on synthetic data as well as on HDR data emulated using multiple exposures of a regular CMOS sensor.

Keywords: binary pixels, maximum likelihood, neural networks, sparse coding

Procedia PDF Downloads 185
2885 Agricultural Knowledge Management System Design, Use, and Consequence for Knowledge Sharing and Integration

Authors: Dejen Alemu, Murray E. Jennex, Temtim Assefa

Abstract:

This paper is investigated to understand the design, the use, and the consequence of Knowledge Management System (KMS) for knowledge systems sharing and integration. A KMS for knowledge systems sharing and integration is designed to meet the challenges raised by knowledge management researchers and practitioners: the technical, the human, and social factors. Agricultural KMS involves various members coming from different Communities of Practice (CoPs) who possess their own knowledge of multiple practices which need to be combined in the system development. However, the current development of the technology ignored the indigenous knowledge of the local communities, which is the key success factor for agriculture. This research employed the multi-methodological approach to KMS research in action research perspective which consists of four strategies: theory building, experimentation, observation, and system development. Using the KMS development practice of Ethiopian agricultural transformation agency as a case study, this research employed an interpretive analysis using primary qualitative data acquired through in-depth semi-structured interviews and participant observations. The Orlikowski's structuration model of technology has been used to understand the design, the use, and the consequence of the KMS. As a result, the research identified three basic components for the architecture of the shared KMS, namely, the people, the resources, and the implementation subsystems. The KMS were developed using web 2.0 tools to promote knowledge sharing and integration among diverse groups of users in a distributed environment. The use of a shared KMS allows users to access diverse knowledge from a number of users in different groups of participants, enhances the exchange of different forms of knowledge and experience, and creates high interaction and collaboration among participants. The consequences of a shared KMS on the social system includes, the elimination of hierarchical structure, enhance participation, collaboration, and negotiation among users from different CoPs having common interest, knowledge and skill development, integration of diverse knowledge resources, and the requirement of policy and guideline. The research contributes methodologically for the application of system development action research for understanding a conceptual framework for KMS development and use. The research have also theoretical contribution in extending structuration model of technology for the incorporation of variety of knowledge and practical implications to provide management understanding in developing strategies for the potential of web 2.0 tools for sharing and integration of indigenous knowledge.

Keywords: communities of practice, indigenous knowledge, participation, structuration model of technology, Web 2.0 tools

Procedia PDF Downloads 239
2884 Spatial Time Series Models for Rice and Cassava Yields Based on Bayesian Linear Mixed Models

Authors: Panudet Saengseedam, Nanthachai Kantanantha

Abstract:

This paper proposes a linear mixed model (LMM) with spatial effects to forecast rice and cassava yields in Thailand at the same time. A multivariate conditional autoregressive (MCAR) model is assumed to present the spatial effects. A Bayesian method is used for parameter estimation via Gibbs sampling Markov Chain Monte Carlo (MCMC). The model is applied to the rice and cassava yields monthly data which have been extracted from the Office of Agricultural Economics, Ministry of Agriculture and Cooperatives of Thailand. The results show that the proposed model has better performance in most provinces in both fitting part and validation part compared to the simple exponential smoothing and conditional auto regressive models (CAR) from our previous study.

Keywords: Bayesian method, linear mixed model, multivariate conditional autoregressive model, spatial time series

Procedia PDF Downloads 381
2883 An Action Toolkit for Health Care Services Driving Disability Inclusion in Universal Health Coverage

Authors: Jill Hanass-Hancock, Bradley Carpenter, Samantha Willan, Kristin Dunkle

Abstract:

Access to quality health care for persons with disabilities is the litmus test in our strive toward universal health coverage. Persons with disabilities experience a variety of health disparities related to increased health risks, greater socioeconomic challenges, and persistent ableism in the provision of health care. In low- and middle-income countries, the support needed to address the diverse needs of persons with disabilities and close the gaps in inclusive and accessible health care can appear overwhelming to staff with little knowledge and tools available. An action-orientated disability inclusion toolkit for health facilities was developed through consensus-building consultations and field testing in South Africa. The co-creation of the toolkit followed a bottom-up approach with healthcare staff and persons with disabilities in two developmental cycles. In cycle one, a disability facility assessment tool was developed to increase awareness of disability accessibility and service delivery gaps in primary healthcare services in a simple and action-orientated way. In cycle two, an intervention menu was created, enabling staff to respond to identified gaps and improve accessibility and inclusion. Each cycle followed five distinct steps of development: a review of needs and existing tools, design of the draft tool, consensus discussion to adapt the tool, pilot-testing and adaptation of the tool, and identification of the next steps. The continued consultations, adaptations, and field-testing allowed the team to discuss and test several adaptations while co-creating a meaningful and feasible toolkit with healthcare staff and persons with disabilities. This approach led to a simplified tool design with ‘key elements’ needed to achieve universal health coverage: universal design of health facilities, reasonable accommodation, health care worker training, and care pathway linkages. The toolkit was adapted for paper or digital data entry, produces automated, instant facility reports, and has easy-to-use training guides and online modules. The cyclic approach enabled the team to respond to emerging needs. The pilot testing of the facility assessment tool revealed that healthcare workers took significant actions to change their facilities after an assessment. However, staff needed information on how to improve disability accessibility and inclusion, where to acquire accredited training, and how to improve disability data collection, referrals, and follow-up. Hence, intervention options were needed for each ‘key element’. In consultation with representatives from the health and disability sectors, tangible and feasible solutions/interventions were identified. This process included the development of immediate/low-cost and long-term solutions. The approach gained buy-in from both sectors, who called for including the toolkit in the standard quality assessments for South Africa’s health care services. Furthermore, the process identified tangible solutions for each ‘key element’ and highlighted where research and development are urgently needed. The cyclic and consultative approach enabled the development of a feasible facility assessment tool and a complementary intervention menu, moving facilities toward universal health coverage for and persons with disabilities in low- or better-resourced contexts while identifying gaps in the availability of interventions.

Keywords: public health, disability, accessibility, inclusive health care, universal health coverage

Procedia PDF Downloads 57
2882 Extending the AOP Joinpoint Model for Memory and Type Safety

Authors: Amjad Nusayr

Abstract:

Software security is a general term used to any type of software architecture or model in which security aspects are incorporated in this architecture. These aspects are not part of the main logic of the underlying program. Software security can be achieved using a combination of approaches, including but not limited to secure software designs, third part component validation, and secure coding practices. Memory safety is one feature in software security where we ensure that any object in memory has a valid pointer or a reference with a valid type. Aspect-Oriented Programming (AOP) is a paradigm that is concerned with capturing the cross-cutting concerns in code development. AOP is generally used for common cross-cutting concerns like logging and DB transaction managing. In this paper, we introduce the concepts that enable AOP to be used for the purpose of memory and type safety. We also present ideas for extending AOP in software security practices.

Keywords: aspect oriented programming, programming languages, software security, memory and type safety

Procedia PDF Downloads 112
2881 Robust Heart Rate Estimation from Multiple Cardiovascular and Non-Cardiovascular Physiological Signals Using Signal Quality Indices and Kalman Filter

Authors: Shalini Rankawat, Mansi Rankawat, Rahul Dubey, Mazad Zaveri

Abstract:

Physiological signals such as electrocardiogram (ECG) and arterial blood pressure (ABP) in the intensive care unit (ICU) are often seriously corrupted by noise, artifacts, and missing data, which lead to errors in the estimation of heart rate (HR) and incidences of false alarm from ICU monitors. Clinical support in ICU requires most reliable heart rate estimation. Cardiac activity, because of its relatively high electrical energy, may introduce artifacts in Electroencephalogram (EEG), Electrooculogram (EOG), and Electromyogram (EMG) recordings. This paper presents a robust heart rate estimation method by detection of R-peaks of ECG artifacts in EEG, EMG & EOG signals, using energy-based function and a novel Signal Quality Index (SQI) assessment technique. SQIs of physiological signals (EEG, EMG, & EOG) were obtained by correlation of nonlinear energy operator (teager energy) of these signals with either ECG or ABP signal. HR is estimated from ECG, ABP, EEG, EMG, and EOG signals from separate Kalman filter based upon individual SQIs. Data fusion of each HR estimate was then performed by weighing each estimate by the Kalman filters’ SQI modified innovations. The fused signal HR estimate is more accurate and robust than any of the individual HR estimate. This method was evaluated on MIMIC II data base of PhysioNet from bedside monitors of ICU patients. The method provides an accurate HR estimate even in the presence of noise and artifacts.

Keywords: ECG, ABP, EEG, EMG, EOG, ECG artifacts, Teager-Kaiser energy, heart rate, signal quality index, Kalman filter, data fusion

Procedia PDF Downloads 683
2880 Measurement Tools of the Maturity Model for IT Service Outsourcing in Higher Education Institutions

Authors: Victoriano Valencia García, Luis Usero Aragonés, Eugenio J. Fernández Vicente

Abstract:

Nowadays, the successful implementation of ICTs is vital for almost any kind of organization. Good governance and ICT management are essential for delivering value, managing technological risks, managing resources and performance measurement. In addition, outsourcing is a strategic IT service solution which complements IT services provided internally in organizations. This paper proposes the measurement tools of a new holistic maturity model based on standards ISO/IEC 20000 and ISO/IEC 38500, and the frameworks and best practices of ITIL and COBIT, with a specific focus on IT outsourcing. These measurement tools allow independent validation and practical application in the field of higher education, using a questionnaire, metrics tables, and continuous improvement plan tables as part of the measurement process. Guidelines and standards are proposed in the model for facilitating adaptation to universities and achieving excellence in the outsourcing of IT services.

Keywords: IT governance, IT management, IT services, outsourcing, maturity model, measurement tools

Procedia PDF Downloads 574
2879 Uncertainty of the Brazilian Earth System Model for Solar Radiation

Authors: Elison Eduardo Jardim Bierhals, Claudineia Brazil, Deivid Pires, Rafael Haag, Elton Gimenez Rossini

Abstract:

This study evaluated the uncertainties involved in the solar radiation projections generated by the Brazilian Earth System Model (BESM) of the Weather and Climate Prediction Center (CPTEC) belonging to Coupled Model Intercomparison Phase 5 (CMIP5), with the aim of identifying efficiency in the projections for solar radiation of said model and in this way establish the viability of its use. Two different scenarios elaborated by Intergovernmental Panel on Climate Change (IPCC) were evaluated: RCP 4.5 (with more optimistic contour conditions) and 8.5 (with more pessimistic initial conditions). The method used to verify the accuracy of the present model was the Nash coefficient and the Statistical bias, as it better represents these atmospheric patterns. The BESM showed a tendency to overestimate the data ​​of solar radiation projections in most regions of the state of Rio Grande do Sul and through the validation methods adopted by this study, BESM did not present a satisfactory accuracy.

Keywords: climate changes, projections, solar radiation, uncertainty

Procedia PDF Downloads 231
2878 Sentiment Classification of Documents

Authors: Swarnadip Ghosh

Abstract:

Sentiment Analysis is the process of detecting the contextual polarity of text. In other words, it determines whether a piece of writing is positive, negative or neutral.Sentiment analysis of documents holds great importance in today's world, when numerous information is stored in databases and in the world wide web. An efficient algorithm to illicit such information, would be beneficial for social, economic as well as medical purposes. In this project, we have developed an algorithm to classify a document into positive or negative. Using our algorithm, we obtained a feature set from the data, and classified the documents based on this feature set. It is important to note that, in the classification, we have not used the independence assumption, which is considered by many procedures like the Naive Bayes. This makes the algorithm more general in scope. Moreover, because of the sparsity and high dimensionality of such data, we did not use empirical distribution for estimation, but developed a method by finding degree of close clustering of the data points. We have applied our algorithm on a movie review data set obtained from IMDb and obtained satisfactory results.

Keywords: sentiment, Run's Test, cross validation, higher dimensional pmf estimation

Procedia PDF Downloads 383
2877 Improving Trainings of Mineral Processing Operators Through Gamification and Modelling and Simulation

Authors: Pedro A. S. Bergamo, Emilia S. Streng, Jan Rosenkranz, Yousef Ghorbani

Abstract:

Within the often-hazardous mineral industry, simulation training has speedily gained appreciation as an important method of increasing site safety and productivity through enhanced operator skill and knowledge. Performance calculations related to froth flotation, one of the most important concentration methods, is probably the hardest topic taught during the training of plant operators. Currently, most training teach those skills by traditional methods like slide presentations and hand-written exercises with a heavy focus on memorization. To optimize certain aspects of these pieces of training, we developed “MinFloat”, which teaches the operation formulas of the froth flotation process with the help of gamification. The simulation core based on a first-principles flotation model was implemented in Unity3D and an instructor tutoring system was developed, which presents didactic content and reviews the selected answers. The game was tested by 25 professionals with extensive experience in the mining industry based on a questionnaire formulated for training evaluations. According to their feedback, the game scored well in terms of quality, didactic efficacy and inspiring character. The feedback of the testers on the main target audience and the outlook of the mentioned solution is presented. This paper aims to provide technical background on the construction of educational games for the mining industry besides showing how feedback from experts can more efficiently be gathered thanks to new technologies such as online forms.

Keywords: training evaluation, simulation based training, modelling, and simulation, froth flotation

Procedia PDF Downloads 102
2876 Validation and Fit of a Biomechanical Bipedal Walking Model for Simulation of Loads Induced by Pedestrians on Footbridges

Authors: Dianelys Vega, Carlos Magluta, Ney Roitman

Abstract:

The simulation of loads induced by walking people in civil engineering structures is still challenging It has been the focus of considerable research worldwide in the recent decades due to increasing number of reported vibration problems in pedestrian structures. One of the most important key in the designing of slender structures is the Human-Structure Interaction (HSI). How moving people interact with structures and the effect it has on their dynamic responses is still not well understood. To rely on calibrated pedestrian models that accurately estimate the structural response becomes extremely important. However, because of the complexity of the pedestrian mechanisms, there are still some gaps in knowledge and more reliable models need to be investigated. On this topic several authors have proposed biodynamic models to represent the pedestrian, whether these models provide a consistent approximation to physical reality still needs to be studied. Therefore, this work comes to contribute to a better understanding of this phenomenon bringing an experimental validation of a pedestrian walking model and a Human-Structure Interaction model. In this study, a bi-dimensional bipedal walking model was used to represent the pedestrians along with an interaction model which was applied to a prototype footbridge. Numerical models were implemented in MATLAB. In parallel, experimental tests were conducted in the Structures Laboratory of COPPE (LabEst), at Federal University of Rio de Janeiro. Different test subjects were asked to walk at different walking speeds over instrumented force platforms to measure the walking force and an accelerometer was placed at the waist of each subject to measure the acceleration of the center of mass at the same time. By fitting the step force and the center of mass acceleration through successive numerical simulations, the model parameters are estimated. In addition, experimental data of a walking pedestrian on a flexible structure was used to validate the interaction model presented, through the comparison of the measured and simulated structural response at mid span. It was found that the pedestrian model was able to adequately reproduce the ground reaction force and the center of mass acceleration for normal and slow walking speeds, being less efficient for faster speeds. Numerical simulations showed that biomechanical parameters such as leg stiffness and damping affect the ground reaction force, and the higher the walking speed the greater the leg length of the model. Besides, the interaction model was also capable to estimate with good approximation the structural response, that remained in the same order of magnitude as the measured response. Some differences in frequency spectra were observed, which are presumed to be due to the perfectly periodic loading representation, neglecting intra-subject variabilities. In conclusion, this work showed that the bipedal walking model could be used to represent walking pedestrians since it was efficient to reproduce the center of mass movement and ground reaction forces produced by humans. Furthermore, although more experimental validations are required, the interaction model also seems to be a useful framework to estimate the dynamic response of structures under loads induced by walking pedestrians.

Keywords: biodynamic models, bipedal walking models, human induced loads, human structure interaction

Procedia PDF Downloads 117
2875 Blockchain’s Feasibility in Military Data Networks

Authors: Brenden M. Shutt, Lubjana Beshaj, Paul L. Goethals, Ambrose Kam

Abstract:

Communication security is of particular interest to military data networks. A relatively novel approach to network security is blockchain, a cryptographically secured distribution ledger with a decentralized consensus mechanism for data transaction processing. Recent advances in blockchain technology have proposed new techniques for both data validation and trust management, as well as different frameworks for managing dataflow. The purpose of this work is to test the feasibility of different blockchain architectures as applied to military command and control networks. Various architectures are tested through discrete-event simulation and the feasibility is determined based upon a blockchain design’s ability to maintain long-term stable performance at industry standards of throughput, network latency, and security. This work proposes a consortium blockchain architecture with a computationally inexpensive consensus mechanism, one that leverages a Proof-of-Identity (PoI) concept and a reputation management mechanism.

Keywords: blockchain, consensus mechanism, discrete-event simulation, fog computing

Procedia PDF Downloads 122
2874 Radiochemical Purity of 68Ga-BCA-Peptides: Separation of All 68Ga Species with a Single iTLC Strip

Authors: Anton A. Larenkov, Alesya Ya Maruk

Abstract:

In the present study, highly effective iTLC single strip method for the determination of radiochemical purity (RCP) of 68Ga-BCA-peptides was developed (with no double-developing, changing of eluents or other additional manipulation). In this method iTLC-SG strips and commonly used eluent TFAaq. (3-5 % (v/v)) are used. The method allows determining each of the key radiochemical forms of 68Ga (colloidal, bound, ionic) separately with the peaks separation being no less than 4 σ. Rf = 0.0-0.1 for 68Ga-colloid; Rf = 0.5-0.6 for 68Ga-BCA-peptides; Rf = 0.9-1.0 for ionic 68Ga. The method is simple and fast: For developing length of 75 mm only 4-6 min is required (versus 18-20 min for pharmacopoeial method). The method has been tested on various compounds (including 68Ga-DOTA-TOC, 68Ga-DOTA-TATE, 68Ga-NODAGA-RGD2 etc.). The cross-validation work for every specific form of 68Ga showed good correlation between method developed and control (pharmacopoeial) methods. The method can become convenient and much more informative replacement for pharmacopoeial methods, including HPLC.

Keywords: DOTA-TATE, 68Ga, quality control, radiochemical purity, radiopharmaceuticals, TLC

Procedia PDF Downloads 275
2873 The Critical Success Factors for Effective ICT Governance in Malaysian Public Sector: A Delphi Study

Authors: Rosida A. Razak, Mohamad Shanudin Zakaria

Abstract:

The fundamental issues in ICT Governance (ICTG) implementation for Malaysian Public Sector (MPS) is how ICT be applied to support improvements in productivity, management effectiveness and the quality of services offered to its citizens. Our main concern is to develop and adopt a common definition and framework to illustrate how ICTG can be used to better align ICT with government’s operations and strategic focus. In particular, we want to identify and categorize factors that drive a successful ICTG process. This paper presents the results of an exploratory study to identify, validate and refine such Critical Success Factors (CSFs) and confirmed seven CSFs and nineteen sub-factors as influential factors that fit MPS after further validated and refined. The Delphi method applied in validation and refining process before being endorsed as appropriate for MPS. The identified CSFs reflect the focus areas that need to be considered strategically to strengthen ICT Governance implementation and ensure business success.

Keywords: IT governance, critical success factors, productivity, CSFs

Procedia PDF Downloads 263
2872 Quality Teaching Evaluation Instrument: A Student Learning-centred Approach

Authors: Thuy T. T. Tran, Hamish Coates, Sophie Arkoudis

Abstract:

Evaluation instruments of teaching are abundant; however, these do not prompt any enhancement in the quality of teaching, not least because these instruments are framed only by teacher-centered conceptions of teaching. There is a need for more sophisticated teaching evaluation measures that focus on student learning and multi-stakeholder involvement. This study aims to develop such an evaluation instrument for Vietnamese higher education. The study uses several kinds of methods. The instrument was initially drafted through in-depth review of research, paying close attention to Vietnamese higher education. Draft evaluation instruments were produced and reviewed by 34 experts. The outcomes of this qualitative and quantitative data reveal an instrument that highlights the value of a multisource student-centered approach, and the rich integration of contextual and cultural traits where Confucian values are emphasized. The validation affirms that evaluating teaching in such way will facilitate the continuous learning growth of all stakeholders involved.

Keywords: multi stakeholders, quality teaching, student learning, teaching evaluation

Procedia PDF Downloads 290
2871 A New Gateway for Rheumatoid Arthritis: COXIBs with a Safety Cardiovascular Profile

Authors: Malvina Hoxha, Valerie Capra, Carola Buccellati, Angelo Sala, Clara Cena, Roberta Fruttero, Massimo Bertinaria, G. Enrico Rovati

Abstract:

Today COXIBs are used in the treatment of arthritis and many other painful conditions in selected patients with high gastrointestinal risk and low CV risk. Previously we found a new mechanism of action of a traditional NSAID (diclofenac) and a COXIB (lumiracoxib) that possess weak competitive antagonism at the TP receptor. We hypothesize that modifying the structure of a known specific inhibitor of cyclooxygenase-2 (COXIB), so that it becomes also a more potent thromboxane antagonist will preserve the anti-inflammatory and gastrointestinal safety typical of COXIBs and prevent the cardiovascular risk associated with long term therapy.

Keywords: cyclooxygenase, inflammation, lumiracoxib, thromboxane A2

Procedia PDF Downloads 285
2870 Determining Full Stage Creep Properties from Miniature Specimen Creep Test

Authors: W. Sun, W. Wen, J. Lu, A. A. Becker

Abstract:

In this work, methods for determining creep properties which can be used to represent the full life until failure from miniature specimen creep tests based on analytical solutions are presented. Examples used to demonstrate the application of the methods include a miniature rectangular thin beam specimen creep test under three-point bending and a miniature two-material tensile specimen creep test subjected to a steady load. Mathematical expressions for deflection and creep strain rate of the two specimens were presented for the Kachanov-Rabotnov creep damage model. On this basis, an inverse procedure was developed which has potential applications for deriving the full life creep damage constitutive properties from a very small volume of material, in particular, for various microstructure constitutive  regions, e.g. within heat-affected zones of power plant pipe weldments. Further work on validation and improvement of the method is addressed.

Keywords: creep damage property, miniature specimen, inverse approach, finite element modeling

Procedia PDF Downloads 218
2869 Bi-Criteria Vehicle Routing Problem for Possibility Environment

Authors: Bezhan Ghvaberidze

Abstract:

A multiple criteria optimization approach for the solution of the Fuzzy Vehicle Routing Problem (FVRP) is proposed. For the possibility environment the levels of movements between customers are calculated by the constructed simulation interactive algorithm. The first criterion of the bi-criteria optimization problem - minimization of the expectation of total fuzzy travel time on closed routes is constructed for the FVRP. A new, second criterion – maximization of feasibility of movement on the closed routes is constructed by the Choquet finite averaging operator. The FVRP is reduced to the bi-criteria partitioning problem for the so called “promising” routes which were selected from the all admissible closed routes. The convenient selection of the “promising” routes allows us to solve the reduced problem in the real-time computing. For the numerical solution of the bi-criteria partitioning problem the -constraint approach is used. An exact algorithm is implemented based on D. Knuth’s Dancing Links technique and the algorithm DLX. The Main objective was to present the new approach for FVRP, when there are some difficulties while moving on the roads. This approach is called FVRP for extreme conditions (FVRP-EC) on the roads. Also, the aim of this paper was to construct the solving model of the constructed FVRP. Results are illustrated on the numerical example where all Pareto-optimal solutions are found. Also, an approach for more complex model FVRP with time windows was developed. A numerical example is presented in which optimal routes are constructed for extreme conditions on the roads.

Keywords: combinatorial optimization, Fuzzy Vehicle routing problem, multiple objective programming, possibility theory

Procedia PDF Downloads 466
2868 Sub-Pixel Level Classification Using Remote Sensing For Arecanut Crop

Authors: S. Athiralakshmi, B.E. Bhojaraja, U. Pruthviraj

Abstract:

In agriculture, remote sensing is applied for monitoring of plant development, evaluating of physiological processes and growth conditions. Especially valuable are the spatio-temporal aspects of the remotely sensed data in detecting crop state differences and stress situations. In this study, hyperion imagery is used for classifying arecanut crops based on their age so that these maps can be used in yield estimation of crops, irrigation purposes, applying fertilizers etc. Traditional hard classifiers assigns the mixed pixels to the dominant classes. The proposed method uses a sub pixel level classifier called linear spectral unmixing available in ENVI software. It provides the relative abundance of surface materials and the context within a pixel that may be a potential solution to effectively identifying the land-cover distribution. Validation is done referring to field spectra collected using spectroradiometer and the ground control points obtained from GPS.

Keywords: FLAASH, Hyperspectral remote sensing, Linear Spectral Unmixing, Spectral Angle Mapper Classifier.

Procedia PDF Downloads 505
2867 Psychometric Properties and Factor Structure of the College Readiness Questionnaire

Authors: Muna Al-Kalbani, Thuwayba Al Barwani, Otherine Neisler, Hussain Alkharusi, David Clayton, Humaira Al-Sulaimani, Mohammad Khan, Hamad Al-Yahmadi

Abstract:

This study describes the psychometric properties and factor structure of the University Readiness Survey (URS). Survey data were collected from sample of 2652 students from Sultan Qaboos University. Exploratory factor analysis identified ten significant factors underlining the structure. The results of Confirmatory factor analysis showed a good fit to the data where the indices for the revised model were χ2(df = 1669) = 6093.4; CFI = 0.900; GFI =0.926; PCLOSE = 1.00 and RMSAE = 0.030 where each of these indices were above threshold. The overall value of Cronbach’s alpha was 0.899 indicating that the instrument score was reliable. Results imply that the URS is a valid measure describing the college readiness pattern among Sultan Qaboos University students and the Arabic version could be used by university counselors to identify students’ readiness factors. Nevertheless, further validation of the of the USR is recommended.

Keywords: college readiness, confirmatory factor analysis, reliability, validity

Procedia PDF Downloads 213
2866 Diagnostic Accuracy of the Tuberculin Skin Test for Tuberculosis Diagnosis: Interest of Using ROC Curve and Fagan’s Nomogram

Authors: Nouira Mariem, Ben Rayana Hazem, Ennigrou Samir

Abstract:

Background and aim: During the past decade, the frequency of extrapulmonary forms of tuberculosis has increased. These forms are under-diagnosed using conventional tests. The aim of this study was to evaluate the performance of the Tuberculin Skin Test (TST) for the diagnosis of tuberculosis, using the ROC curve and Fagan’s Nomogram methodology. Methods: This was a case-control, multicenter study in 11 anti-tuberculosis centers in Tunisia, during the period from June to November2014. The cases were adults aged between 18 and 55 years with confirmed tuberculosis. Controls were free from tuberculosis. A data collection sheet was filled out and a TST was performed for each participant. Diagnostic accuracy measures of TST were estimated using ROC curve and Area Under Curve to estimate sensitivity and specificity of a determined cut-off point. Fagan’s nomogram was used to estimate its predictive values. Results: Overall, 1053 patients were enrolled, composed of 339 cases (sex-ratio (M/F)=0.87) and 714 controls (sex-ratio (M/F)=0.99). The mean age was 38.3±11.8 years for cases and 33.6±11 years for controls. The mean diameter of the TST induration was significantly higher among cases than controls (13.7mm vs.6.2mm;p=10-6). Area Under Curve was 0.789 [95% CI: 0.758-0.819; p=0.01], corresponding to a moderate discriminating power for this test. The most discriminative cut-off value of the TST, which were associated with the best sensitivity (73.7%) and specificity (76.6%) couple was about 11 mm with a Youden index of 0.503. Positive and Negative predictive values were 3.11% and 99.52%, respectively. Conclusion: In view of these results, we can conclude that the TST can be used for tuberculosis diagnosis with a good sensitivity and specificity. However, the skin induration measurement and its interpretation is operator dependent and remains difficult and subjective. The combination of the TST with another test such as the Quantiferon test would be a good alternative.

Keywords: tuberculosis, tuberculin skin test, ROC curve, cut-off

Procedia PDF Downloads 54
2865 Normalized Enterprises Architectures: Portugal's Public Procurement System Application

Authors: Tiago Sampaio, André Vasconcelos, Bruno Fragoso

Abstract:

The Normalized Systems Theory, which is designed to be applied to software architectures, provides a set of theorems, elements and rules, with the purpose of enabling evolution in Information Systems, as well as ensuring that they are ready for change. In order to make that possible, this work’s solution is to apply the Normalized Systems Theory to the domain of enterprise architectures, using Archimate. This application is achieved through the adaptation of the elements of this theory, making them artifacts of the modeling language. The theorems are applied through the identification of the viewpoints to be used in the architectures, as well as the transformation of the theory’s encapsulation rules into architectural rules. This way, it is possible to create normalized enterprise architectures, thus fulfilling the needs and requirements of the business. This solution was demonstrated using the Portuguese Public Procurement System. The Portuguese government aims to make this system as fair as possible, allowing every organization to have the same business opportunities. The aim is for every economic operator to have access to all public tenders, which are published in any of the 6 existing platforms, independently of where they are registered. In order to make this possible, we applied our solution to the construction of two different architectures, which are able of fulfilling the requirements of the Portuguese government. One of those architectures, TO-BE A, has a Message Broker that performs the communication between the platforms. The other, TO-BE B, represents the scenario in which the platforms communicate with each other directly. Apart from these 2 architectures, we also represent the AS-IS architecture that demonstrates the current behavior of the Public Procurement Systems. Our evaluation is based on a comparison between the AS-IS and the TO-BE architectures, regarding the fulfillment of the rules and theorems of the Normalized Systems Theory and some quality metrics.

Keywords: archimate, architecture, broker, enterprise, evolvable systems, interoperability, normalized architectures, normalized systems, normalized systems theory, platforms

Procedia PDF Downloads 340
2864 Exploring Transitions between Communal- and Market-Based Knowledge Sharing

Authors: Benbya Hind, Belbaly Nassim

Abstract:

Markets and communities are often cast as alternative forms of knowledge sharing, but an open question is how and why people dynamically transition between them. To study these transitions, we design a technology that allows geographically distributed participants to either buy knowledge (using virtual points) or request it for free. We use a data-driven, inductive approach, studying 550 members in over 5000 interactions, during nine months. Because the technology offered participants choices between market or community forms, we can document both individual and collective transitions that emerge as people cycle between these forms. Our inductive analysis revealed that uncertainties endemic to knowledge sharing were the impetus for these transitions. Communities evoke uncertainties about knowledge sharing’s costs and benefits, which markets resolve by quantifying explicit prices. However, if people manipulate markets, they create uncertainties about the validity of those prices, allowing communities to reemerge to establish certainty via identity-based validation.

Keywords: knowledge sharing, communities, information technology design, transitions, markets

Procedia PDF Downloads 163
2863 Hysteresis Modeling in Iron-Dominated Magnets Based on a Deep Neural Network Approach

Authors: Maria Amodeo, Pasquale Arpaia, Marco Buzio, Vincenzo Di Capua, Francesco Donnarumma

Abstract:

Different deep neural network architectures have been compared and tested to predict magnetic hysteresis in the context of pulsed electromagnets for experimental physics applications. Modelling quasi-static or dynamic major and especially minor hysteresis loops is one of the most challenging topics for computational magnetism. Recent attempts at mathematical prediction in this context using Preisach models could not attain better than percent-level accuracy. Hence, this work explores neural network approaches and shows that the architecture that best fits the measured magnetic field behaviour, including the effects of hysteresis and eddy currents, is the nonlinear autoregressive exogenous neural network (NARX) model. This architecture aims to achieve a relative RMSE of the order of a few 100 ppm for complex magnetic field cycling, including arbitrary sequences of pseudo-random high field and low field cycles. The NARX-based architecture is compared with the state-of-the-art, showing better performance than the classical operator-based and differential models, and is tested on a reference quadrupole magnetic lens used for CERN particle beams, chosen as a case study. The training and test datasets are a representative example of real-world magnet operation; this makes the good result obtained very promising for future applications in this context.

Keywords: deep neural network, magnetic modelling, measurement and empirical software engineering, NARX

Procedia PDF Downloads 118
2862 Landslide Hazard Assessment Using Physically Based Mathematical Models in Agricultural Terraces at Douro Valley in North of Portugal

Authors: C. Bateira, J. Fernandes, A. Costa

Abstract:

The Douro Demarked Region (DDR) is a production Porto wine region. On the NE of Portugal, the strong incision of the Douro valley developed very steep slopes, organized with agriculture terraces, have experienced an intense and deep transformation in order to implement the mechanization of the work. The old terrace system, based on stone vertical wall support structure, replaced by terraces with earth embankments experienced a huge terrace instability. This terrace instability has important economic and financial consequences on the agriculture enterprises. This paper presents and develops cartographic tools to access the embankment instability and identify the area prone to instability. The priority on this evaluation is related to the use of physically based mathematical models and develop a validation process based on an inventory of the past embankment instability. We used the shallow landslide stability model (SHALSTAB) based on physical parameters such us cohesion (c’), friction angle(ф), hydraulic conductivity, soil depth, soil specific weight (ϱ), slope angle (α) and contributing areas by Multiple Flow Direction Method (MFD). A terraced area can be analysed by this models unless we have very detailed information representative of the terrain morphology. The slope angle and the contributing areas depend on that. We can achieve that propose using digital elevation models (DEM) with great resolution (pixel with 40cm side), resulting from a set of photographs taken by a flight at 100m high with pixel resolution of 12cm. The slope angle results from this DEM. In the other hand, the MFD contributing area models the internal flow and is an important element to define the spatial variation of the soil saturation. That internal flow is based on the DEM. That is supported by the statement that the interflow, although not coincident with the superficial flow, have important similitude with it. Electrical resistivity monitoring values which related with the MFD contributing areas build from a DEM of 1m resolution and revealed a consistent correlation. That analysis, performed on the area, showed a good correlation with R2 of 0,72 and 0,76 at 1,5m and 2m depth, respectively. Considering that, a DEM with 1m resolution was the base to model the real internal flow. Thus, we assumed that the contributing area of 1m resolution modelled by MFD is representative of the internal flow of the area. In order to solve this problem we used a set of generalized DEMs to build the contributing areas used in the SHALSTAB. Those DEMs, with several resolutions (1m and 5m), were built from a set of photographs with 50cm resolution taken by a flight with 5km high. Using this maps combination, we modelled several final maps of terrace instability and performed a validation process with the contingency matrix. The best final instability map resembles the slope map from a DEM of 40cm resolution and a MFD map from a DEM of 1m resolution with a True Positive Rate (TPR) of 0,97, a False Positive Rate of 0,47, Accuracy (ACC) of 0,53, Precision (PVC) of 0,0004 and a TPR/FPR ratio of 2,06.

Keywords: agricultural terraces, cartography, landslides, SHALSTAB, vineyards

Procedia PDF Downloads 160
2861 Mapping the Intrinsic Vulnerability of the Quaternary Aquifer of the Eastern Mitidja (Northern Algeria)

Authors: Abida Haddouche, Ahmed Chrif Toubal

Abstract:

The Neogene basin of the Eastern Mitidja, object of the study area, represents potential water resources and especially groundwater reserves. This water is an important economic; this resource is highly sensitive which need protection and preservation. Unfortunately, these waters are exposed to various forms of pollution, whether from urban, agricultural, industrial or merely accidental. This pollution is a permanent risk of limiting resource. In this context, the work aims to evaluate the intrinsic vulnerability of the aquifer to protect and preserve the quality of this resource. It will focus on the disposal of water and land managers a cartographic document accessible to locate the areas where the water has a high vulnerability. Vulnerability mapping of the Easter Mitidja quaternary aquifer is performed by applying three methods (DRASTIC, DRIST, and GOD). Comparison and validation results show that the DRASTIC method is the most suitable method for aquifer vulnerability of the study area.

Keywords: Aquifer of Mitidja, DRASTIC method, geographic information system (GIS), vulnerability mapping

Procedia PDF Downloads 370