Search results for: Finite element method to model a slab
1220 Integrating Microcontroller-Based Projects in a Human-Computer Interaction Course
Authors: Miguel Angel Garcia-Ruiz, Pedro Cesar Santana-Mancilla, Laura Sanely Gaytan-Lugo
Abstract:
This paper describes the design and application of a short in-class project conducted in Algoma University’s Human-Computer Interaction (HCI) course taught at the Bachelor of Computer Science. The project was based on the Maker Movement (people using and reusing electronic components and everyday materials to tinker with technology and make interactive applications), where students applied low-cost and easy-to-use electronic components, the Arduino Uno microcontroller board, software tools, and everyday objects. Students collaborated in small teams by completing hands-on activities with them, making an interactive walking cane for blind people. At the end of the course, students filled out a Technology Acceptance Model version 2 (TAM2) questionnaire where they evaluated microcontroller boards’ applications in HCI classes. We also asked them about applying the Maker Movement in HCI classes. Results showed overall students’ positive opinions and response about using microcontroller boards in HCI classes. We strongly suggest that every HCI course should include practical activities related to tinkering with technology such as applying microcontroller boards, where students actively and constructively participate in teams for achieving learning objectives.
Keywords: Maker movement, microcontrollers, learning, projects, course, technology acceptance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8581219 Design of Parity-Preserving Reversible Logic Signed Array Multipliers
Authors: Mojtaba Valinataj
Abstract:
Reversible logic as a new favorable design domain can be used for various fields especially creating quantum computers because of its speed and intangible power consumption. However, its susceptibility to a variety of environmental effects may lead to yield the incorrect results. In this paper, because of the importance of multiplication operation in various computing systems, some novel reversible logic array multipliers are proposed with error detection capability by incorporating the parity-preserving gates. The new designs are presented for two main parts of array multipliers, partial product generation and multi-operand addition, by exploiting the new arrangements of existing gates, which results in two signed parity-preserving array multipliers. The experimental results reveal that the best proposed 4×4 multiplier in this paper reaches 12%, 24%, and 26% enhancements in the number of constant inputs, number of required gates, and quantum cost, respectively, compared to previous design. Moreover, the best proposed design is generalized for n×n multipliers with general formulations to estimate the main reversible logic criteria as the functions of the multiplier size.Keywords: Array multipliers, Baugh-Wooley method, error detection, parity-preserving gates, quantum computers, reversible logic.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10261218 Anticancer Effect of Doxorubicin Loaded Heparin based Super-paramagnetic Iron oxide Nanoparticles against the Human Ovarian Cancer Cells
Authors: Amaneh Javid, Shahin Ahmadian, Ali A. Saboury, Saeed Rezaei-Zarchi
Abstract:
This study determines the effect of naked and heparinbased super-paramagnetic iron oxide nanoparticles on the human cancer cell lines of A2780. Doxorubicin was used as the anticancer drug, entrapped in the SPIO-NPs. This study aimed to decorate nanoparticles with heparin, a molecular ligand for 'active' targeting of cancerous cells and the application of modified-nanoparticles in cancer treatment. The nanoparticles containing the anticancer drug DOX were prepared by a solvent evaporation and emulsification cross-linking method. The physicochemical properties of the nanoparticles were characterized by various techniques, and uniform nanoparticles with an average particle size of 110±15 nm with high encapsulation efficiencies (EE) were obtained. Additionally, a sustained release of DOX from the SPIO-NPs was successful. Cytotoxicity tests showed that the SPIO-DOX-HP had higher cell toxicity than the individual HP and confocal microscopy analysis confirmed excellent cellular uptake efficiency. These results indicate that HP based SPIO-NPs have potential uses as anticancer drug carriers and also have an enhanced anticancer effect.Keywords: Heparin, A2780 cells, ovarian cancer, nanoparticles, doxorubicin.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24651217 From Vertigo to Verticality: An Example of Phenomenological Design in Architecture
Authors: E. Osorio Schmied
Abstract:
Architects commonly attempt a depiction of organic forms when their works are inspired by nature, regardless of the building site. Nevertheless it is also possible to try matching structures with natural scenery, by applying a phenomenological approach in terms of spatial operations, regarding perceptions from nature through architectural aspects such as protection, views, and orientation. This method acknowledges a relationship between place and space, where intentions towards tangible facts then become design statements. Although spaces resulting from such a process may present an effective response to the environment, they can also offer further outcomes beyond the realm of form. The hypothesis is that, in addition to recognising a bond between architecture and nature, it is also plausible to associate such perceptions with the inner ambient of buildings, by analysing features such as daylight. The case study of a single-family house in a rainforest near Valdivia, Chilean Patagonia is presented, with the intention of addressing the above notions through a discussion of the actual effects of inhabiting a place by way of a series of insights, including a revision of diagrams and photographs that assist in understanding the implications of this design practice. In addition, figures based on post-occupancy behaviour and daylighting performance relate both architectural and environmental issues to a decision-making process motivated by the observation of nature.
Keywords: Architecture, design statements, nature, perception.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10781216 Effect of Fractional Flow Curves on the Heavy Oil and Light Oil Recoveries in Petroleum Reservoirs
Authors: Abdul Jamil Nazari, Shigeo Honma
Abstract:
This paper evaluates and compares the effect of fractional flow curves on the heavy oil and light oil recoveries in a petroleum reservoir. Fingering of flowing water is one of the serious problems of the oil displacement by water and another problem is the estimation of the amount of recover oil from a petroleum reservoir. To address these problems, the fractional flow of heavy oil and light oil are investigated. The fractional flow approach treats the multi-phases flow rate as a total mixed fluid and then describes the individual phases as fractional of the total flow. Laboratory experiments are implemented for two different types of oils, heavy oil, and light oil, to experimentally obtain relative permeability and fractional flow curves. Application of the light oil fractional curve, which exhibits a regular S-shape, to the water flooding method showed that a large amount of mobile oil in the reservoir is displaced by water injection. In contrast, the fractional flow curve of heavy oil does not display an S-shape because of its high viscosity. Although the advance of the injected waterfront is faster than in light oil reservoirs, a significant amount of mobile oil remains behind the waterfront.
Keywords: Fractional flow curve, oil recovery, relative permeability, water fingering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14741215 Study of Chest Pain and its Risk Factors in Over 30 Year-Old Individuals
Authors: S. Dabiran
Abstract:
Chest pain is one of the most prevalent complaints among adults that cause the people to attend to medical centers. The aim was to determine the prevalence and risk factors of chest pain among over 30 years old people in Tehran. In this cross-sectional study, 787 adults took part from Apr 2005 until Apr 2006. The sampling method was random cluster sampling and there were 25 clusters. In each cluster, interviews were performed with 32 over 30 years old, people lived in those houses. In cases with chest pain, extra questions asked. The prevalence of CP was 9% (71 cases). Of them 21 cases (6.5%) were in 41-60 year age ranges and the remainders were over 61 year old. 19 cases (26.8%) mentioned CP in resting state and all of the cases had exertion onset CP. The CP duration was 10 minutes or less in all of the cases and in most of them (84.5%), the location of pain mentioned left anterior part of chest, left anterior part of sternum and or left arm. There was positive history of myocardial infarction in 12 cases (17%). There was significant relation between CP and age, sex and between history of myocardial infarction and marital state of study people. Our results are similar to other studies- results in most parts, however it is necessary to perform supplementary tests and follow up studies to differentiate between cardiac and non-cardiac CP exactly.Keywords: Chest pain, myocardial infarction, risk factor, prevalence
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14651214 Optimal Sliding Mode Controller for Knee Flexion During Walking
Authors: Gabriel Sitler, Yousef Sardahi, Asad Salem
Abstract:
This paper presents an optimal and robust sliding mode controller (SMC) to regulate the position of the knee joint angle for patients suffering from knee injuries. The controller imitates the role of active orthoses that produce the joint torques required to overcome gravity and loading forces and regain natural human movements. To this end, a mathematical model of the shank, the lower part of the leg, is derived first and then used for the control system design and computer simulations. The design of the controller is carried out in optimal and multi-objective settings. Four objectives are considered: minimization of the control effort and tracking error; and maximization of the control signal smoothness and closed-loop system’s speed of response. Optimal solutions in terms of the Pareto set and its image, the Pareto front, are obtained. The results show that there are trade-offs among the design objectives and many optimal solutions from which the decision-maker can choose to implement. Also, computer simulations conducted at different points from the Pareto set and assuming knee squat movement demonstrate competing relationships among the design goals. In addition, the proposed control algorithm shows robustness in tracking a standard gait signal when accounting for uncertainty in the shank’s parameters.
Keywords: Optimal control, multi-objective optimization, sliding mode control, wearable knee exoskeletons.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1831213 Survey of Access Controls in Cloud Computing
Authors: Monirah Alkathiry, Hanan Aljarwan
Abstract:
Cloud computing is one of the most significant technologies that the world deals with, in different sectors with different purposes and capabilities. The cloud faces various challenges in securing data from unauthorized access or modification. Consequently, security risks and levels have greatly increased. Therefore, cloud service providers (CSPs) and users need secure mechanisms that ensure that data are kept secret and safe from any disclosures or exploits. For this reason, CSPs need a number of techniques and technologies to manage and secure access to the cloud services to achieve security goals, such as confidentiality, integrity, identity access management (IAM), etc. Therefore, this paper will review and explore various access controls implemented in a cloud environment that achieve different security purposes. The methodology followed in this survey was conducting an assessment, evaluation, and comparison between those access controls mechanisms and technologies based on different factors, such as the security goals it achieves, usability, and cost-effectiveness. This assessment resulted in the fact that the technology used in an access control affects the security goals it achieves as well as there is no one access control method that achieves all security goals. Consequently, such a comparison would help decision-makers to choose properly the access controls that meet their requirements.Keywords: Access controls, cloud computing, confidentiality, identity and access management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7301212 A Comparative Analysis of Performance and QoS Issues in MANETs
Authors: Javed Parvez, Mushtaq Ahmad Peer
Abstract:
Mobile Ad hoc networks (MANETs) are collections of wireless mobile nodes dynamically reconfiguring and collectively forming a temporary network. These types of networks assume existence of no fixed infrastructure and are often useful in battle-field tactical operations or emergency search-and-rescue type of operations where fixed infrastructure is neither feasible nor practical. They also find use in ad hoc conferences, campus networks and commercial recreational applications carrying multimedia traffic. All of the above applications of MANETs require guaranteed levels of performance as experienced by the end-user. This paper focuses on key challenges in provisioning predetermined levels of such Quality of Service (QoS). It also identifies functional areas where QoS models are currently defined and used. Evolving functional areas where performance and QoS provisioning may be applied are also identified and some suggestions are provided for further research in this area. Although each of the above functional areas have been discussed separately in recent research studies, since these QoS functional areas are highly correlated and interdependent, a comprehensive and comparative analysis of these areas and their interrelationships is desired. In this paper we have attempted to provide such an overview.Keywords: Bandwidth Reservation, Congestion, DynamicNetwork Topology, End-to-End Delay, Flexible QoS Model forMANET(FQMM), Hidden Terminal, Mobile AdhocNetwork(MANET), Packet Jitter, Queuing, Quality-of-Service(QoS), Relative Bandwidth Service Differentiation(RBSD), Resource ReSerVation Protocol (RSVP).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21461211 Off-Policy Q-learning Technique for Intrusion Response in Network Security
Authors: Zheni S. Stefanova, Kandethody M. Ramachandran
Abstract:
With the increasing dependency on our computer devices, we face the necessity of adequate, efficient and effective mechanisms, for protecting our network. There are two main problems that Intrusion Detection Systems (IDS) attempt to solve. 1) To detect the attack, by analyzing the incoming traffic and inspect the network (intrusion detection). 2) To produce a prompt response when the attack occurs (intrusion prevention). It is critical creating an Intrusion detection model that will detect a breach in the system on time and also challenging making it provide an automatic and with an acceptable delay response at every single stage of the monitoring process. We cannot afford to adopt security measures with a high exploiting computational power, and we are not able to accept a mechanism that will react with a delay. In this paper, we will propose an intrusion response mechanism that is based on artificial intelligence, and more precisely, reinforcement learning techniques (RLT). The RLT will help us to create a decision agent, who will control the process of interacting with the undetermined environment. The goal is to find an optimal policy, which will represent the intrusion response, therefore, to solve the Reinforcement learning problem, using a Q-learning approach. Our agent will produce an optimal immediate response, in the process of evaluating the network traffic.This Q-learning approach will establish the balance between exploration and exploitation and provide a unique, self-learning and strategic artificial intelligence response mechanism for IDS.Keywords: Intrusion prevention, network security, optimal policy, Q-learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10221210 Sensitivity Comparison between Rapid Immuno-Chromatographic Device Test and ELISA in Detection and Sero-Prevalence of HBsAg and Anti-HCV antibodies in Apparently Healthy Blood Donors of Lahore, Pakistan
Authors: Natasha Hussain, Maleeha Aslam, Robina Farooq
Abstract:
Hepatitis B and hepatitis C are among the most significant hepatic infections all around the world that may lead to hepatocellular carcinoma. This study is first time performed at the blood transfussion centre of Omar hospital, Lahore. It aims to determine the sero-prevalence of these diseases by screening the apparently healthy blood donors who might be the carriers of HBV or HCV and pose a high risk in the transmission. It also aims the comparison between the sensitivity of two diagnostic tests; chromatographic immunoassay – one step test device and Enzyme Linked Immuno Sorbant Assay (ELISA). Blood serum of 855 apparently healthy blood donors was screened for Hepatitis B surface antigen (HBsAg) and for anti HCV antibodies. SPSS version 12.0 and X2 (Chi-square) test were used for statistical analysis. The seroprevalence of HCV was 8.07% by the device method and by ELISA 9.12% and that of HBV was 5.6% by the device and 6.43% by ELISA. The unavailability of vaccination against HCV makes it more prevalent. Comparing the two diagnostic methods, ELISA proved to be more sensitive.Keywords: ELISA, Sensitivity comparison of diagnostic tests, seroprevalence of Hepatitis B and C
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21981209 A Multigrid Approach for Three-Dimensional Inverse Heat Conduction Problems
Authors: Jianhua Zhou, Yuwen Zhang
Abstract:
A two-step multigrid approach is proposed to solve the inverse heat conduction problem in a 3-D object under laser irradiation. In the first step, the location of the laser center is estimated using a coarse and uniform grid system. In the second step, the front-surface temperature is recovered in good accuracy using a multiple grid system in which fine mesh is used at laser spot center to capture the drastic temperature rise in this region but coarse mesh is employed in the peripheral region to reduce the total number of sensors required. The effectiveness of the two-step approach and the multiple grid system are demonstrated by the illustrative inverse solutions. If the measurement data for the temperature and heat flux on the back surface do not contain random error, the proposed multigrid approach can yield more accurate inverse solutions. When the back-surface measurement data contain random noise, accurate inverse solutions cannot be obtained if both temperature and heat flux are measured on the back surface.
Keywords: Conduction, inverse problems, conjugated gradient method, laser.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8441208 On the Early Development of Dispersion in Flow through a Tube with Wall Reactions
Abstract:
This is a study on numerical simulation of the convection-diffusion transport of a chemical species in steady flow through a small-diameter tube, which is lined with a very thin layer made up of retentive and absorptive materials. The species may be subject to a first-order kinetic reversible phase exchange with the wall material and irreversible absorption into the tube wall. Owing to the velocity shear across the tube section, the chemical species may spread out axially along the tube at a rate much larger than that given by the molecular diffusion; this process is known as dispersion. While the long-time dispersion behavior, well described by the Taylor model, has been extensively studied in the literature, the early development of the dispersion process is by contrast much less investigated. By early development, that means a span of time, after the release of the chemical into the flow, that is shorter than or comparable to the diffusion time scale across the tube section. To understand the early development of the dispersion, the governing equations along with the reactive boundary conditions are solved numerically using the Flux Corrected Transport Algorithm (FCTA). The computation has enabled us to investigate the combined effects on the early development of the dispersion coefficient due to the reversible and irreversible wall reactions. One of the results is shown that the dispersion coefficient may approach its steady-state limit in a short time under the following conditions: (i) a high value of Damkohler number (say Da ≥ 10); (ii) a small but non-zero value of absorption rate (say Γ* ≤ 0.5).
Keywords: Dispersion coefficient, early development of dispersion, FCTA, wall reactions.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13391207 Study on Phytochemical Properties, Antibacterial Activity and Cytotoxicity of Aloe vera L.
Authors: K. Thu, Yin Y. Mon, Tin A. Khaing, Ohn M. Tun
Abstract:
The aim of the study was to investigate phytochemical properties, antimicrobial activity and cytotoxicity of Aloe vera. The phytochemical screening of the extracts of leaves of A. vera revealed the presence of bioactive compounds such as alkaloids, tannins, flavonoids phenolic compounds, and etc. with absence of cyanogenic glycosides. Three different solvents such as methanol, ethanol and Di-Methyl sulfoxide were used to screen the antimicrobial activity of A. vera leaves against four human clinical pathogens by agar well diffusion method. The maximum antibacterial activities were observed in methanol extract followed by ethanol and Di-Methyl sulfoxide. It was also found that remarkable antibacterial activities with methanolic and ethanolic extracts of A. vera compared with the standard antibiotic, tetracycline that was not active against E. coli and S. boydii and supported the view that A. vera is a potent antimicrobial agent compared with the conventional antibiotic. Moreover, the brine shrimps (Artemia salina) toxicity test exhibited LC50 value was 569.52 ppm. The resulting data indicated that the A. vera plant have less toxic effects on brine shrimp. Hence, it is signified that Aloe vera plant extract is safe to be used as an antimicrobial agent.Keywords: Aloe vera L., antimicrobial activity, brine shrimp, cytotoxicity, phytochemical properties.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 60511206 Applied Actuator Fault Accommodation in Flight Control Systems Using Fault Reconstruction Based FDD and SMC Reconfiguration
Authors: A. Ghodbane, M. Saad, J.-F. Boland, C. Thibeault
Abstract:
Historically, actuators’ redundancy was used to deal with faults occurring suddenly in flight systems. This technique was generally expensive, time consuming and involves increased weight and space in the system. Therefore, nowadays, the on-line fault diagnosis of actuators and accommodation plays a major role in the design of avionic systems. These approaches, known as Fault Tolerant Flight Control systems (FTFCs) are able to adapt to such sudden faults while keeping avionics systems lighter and less expensive. In this paper, a (FTFC) system based on the Geometric Approach and a Reconfigurable Flight Control (RFC) are presented. The Geometric approach is used for cosmic ray fault reconstruction, while Sliding Mode Control (SMC) based on Lyapunov stability theory is designed for the reconfiguration of the controller in order to compensate the fault effect. Matlab®/Simulink® simulations are performed to illustrate the effectiveness and robustness of the proposed flight control system against actuators’ faulty signal caused by cosmic rays. The results demonstrate the successful real-time implementation of the proposed FTFC system on a non-linear 6 DOF aircraft model.
Keywords: Actuators’ faults, Fault detection and diagnosis, Fault tolerant flight control, Sliding mode control, Geometric approach for fault reconstruction, Lyapunov stability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25761205 Clustering Mixed Data Using Non-normal Regression Tree for Process Monitoring
Authors: Youngji Yoo, Cheong-Sool Park, Jun Seok Kim, Young-Hak Lee, Sung-Shick Kim, Jun-Geol Baek
Abstract:
In the semiconductor manufacturing process, large amounts of data are collected from various sensors of multiple facilities. The collected data from sensors have several different characteristics due to variables such as types of products, former processes and recipes. In general, Statistical Quality Control (SQC) methods assume the normality of the data to detect out-of-control states of processes. Although the collected data have different characteristics, using the data as inputs of SQC will increase variations of data, require wide control limits, and decrease performance to detect outof- control. Therefore, it is necessary to separate similar data groups from mixed data for more accurate process control. In the paper, we propose a regression tree using split algorithm based on Pearson distribution to handle non-normal distribution in parametric method. The regression tree finds similar properties of data from different variables. The experiments using real semiconductor manufacturing process data show improved performance in fault detecting ability.Keywords: Semiconductor, non-normal mixed process data, clustering, Statistical Quality Control (SQC), regression tree, Pearson distribution system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17801204 An Intelligent Nondestructive Testing System of Ultrasonic Infrared Thermal Imaging Based on Embedded Linux
Authors: Hao Mi, Ming Yang, Tian-yue Yang
Abstract:
Ultrasonic infrared nondestructive testing is a kind of testing method with high speed, accuracy and localization. However, there are still some problems, such as the detection requires manual real-time field judgment, the methods of result storage and viewing are still primitive. An intelligent non-destructive detection system based on embedded linux is put forward in this paper. The hardware part of the detection system is based on the ARM (Advanced Reduced Instruction Set Computer Machine) core and an embedded linux system is built to realize image processing and defect detection of thermal images. The CLAHE algorithm and the Butterworth filter are used to process the thermal image, and then the boa server and CGI (Common Gateway Interface) technology are used to transmit the test results to the display terminal through the network for real-time monitoring and remote monitoring. The system also liberates labor and eliminates the obstacle of manual judgment. According to the experiment result, the system provides a convenient and quick solution for industrial non-destructive testing.Keywords: Remote monitoring, non-destructive testing, embedded linux system, image processing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9671203 Objectivity, Reliability and Validity of the 90º Push-Ups Test Protocol Among Male and Female Students of Sports Science Program
Authors: Ahmad Hashim, Mohd Sani Madon
Abstract:
This study was conducted to determine the objectivity, reliability and validity of the 90º push-ups test protocol among male and female students of Sports Science Program, Faculty of Sports Science and Coaching Sultan Idris University of Education. Samples (n = 300), consisted of males (n = 168) and females (n = 132) students were randomly selected for this study. Researchers tested the 90º push-ups on the sample twice in a single trial, test and re-test protocol in the bench press test. Pearson-Product Moment Correlation method's was used to determine the value of objectivity, reliability and validity testing. The findings showed that the 900 pushups test protocol showed high consistency between the two testers with a value of r = .99. Likewise, The reliability value between test and re-test for the 90º push-ups test for the male (r=.93) and female (r=.93) students was also high. The results showed a correlation between 90º push-ups test and bench press test for boys was r = .64 and girls was r = .28. This finding indicates that the use of the 90º push-ups to test muscular strength and endurance in the upper body of males has a higher validity values than female students.Keywords: Arm and shoulder girdle strength and endurance, 900 push-ups, bench press
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 92171202 An Implementation of MacMahon's Partition Analysis in Ordering the Lower Bound of Processing Elements for the Algorithm of LU Decomposition
Authors: Halil Snopce, Ilir Spahiu, Lavdrim Elmazi
Abstract:
A lot of Scientific and Engineering problems require the solution of large systems of linear equations of the form bAx in an effective manner. LU-Decomposition offers good choices for solving this problem. Our approach is to find the lower bound of processing elements needed for this purpose. Here is used the so called Omega calculus, as a computational method for solving problems via their corresponding Diophantine relation. From the corresponding algorithm is formed a system of linear diophantine equalities using the domain of computation which is given by the set of lattice points inside the polyhedron. Then is run the Mathematica program DiophantineGF.m. This program calculates the generating function from which is possible to find the number of solutions to the system of Diophantine equalities, which in fact gives the lower bound for the number of processors needed for the corresponding algorithm. There is given a mathematical explanation of the problem as well. Keywordsgenerating function, lattice points in polyhedron, lower bound of processor elements, system of Diophantine equationsand : calculus.
Keywords: generating function, lattice points in polyhedron, lower bound of processor elements, system of Diophantine equations and calculus.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14751201 Person-Environment Fit (PE Fit): Evidence from Brazil
Authors: Jucelia Appio, Danielle Deimling De Carli, Bruno Henrique Rocha Fernandes, Nelson Natalino Frizon
Abstract:
The purpose of this paper is to investigate if there are positive and significant correlations between the dimensions of Person-Environment Fit (Person-Job, Person-Organization, Person-Group and Person-Supervisor) at the “Best Companies to Work for” in Brazil in 2017. For that, a quantitative approach was used with a descriptive method being defined as a research sample the "150 Best Companies to Work for", according to data base collected in 2017 and provided by Fundação Instituto of Administração (FIA) of the University of São Paulo (USP). About the data analysis procedures, asymmetry and kurtosis, factorial analysis, Kaiser-Meyer-Olkin (KMO) tests, Bartlett sphericity and Cronbach's alpha were used for the 69 research variables, and as a statistical technique for the purpose of analyzing the hypothesis, Pearson's correlation analysis was performed. As a main result, we highlight that there was a positive and significant correlation between the dimensions of Person-Environment Fit, corroborating the H1 hypothesis that there is a positive and significant correlation between Person-Job Fit, Person-Organization Fit, Person-Group Fit and Person-Supervisor Fit.
Keywords: Human resource management, person-environment fit, strategic people management, best companies to work for.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10001200 New Features for Specific JPEG Steganalysis
Authors: Johann Barbier, Eric Filiol, Kichenakoumar Mayoura
Abstract:
We present in this paper a new approach for specific JPEG steganalysis and propose studying statistics of the compressed DCT coefficients. Traditionally, steganographic algorithms try to preserve statistics of the DCT and of the spatial domain, but they cannot preserve both and also control the alteration of the compressed data. We have noticed a deviation of the entropy of the compressed data after a first embedding. This deviation is greater when the image is a cover medium than when the image is a stego image. To observe this deviation, we pointed out new statistic features and combined them with the Multiple Embedding Method. This approach is motivated by the Avalanche Criterion of the JPEG lossless compression step. This criterion makes possible the design of detectors whose detection rates are independent of the payload. Finally, we designed a Fisher discriminant based classifier for well known steganographic algorithms, Outguess, F5 and Hide and Seek. The experiemental results we obtained show the efficiency of our classifier for these algorithms. Moreover, it is also designed to work with low embedding rates (< 10-5) and according to the avalanche criterion of RLE and Huffman compression step, its efficiency is independent of the quantity of hidden information.
Keywords: Compressed frequency domain, Fisher discriminant, specific JPEG steganalysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21621199 Application of Scientific Metrics to Evaluate Academic Reputation in Different Research Areas
Authors: Cristiano R. Cervi, Renata Galante, José Palazzo M. de Oliveira
Abstract:
In this paper, we address the problem of identifying academic reputation of researchers using scientific metrics in different research areas. Due to the characteristics of each area, researchers can present different behaviors. In previous work, we define Rep-Index that makes use of a profile template to individually identify the reputation of researchers. The Rep-Index is comprehensive and adaptive because involves hole trajectory of the researcher built throughout his career and can be used in different areas and in different contexts. Now, we compare our metric (Rep-Index) with the h-index and the g-index through experiments with researchers in the fields of Economics, Dentistry and Computer Science. We analyze the trajectory of 830 Brazilian researchers from the National Council of Technological and Scientific Development (CNPq), which receive grants research productivity. The grants are aimed at productivity researchers that stand out among their peers, enhancing their scientific normative criteria established by CNPq. Of the 830 researchers, 210 are in the area of Economics, 216 of Dentistry e 404 of Computer Science. The experiments show that our metric is strongly correlated with h-index, g-index and CNPq ranking. We also show good results for our hypothesis that our metric can be used to evaluate research in several areas. We apply our metric (Rep-Index) to compare the behavior of researchers in relation to their h-index and g-index through extensive experiments. The experiments showed that our metric is strongly correlated with h-index, g-index and CNPq ranking.
Keywords: Researcher reputation, profile model, scientific metrics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20011198 Motor Imagery Signal Classification for a Four State Brain Machine Interface
Authors: Hema C. R., Paulraj M. P., S. Yaacob, A. H. Adom, R. Nagarajan
Abstract:
Motor imagery classification provides an important basis for designing Brain Machine Interfaces [BMI]. A BMI captures and decodes brain EEG signals and transforms human thought into actions. The ability of an individual to control his EEG through imaginary mental tasks enables him to control devices through the BMI. This paper presents a method to design a four state BMI using EEG signals recorded from the C3 and C4 locations. Principle features extracted through principle component analysis of the segmented EEG are analyzed using two novel classification algorithms using Elman recurrent neural network and functional link neural network. Performance of both classifiers is evaluated using a particle swarm optimization training algorithm; results are also compared with the conventional back propagation training algorithm. EEG motor imagery recorded from two subjects is used in the offline analysis. From overall classification performance it is observed that the BP algorithm has higher average classification of 93.5%, while the PSO algorithm has better training time and maximum classification. The proposed methods promises to provide a useful alternative general procedure for motor imagery classification
Keywords: Motor Imagery, Brain Machine Interfaces, Neural Networks, Particle Swarm Optimization, EEG signal processing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24561197 In vitro Cytotoxic and Genotoxic Effects of Arsenic Trioxide on Human Keratinocytes
Authors: H. Bouaziz, M. Sefi, J. de Lapuente, M. Borras, N. Zeghal
Abstract:
Although, arsenic trioxide has been the subject of toxicological research, in vitro cytotoxicity and genotoxicity studies using relevant cell models and uniform methodology are not well elucidated. Hence, the aim of the present study was to evaluate the cytotoxicity and genotoxicity induced by arsenic trioxide in human keratinocytes (HaCaT) using the MTT [3-(4, 5-dimethylthiazol-2-yl)- 2,5-diphenyltetrazolium bromide] and alkaline single cell gel electrophoresis (Comet) assays, respectively. Human keratinocytes were treated with different doses of arsenic trioxide for 4 h prior to cytogenetic assessment. Data obtained from the MTT assay indicated that arsenic trioxide significantly reduced the viability of HaCaT cells in a dose-dependent manner, showing an IC50 value of 34.18 ± 0.6 μM. Data generated from the comet assay also indicated a significant dose-dependent increase in DNA damage in HaCaT cells associated with arsenic trioxide exposure. We observed a significant increase in comet tail length and tail moment, showing an evidence of arsenic trioxide -induced genotoxic damage in HaCaT cells. This study confirms that the comet assay is a sensitive and effective method to detect DNA damage caused by arsenic.
Keywords: Arsenic trioxide, cytotoxixity, genotoxicity, HaCaT.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22281196 The Pitch Diameter of Pipe Taper Thread Measurement and Uncertainty Using Three-Wire Probe
Authors: J. Kloypayan, W. Pimpakan
Abstract:
The pipe taper thread measurement and uncertainty normally used the four-wire probe according to the JIS B 0262. Besides, according to the EA-10/10 standard, the pipe thread could be measured using the three-wire probe. This research proposed to use the three-wire probe measuring the pitch diameter of the pipe taper thread. The measuring accessory component was designed and made, then, assembled to one side of the ULM 828 CiM machine. Therefore, this machine could be used to measure and calibrate both the pipe thread and the pipe taper thread. The equations and the expanded uncertainty for pitch diameter measurement were formulated. After the experiment, the results showed that the pipe taper thread had the pitch diameter equal to 19.165mm and the expanded uncertainty equal to 1.88µm. Then, the experiment results were compared to the results from the National Institute of Metrology Thailand. The equivalence ratio from the comparison showed that both results were related. Thus, the proposed method of using the three-wire probe measured the pitch diameter of the pipe taper thread was acceptable.
Keywords: Pipe taper thread, Three-wire probe, Measure and Calibration, The Universal length measuring machine.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 71051195 Application of De-Laval Nozzle Transonic Flow Field Computation Approaches
Abstract:
A supersonic expansion cannot be achieved within a convergent-divergent nozzle if the flow velocity does not reach that of the sound at the throat. The computation of the flow field characteristics at the throat is thus essential to the nozzle developed thrust value and therefore to the aircraft or rocket it propels. Several approaches were developed in order to describe the transonic expansion, which takes place through the throat of a De-Laval convergent-divergent nozzle. They all allow reaching good results but showing a major shortcoming represented by their inability to describe the transonic flow field for nozzles having a small throat radius. The approach initially developed by Kliegel & Levine uses the velocity series development in terms of the normalized throat radius added to unity instead of solely the normalized throat radius or the traditional small disturbances theory approach. The present investigation carries out the application of these three approaches for different throat radiuses of curvature. The method using the normalized throat radius added to unity shows better results when applied to geometries integrating small throat radiuses.
Keywords: De-Laval nozzles, transonic calculations, transonic flow, supersonic nozzle.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32861194 CFD Simulation for Air-Borne Infection Analysis in AII-Room
Authors: Young Kwon Yang, In Sung Kang, Jung Ha Hwang, Jin Chul Park
Abstract:
The present study is a foundational study for performance improvements on isolation wards to prevent proliferation of secondary infection of infectious diseases such as SARS, H1N1, and MERS inside hospitals. Accordingly, the present study conducted an analysis of the effect of sealing mechanisms and filling of openings on ensuring air tightness performance in isolation wards as well as simulation on air currents in improved isolation wards. The study method is as follows. First, previous studies on aerial infection type and mechanism were reviewed, and the review results were utilized as basic data of analysis on simulation of air current. Second, national and international legislations and regulations in relation to isolation wards as well as case studies on developed nations were investigated in order to identify the problems in isolation wards in Korea and improvement plans. Third, construction and facility plans were compared and analyzed between general and isolation wards focusing on large general hospitals in Korea, thereby conducting comparison and analysis on the performance and effects of air-tightness of general and isolation wards through CFD simulations. The study results showed that isolation wards had better air-tightness performance than that of general wards.
Keywords: AII Room, air-borne infection, CFD, computational fluid dynamics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15241193 Smart Side View Mirror Camera for Real Time System
Authors: Nunziata Ivana Guarneri, Arcangelo Bruna, Giuseppe Spampinato, Antonio Buemi
Abstract:
In the last decade, automotive companies have invested a lot in terms of innovation about many aspects regarding the automatic driver assistance systems. One innovation regards the usage of a smart camera placed on the car’s side mirror for monitoring the back and lateral road situation. A common road scenario is the overtaking of the preceding car and, in this case, a brief distraction or a loss of concentration can lead the driver to undertake this action, even if there is an already overtaking vehicle, leading to serious accidents. A valid support for a secure drive can be a smart camera system, which is able to automatically analyze the road scenario and consequentially to warn the driver when another vehicle is overtaking. This paper describes a method for monitoring the side view of a vehicle by using camera optical flow motion vectors. The proposed solution detects the presence of incoming vehicles, assesses their distance from the host car, and warns the driver through different levels of alert according to the estimated distance. Due to the low complexity and computational cost, the proposed system ensures real time performances.
Keywords: Camera calibration, ego motion, kalman filters, object tracking, real time systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9021192 On Determining the Most Effective Technique Available in Software Testing
Authors: Qasim Zafar, Matthew Anderson, Esteban Garcia, Steven Drager
Abstract:
Software failures can present an enormous detriment to people's lives and cost millions of dollars to repair when they are unexpectedly encountered in the wild. Despite a significant portion of the software development lifecycle and resources are dedicated to testing, software failures are a relatively frequent occurrence. Nevertheless, the evaluation of testing effectiveness remains at the forefront of ensuring high-quality software and software metrics play a critical role in providing valuable insights into quantifiable objectives to assess the level of assurance and confidence in the system. As the selection of appropriate metrics can be an arduous process, the goal of this paper is to shed light on the significance of software metrics by examining a range of testing techniques and metrics as well as identifying key areas for improvement. In doing so, this paper presents a method to compare the effectiveness of testing techniques with heterogeneous output metrics. Additionally, through this investigation, readers will gain a deeper understanding of how metrics can help to drive informed decision-making on delivering high-quality software and facilitate continuous improvement in testing practices.
Keywords: Software testing, software metrics, testing effectiveness, black box testing, random testing, adaptive random testing, combinatorial testing, fuzz testing, equivalence partition, boundary value analysis, white box testings.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 671191 Comparative Spatial Analysis of a Re-arranged Hospital Building
Authors: Burak Köken, Hatice D. Arslan, Bilgehan Y. Çakmak
Abstract:
Analyzing the relation networks between the hospital buildings which have complex structure and distinctive spatial relationships is quite difficult. The hospital buildings which require specialty in spatial relationship solutions during design and selfinnovation through the developing technology should survive and keep giving service even after the disasters such as earthquakes. In this study, a hospital building where the load-bearing system was strengthened because of the insufficient earthquake performance and the construction of an additional building was required to meet the increasing need for space was discussed and a comparative spatial evaluation of the hospital building was made with regard to its status before the change and after the change. For this reason, spatial organizations of the building before change and after the change were analyzed by means of Space Syntax method and the effects of the change on space organization parameters were searched by applying an analytical procedure. Using Depthmap UCL software, Connectivity, Visual Mean Depth, Beta and Visual Integration analyses were conducted. Based on the data obtained after the analyses, it was seen that the relationships between spaces of the building increased after the change and the building has become more explicit and understandable for the occupants. Furthermore, it was determined according to findings of the analysis that the increase in depth causes difficulty in perceiving the spaces and the changes considering this problem generally ease spatial use.Keywords: Architecture, hospital building, space syntax, strengthening.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2209