Search results for: computer node
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2732

Search results for: computer node

2552 Numerical Analysis of the Flexural Behaviour of Concrete-Filled Rectangular Flange Girders

Authors: R. Al-Dujele, K. A. Cashell

Abstract:

A tubular flange girder is an I-shaped steel girder with either one of both of the usual flat flange plates replaced with a hollow section. Typically, these hollow sections are either rectangular or circular in shape. Concrete filled tubular flange girders (CFTFGs) are unconventional I-shaped beams that use a hollow structural section as the top flange which is filled with concrete. The resulting section offers very high lateral torsional buckling strength and stiffness compared with conventional steel I-beams of similar depth, width and weight, typically leading to a reduction in lateral bracing requirements. This paper is focussed on investigating the ultimate capacity of concrete filled rectangular tubular flange girders (CFRTFGs). These are complex members and their behaviour is governed by a number of inter-related parameters. The FE model is developed using ABAQUS software, 3-D finite element (FE) model for simply supported CFRTFGs subjected to two point loads applied at the third-span points is built. An initial geometrical imperfection of (L/1000), as well as geometrical and material nonlinearities, are introduced into the model, where L denotes the span of the girder. In this numerical model, the concrete and steel materials are modelled using eight-node solid and four-node shell elements, respectively. In addition to the FE model, simplified analytical expressions for the flexural capacity are also proposed, and the results are compared to those from the FE analyses. The analytical expressions, which are suitable for design, are also shown to be capable of providing an accurate depiction of the bending moment capacity.

Keywords: concrete-filled rectangular tubular flange girders, ultimate capacity, confining effect, finite element analysis

Procedia PDF Downloads 121
2551 A Pattern Practise for Awareness Educations on Information Security: Information Security Project

Authors: Fati̇h Apaydin

Abstract:

Education technology is an area which constantly changes and creates innovations. As an inevitable part of the changing circumstances, the societies who have a tendency to the improvements keep up with these innovations by using the methods and strategies which have been designed for education technology. At this point, education technology has taken the responsibility to help the individuals improve themselves and teach the effective teaching methods by filling the airs in theoretical information, information security and the practice. The technology which comes to the core of our lives by raising the importance of it day by day and it enforced its position in computer- based environments. As a result, ‘being ready for technological innovations, improvement on computer-based talent, information, ability and attitude’ doctrines have to be given. However, it is today quite hard to deal with the security and reinforcement of this information. The information which is got illegally gives harm to society from every aspect, especially education. This study includes how and to what extent to use these innovative appliances such as computers and the factor of information security of these appliances in computer-based education. As the use of computer is constantly becoming prevalent in our country, both education and computer will never become out of date, so how computer-based education affects our lives and the study of information security for this type of education are important topics.

Keywords: computer, information security, education, technology, development

Procedia PDF Downloads 559
2550 Leakage Current Analysis of FinFET Based 7T SRAM at 32nm Technology

Authors: Chhavi Saxena

Abstract:

FinFETs can be a replacement for bulk-CMOS transistors in many different designs. Its low leakage/standby power property makes FinFETs a desirable option for memory sub-systems. Memory modules are widely used in most digital and computer systems. Leakage power is very important in memory cells since most memory applications access only one or very few memory rows at a given time. As technology scales down, the importance of leakage current and power analysis for memory design is increasing. In this paper, we discover an option for low power interconnect synthesis at the 32nm node and beyond, using Fin-type Field-Effect Transistors (FinFETs) which are a promising substitute for bulk CMOS at the considered gate lengths. We consider a mechanism for improving FinFETs efficiency, called variable supply voltage schemes. In this paper, we’ve illustrated the design and implementation of FinFET based 4x4 SRAM cell array by means of one bit 7T SRAM. FinFET based 7T SRAM has been designed and analysis have been carried out for leakage current, dynamic power and delay. For the validation of our design approach, the output of FinFET SRAM array have been compared with standard CMOS SRAM and significant improvements are obtained in proposed model.

Keywords: FinFET, 7T SRAM cell, leakage current, delay

Procedia PDF Downloads 426
2549 Stochastic Multicast Routing Protocol for Flying Ad-Hoc Networks

Authors: Hyunsun Lee, Yi Zhu

Abstract:

Wireless ad-hoc network is a decentralized type of temporary machine-to-machine connection that is spontaneous or impromptu so that it does not rely on any fixed infrastructure and centralized administration. As unmanned aerial vehicles (UAVs), also called drones, have recently become more accessible and widely utilized in military and civilian domains such as surveillance, search and detection missions, traffic monitoring, remote filming, product delivery, to name a few. The communication between these UAVs become possible and materialized through Flying Ad-hoc Networks (FANETs). However, due to the high mobility of UAVs that may cause different types of transmission interference, it is vital to design robust routing protocols for FANETs. In this talk, the multicast routing method based on a modified stochastic branching process is proposed. The stochastic branching process is often used to describe an early stage of an infectious disease outbreak, and the reproductive number in the process is used to classify the outbreak into a major or minor outbreak. The reproductive number to regulate the local transmission rate is adapted and modified for flying ad-hoc network communication. The performance of the proposed routing method is compared with other well-known methods such as flooding method and gossip method based on three measures; average reachability, average node usage and average branching factor. The proposed routing method achieves average reachability very closer to flooding method, average node usage closer to gossip method, and outstanding average branching factor among methods. It can be concluded that the proposed multicast routing scheme is more efficient than well-known routing schemes such as flooding and gossip while it maintains high performance.

Keywords: Flying Ad-hoc Networks, Multicast Routing, Stochastic Branching Process, Unmanned Aerial Vehicles

Procedia PDF Downloads 85
2548 Comparison of Different Hydrograph Routing Techniques in XPSTORM Modelling Software: A Case Study

Authors: Fatema Akram, Mohammad Golam Rasul, Mohammad Masud Kamal Khan, Md. Sharif Imam Ibne Amir

Abstract:

A variety of routing techniques are available to develop surface runoff hydrographs from rainfall. The selection of runoff routing method is very vital as it is directly related to the type of watershed and the required degree of accuracy. There are different modelling softwares available to explore the rainfall-runoff process in urban areas. XPSTORM, a link-node based, integrated storm-water modelling software, has been used in this study for developing surface runoff hydrograph for a Golf course area located in Rockhampton in Central Queensland in Australia. Four commonly used methods, namely SWMM runoff, Kinematic wave, Laurenson, and Time-Area are employed to generate runoff hydrograph for design storm of this study area. In runoff mode of XPSTORM, the rainfall, infiltration, evaporation and depression storage for sub-catchments were simulated and the runoff from the sub-catchment to collection node was calculated. The simulation results are presented, discussed and compared. The total surface runoff generated by SWMM runoff, Kinematic wave and Time-Area methods are found to be reasonably close, which indicates any of these methods can be used for developing runoff hydrograph of the study area. Laurenson method produces a comparatively less amount of surface runoff, however, it creates highest peak of surface runoff among all which may be suitable for hilly region. Although the Laurenson hydrograph technique is widely acceptable surface runoff routing technique in Queensland (Australia), extensive investigation is recommended with detailed topographic and hydrologic data in order to assess its suitability for use in the case study area.

Keywords: ARI, design storm, IFD, rainfall temporal pattern, routing techniques, surface runoff, XPSTORM

Procedia PDF Downloads 427
2547 Maximization of Lifetime for Wireless Sensor Networks Based on Energy Efficient Clustering Algorithm

Authors: Frodouard Minani

Abstract:

Since last decade, wireless sensor networks (WSNs) have been used in many areas like health care, agriculture, defense, military, disaster hit areas and so on. Wireless Sensor Networks consist of a Base Station (BS) and more number of wireless sensors in order to monitor temperature, pressure, motion in different environment conditions. The key parameter that plays a major role in designing a protocol for Wireless Sensor Networks is energy efficiency which is a scarcest resource of sensor nodes and it determines the lifetime of sensor nodes. Maximizing sensor node’s lifetime is an important issue in the design of applications and protocols for Wireless Sensor Networks. Clustering sensor nodes mechanism is an effective topology control approach for helping to achieve the goal of this research. In this paper, the researcher presents an energy efficiency protocol to prolong the network lifetime based on Energy efficient clustering algorithm. The Low Energy Adaptive Clustering Hierarchy (LEACH) is a routing protocol for clusters which is used to lower the energy consumption and also to improve the lifetime of the Wireless Sensor Networks. Maximizing energy dissipation and network lifetime are important matters in the design of applications and protocols for wireless sensor networks. Proposed system is to maximize the lifetime of the Wireless Sensor Networks by choosing the farthest cluster head (CH) instead of the closest CH and forming the cluster by considering the following parameter metrics such as Node’s density, residual-energy and distance between clusters (inter-cluster distance). In this paper, comparisons between the proposed protocol and comparative protocols in different scenarios have been done and the simulation results showed that the proposed protocol performs well over other comparative protocols in various scenarios.

Keywords: base station, clustering algorithm, energy efficient, sensors, wireless sensor networks

Procedia PDF Downloads 109
2546 Learning Based on Computer Science Unplugged in Computer Science Education: Design, Development, and Assessment

Authors: Eiko Takaoka, Yoshiyuki Fukushima, Koichiro Hirose, Tadashi Hasegawa

Abstract:

Although all high school students in Japan are required to learn informatics, many of them do not learn this topic sufficiently. In response to this situation, we propose a support package for high school informatics classes. To examine what students learned and if they sufficiently understood the context of the lessons, a questionnaire survey was distributed to 186 students. We analyzed the results of the questionnaire and determined the weakest units, which were “basic computer configuration” and “memory and secondary storage”. We then developed a package for teaching these units. We propose that our package be applied in high school classrooms.

Keywords: computer science unplugged, computer science outreach, high school curriculum, experimental evaluation

Procedia PDF Downloads 358
2545 A Tutorial on Network Security: Attacks and Controls

Authors: Belbahi Ahlam

Abstract:

With the phenomenal growth in the Internet, network security has become an integral part of computer and information security. In order to come up with measures that make networks more secure, it is important to learn about the vulnerabilities that could exist in a computer network and then have an understanding of the typical attacks that have been carried out in such networks. The first half of this paper will expose the readers to the classical network attacks that have exploited the typical vulnerabilities of computer networks in the past and solutions that have been adopted since then to prevent or reduce the chances of some of these attacks. The second half of the paper will expose the readers to the different network security controls including the network architecture, protocols, standards and software/ hardware tools that have been adopted in modern day computer networks.

Keywords: network security, attacks and controls, computer and information, solutions

Procedia PDF Downloads 418
2544 The Current Use of Computer Technology in Arabic Language

Authors: Saad Alkahtani

Abstract:

This study aims to identify the extent to which the faculty members who teach Arabic to speakers of other languages in Arabic language institutes at Saudi universities use computer technologies such as language laboratories, websites, software programs, and learning management system (LMS). It also seeks to identify critical difficulties that hinder the use of these technologies by faculty members. The population of the study consisted of 103 faculty members in four Arabic language institutes at Saudi universities. The results of the study showed a disparity in the use of computer technologies in teaching Arabic to non-native speakers. The means of degree of use ranged from 1.20 through 2.83. The study also identified difficulties limiting the use of computer technology in teaching Arabic. And the means of averages of difficulty of use ranged from 1.50 to 2.89. The differences were not statistically significant among the institutes (at 0.05).

Keywords: Arabic language programs, computer technology, using technology in teaching Arabic language, Arabic as a second language, computer skills

Procedia PDF Downloads 429
2543 A Cloud-Based Federated Identity Management in Europe

Authors: Jesus Carretero, Mario Vasile, Guillermo Izquierdo, Javier Garcia-Blas

Abstract:

Currently, there is a so called ‘identity crisis’ in cybersecurity caused by the substantial security, privacy and usability shortcomings encountered in existing systems for identity management. Federated Identity Management (FIM) could be solution for this crisis, as it is a method that facilitates management of identity processes and policies among collaborating entities without enforcing a global consistency, that is difficult to achieve when there are ID legacy systems. To cope with this problem, the Connecting Europe Facility (CEF) initiative proposed in 2014 a federated solution in anticipation of the adoption of the Regulation (EU) N°910/2014, the so-called eIDAS Regulation. At present, a network of eIDAS Nodes is being deployed at European level to allow that every citizen recognized by a member state is to be recognized within the trust network at European level, enabling the consumption of services in other member states that, until now were not allowed, or whose concession was tedious. This is a very ambitious approach, since it tends to enable cross-border authentication of Member States citizens without the need to unify the authentication method (eID Scheme) of the member state in question. However, this federation is currently managed by member states and it is initially applied only to citizens and public organizations. The goal of this paper is to present the results of a European Project, named eID@Cloud, that focuses on the integration of eID in 5 cloud platforms belonging to authentication service providers of different EU Member States to act as Service Providers (SP) for private entities. We propose an initiative based on a private eID Scheme both for natural and legal persons. The methodology followed in the eID@Cloud project is that each Identity Provider (IdP) is subscribed to an eIDAS Node Connector, requesting for authentication, that is subscribed to an eIDAS Node Proxy Service, issuing authentication assertions. To cope with high loads, load balancing is supported in the eIDAS Node. The eID@Cloud project is still going on, but we already have some important outcomes. First, we have deployed the federation identity nodes and tested it from the security and performance point of view. The pilot prototype has shown the feasibility of deploying this kind of systems, ensuring good performance due to the replication of the eIDAS nodes and the load balance mechanism. Second, our solution avoids the propagation of identity data out of the native domain of the user or entity being identified, which avoids problems well known in cybersecurity due to network interception, man in the middle attack, etc. Last, but not least, this system allows to connect any country or collectivity easily, providing incremental development of the network and avoiding difficult political negotiations to agree on a single authentication format (which would be a major stopper).

Keywords: cybersecurity, identity federation, trust, user authentication

Procedia PDF Downloads 140
2542 Efficient Computer-Aided Design-Based Multilevel Optimization of the LS89

Authors: A. Chatel, I. S. Torreguitart, T. Verstraete

Abstract:

The paper deals with a single point optimization of the LS89 turbine using an adjoint optimization and defining the design variables within a CAD system. The advantage of including the CAD model in the design system is that higher level constraints can be imposed on the shape, allowing the optimized model or component to be manufactured. However, CAD-based approaches restrict the design space compared to node-based approaches where every node is free to move. In order to preserve a rich design space, we develop a methodology to refine the CAD model during the optimization and to create the best parameterization to use at each time. This study presents a methodology to progressively refine the design space, which combines parametric effectiveness with a differential evolutionary algorithm in order to create an optimal parameterization. In this manuscript, we show that by doing the parameterization at the CAD level, we can impose higher level constraints on the shape, such as the axial chord length, the trailing edge radius and G2 geometric continuity between the suction side and pressure side at the leading edge. Additionally, the adjoint sensitivities are filtered out and only smooth shapes are produced during the optimization process. The use of algorithmic differentiation for the CAD kernel and grid generator allows computing the grid sensitivities to machine accuracy and avoid the limited arithmetic precision and the truncation error of finite differences. Then, the parametric effectiveness is computed to rate the ability of a set of CAD design parameters to produce the design shape change dictated by the adjoint sensitivities. During the optimization process, the design space is progressively enlarged using the knot insertion algorithm which allows introducing new control points whilst preserving the initial shape. The position of the inserted knots is generally assumed. However, this assumption can hinder the creation of better parameterizations that would allow producing more localized shape changes where the adjoint sensitivities dictate. To address this, we propose using a differential evolutionary algorithm to maximize the parametric effectiveness by optimizing the location of the inserted knots. This allows the optimizer to gradually explore larger design spaces and to use an optimal CAD-based parameterization during the course of the optimization. The method is tested on the LS89 turbine cascade and large aerodynamic improvements in the entropy generation are achieved whilst keeping the exit flow angle fixed. The trailing edge and axial chord length, which are kept fixed as manufacturing constraints. The optimization results show that the multilevel optimizations were more efficient than the single level optimization, even though they used the same number of design variables at the end of the multilevel optimizations. Furthermore, the multilevel optimization where the parameterization is created using the optimal knot positions results in a more efficient strategy to reach a better optimum than the multilevel optimization where the position of the knots is arbitrarily assumed.

Keywords: adjoint, CAD, knots, multilevel, optimization, parametric effectiveness

Procedia PDF Downloads 86
2541 Emerging Research Trends in Routing Protocol for Wireless Sensor Network

Authors: Subhra Prosun Paul, Shruti Aggarwal

Abstract:

Now a days Routing Protocol in Wireless Sensor Network has become a promising technique in the different fields of the latest computer technology. Routing in Wireless Sensor Network is a demanding task due to the different design issues of all sensor nodes. Network architecture, no of nodes, traffic of routing, the capacity of each sensor node, network consistency, service value are the important factor for the design and analysis of Routing Protocol in Wireless Sensor Network. Additionally, internal energy, the distance between nodes, the load of sensor nodes play a significant role in the efficient routing protocol. In this paper, our intention is to analyze the research trends in different routing protocols of Wireless Sensor Network in terms of different parameters. In order to explain the research trends on Routing Protocol in Wireless Sensor Network, different data related to this research topic are analyzed with the help of Web of Science and Scopus databases. The data analysis is performed from global perspective-taking different parameters like author, source, document, country, organization, keyword, year, and a number of the publication. Different types of experiments are also performed, which help us to evaluate the recent research tendency in the Routing Protocol of Wireless Sensor Network. In order to do this, we have used Web of Science and Scopus databases separately for data analysis. We have observed that there has been a tremendous development of research on this topic in the last few years as it has become a very popular topic day by day.

Keywords: analysis, routing protocol, research trends, wireless sensor network

Procedia PDF Downloads 185
2540 Addressing Scheme for IOT Network Using IPV6

Authors: H. Zormati, J. Chebil, J. Bel Hadj Taher

Abstract:

The goal of this paper is to present an addressing scheme that allows for assigning a unique IPv6 address to each node in the Internet of Things (IoT) network. This scheme guarantees uniqueness by extracting the clock skew of each communication device and converting it into an IPv6 address. Simulation analysis confirms that the presented scheme provides reductions in terms of energy consumption, communication overhead and response time as compared to four studied addressing schemes Strong DAD, LEADS, SIPA and CLOSA.

Keywords: addressing, IoT, IPv6, network, nodes

Procedia PDF Downloads 262
2539 A Numerical Model for Simulation of Blood Flow in Vascular Networks

Authors: Houman Tamaddon, Mehrdad Behnia, Masud Behnia

Abstract:

An accurate study of blood flow is associated with an accurate vascular pattern and geometrical properties of the organ of interest. Due to the complexity of vascular networks and poor accessibility in vivo, it is challenging to reconstruct the entire vasculature of any organ experimentally. The objective of this study is to introduce an innovative approach for the reconstruction of a full vascular tree from available morphometric data. Our method consists of implementing morphometric data on those parts of the vascular tree that are smaller than the resolution of medical imaging methods. This technique reconstructs the entire arterial tree down to the capillaries. Vessels greater than 2 mm are obtained from direct volume and surface analysis using contrast enhanced computed tomography (CT). Vessels smaller than 2mm are reconstructed from available morphometric and distensibility data and rearranged by applying Murray’s Laws. Implementation of morphometric data to reconstruct the branching pattern and applying Murray’s Laws to every vessel bifurcation simultaneously, lead to an accurate vascular tree reconstruction. The reconstruction algorithm generates full arterial tree topography down to the first capillary bifurcation. Geometry of each order of the vascular tree is generated separately to minimize the construction and simulation time. The node-to-node connectivity along with the diameter and length of every vessel segment is established and order numbers, according to the diameter-defined Strahler system, are assigned. During the simulation, we used the averaged flow rate for each order to predict the pressure drop and once the pressure drop is predicted, the flow rate is corrected to match the computed pressure drop for each vessel. The final results for 3 cardiac cycles is presented and compared to the clinical data.

Keywords: blood flow, morphometric data, vascular tree, Strahler ordering system

Procedia PDF Downloads 243
2538 Integrated Teaching of Hardware Courses for the Undergraduates of Computer Science and Engineering to Attain Focused Outcomes

Authors: Namrata D. Hiremath, Mahalaxmi Bhille, P. G. Sunitha Hiremath

Abstract:

Computer systems play an integral role in all facets of the engineering profession. This calls for an understanding of the processor-level components of computer systems, their design and operation, and their impact on the overall performance of the systems. Systems users are always in need of faster, more powerful, yet cheaper computer systems. The focus of Computer Science engineering graduates is inclined towards software oriented base. To be an efficient programmer there is a need to understand the role of hardware architecture towards the same. It is essential for the students of Computer Science and Engineering to know the basic building blocks of any computing device and how the digital principles can be used to build them. Hence two courses Digital Electronics of 3 credits, which is associated with lab of 1.5 credits and Computer Organization of 5 credits, were introduced at the sophomore level. Activity was introduced with the objective to teach the hardware concepts to the students of Computer science engineering through structured lab. The students were asked to design and implement a component of a computing device using MultiSim simulation tool and build the same using hardware components. The experience of the activity helped the students to understand the real time applications of the SSI and MSI components. The impact of the activity was evaluated and the performance was measured. The paper explains the achievement of the ABET outcomes a, c and k.

Keywords: digital, computer organization, ABET, structured enquiry, course activity

Procedia PDF Downloads 463
2537 Computation and Validation of the Stress Distribution around a Circular Hole in a Slab Undergoing Plastic Deformation

Authors: Sherif D. El Wakil, John Rice

Abstract:

The aim of the current work was to employ the finite element method to model a slab, with a small hole across its width, undergoing plastic plane strain deformation. The computational model had, however, to be validated by comparing its results with those obtained experimentally. Since they were in good agreement, the finite element method can therefore be considered a reliable tool that can help gain better understanding of the mechanism of ductile failure in structural members having stress raisers. The finite element software used was ANSYS, and the PLANE183 element was utilized. It is a higher order 2-D, 8-node or 6-node element with quadratic displacement behavior. A bilinear stress-strain relationship was used to define the material properties, with constants similar to those of the material used in the experimental study. The model was run for several tensile loads in order to observe the progression of the plastic deformation region, and the stress concentration factor was determined in each case. The experimental study involved employing the visioplasticity technique, where a circular mesh (each circle was 0.5 mm in diameter, with 0.05 mm line thickness) was initially printed on the side of an aluminum slab having a small hole across its width. Tensile loading was then applied to produce a small increment of plastic deformation. Circles in the plastic region became ellipses, where the directions of the principal strains and stresses coincided with the major and minor axes of the ellipses. Next, we were able to determine the directions of the maximum and minimum shear stresses at the center of each ellipse, and the slip-line field was then constructed. We were then able to determine the stress at any point in the plastic deformation zone, and hence the stress concentration factor. The experimental results were found to be in good agreement with the analytical ones.

Keywords: finite element method to model a slab, slab undergoing plastic deformation, stress distribution around a circular hole, visioplasticity

Procedia PDF Downloads 298
2536 Analyzing the Perceptions of Emotions in Aesthetic Music

Authors: Abigail Wiafe, Charles Nutrokpor, Adelaide Oduro-Asante

Abstract:

The advancement of technology is rapidly making people more receptive to music as computer-generated music requires minimal human interventions. Though algorithms are applied to generate music, the human experience of emotions is still explored. Thus, this study investigates the emotions humans experience listening to computer-generated music that possesses aesthetic qualities. Forty-two subjects participated in the survey. The selection process was purely arbitrary since it was based on convenience. Subjects listened and evaluated the emotions experienced from the computer-generated music through an online questionnaire. The Likert scale was used to rate the emotional levels after the music listening experience. The findings suggest that computer-generated music possesses aesthetic qualities that do not affect subjects' emotions as long as they are pleased with the music. Furthermore, computer-generated music has unique creativity, and expressioneven though the music produced is meaningless, the computational models developed are unable to present emotional contents in music as humans do.

Keywords: aesthetic, algorithms, emotions, computer-generated music

Procedia PDF Downloads 99
2535 Heart Rate Variability Analysis for Early Stage Prediction of Sudden Cardiac Death

Authors: Reeta Devi, Hitender Kumar Tyagi, Dinesh Kumar

Abstract:

In present scenario, cardiovascular problems are growing challenge for researchers and physiologists. As heart disease have no geographic, gender or socioeconomic specific reasons; detecting cardiac irregularities at early stage followed by quick and correct treatment is very important. Electrocardiogram is the finest tool for continuous monitoring of heart activity. Heart rate variability (HRV) is used to measure naturally occurring oscillations between consecutive cardiac cycles. Analysis of this variability is carried out using time domain, frequency domain and non-linear parameters. This paper presents HRV analysis of the online dataset for normal sinus rhythm (taken as healthy subject) and sudden cardiac death (SCD subject) using all three methods computing values for parameters like standard deviation of node to node intervals (SDNN), square root of mean of the sequences of difference between adjacent RR intervals (RMSSD), mean of R to R intervals (mean RR) in time domain, very low-frequency (VLF), low-frequency (LF), high frequency (HF) and ratio of low to high frequency (LF/HF ratio) in frequency domain and Poincare plot for non linear analysis. To differentiate HRV of healthy subject from subject died with SCD, k –nearest neighbor (k-NN) classifier has been used because of its high accuracy. Results show highly reduced values for all stated parameters for SCD subjects as compared to healthy ones. As the dataset used for SCD patients is recording of their ECG signal one hour prior to their death, it is therefore, verified with an accuracy of 95% that proposed algorithm can identify mortality risk of a patient one hour before its death. The identification of a patient’s mortality risk at such an early stage may prevent him/her meeting sudden death if in-time and right treatment is given by the doctor.

Keywords: early stage prediction, heart rate variability, linear and non-linear analysis, sudden cardiac death

Procedia PDF Downloads 317
2534 The Effect of a Computer-Assisted Glycemic Surveillance Protocol on Nursing Workload

Authors: Özlem Canbolat, Sevgisun Kapucu

Abstract:

The aim of this study was to determine the effect of a computer-assisted glycemic surveillance protocol on nursing workload in intensive care unit. The study is completed in an Education and Research Hospital in Ankara with the attendance of volunteered 19 nurse who had been worked in reanimation unit. Nurses used the written protocol and computer-assisted glycemic surveillance protocol for glycemic follow-up approach of the intensive care patients. Nurses used the written protocol first in the glycemic follow-up of the patient, then used the computer-assisted protocol. (Nurses used the written protocol first, then the computer-assisted protocol in the glycemic follow-up of the patient). Less time was spent in glycemic control with computerized protocol than written protocol and this difference is statistically significant (p < 0.001). It was determined that the computerized protocol application was completed in about 10 seconds (25% shorter) than the written protocol implementation. The computer-assisted glycemic surveillance protocol was found to be more easy and appropriate by nurses and the satisfaction level of the users was higher than with written protocol. While 79% of the nurses find it confusing to implement the written protocol, 79% were satisfied with the use of computerized protocol.

Keywords: computer-assisted protocol, glycemic control, insulin infusion protocol, intensive care, nursing workload

Procedia PDF Downloads 188
2533 Comparison of the Indocyanine Green Dye Method versus the Combined Method of Indigo Carmine Blue Dye with Indocyanine Green Fluorescence Imaging for Sentinel Lymph Node Biopsy in Breast Conservative Therapy for Early Breast Cancer

Authors: Nobuyuki Takemoto, Ai Koyanagi, Masanori Yasuda, Hiroshi Yamamoto

Abstract:

Background: Fluorescence imaging (FI) is one of the methods to identify sentinel lymph nodes (SLNs). However, the procedure is technically complicated and requires procedural skills, as SLN biopsy must be conducted in dim light conditions. As an improved version of this method, we introduced a combined method (Combined mixed dye and fluorescence; CMF) consisting of indigo carmine blue dye and FI. The direct visualization of SLNs under shadowless surgical light conditions is facilitated by the addition of the blue dye. We compared the SLN detection rates of CMF with that of the indocyanine green (ICG) dye method (ICG-D). Methods: A total of 202 patients with stage ≤ IIA breast cancer who underwent breast conservative therapy with separate incision from January 2004 to February 2017 were reviewed. Details of the two methods are as follows: (1) ICG-D: 2ml of ICG (10mg) was used and the green-stained SLNs were resected via a 3-4cm axillary incision; (2) CMF: A combination of 1ml of ICG (5mg) and 1-3ml of indigo carmine (4-12mg) was used. Using Photodynamic Eye (PDE), a 1.5-2 cm incision was made near the point of disappearance of the fluorescence and SLNs with intermediate color of blue and green were resected. Results: There were 92 ICG-D and 110 CMF cases. CMF resulted in a significantly higher detection rate than ICG-D (96.4% vs. 83.7%; p=0.003). This difference was particularly notable in those aged ≥ 60 years (98.3% vs. 74.3%) and individuals with BMI ≥ 25kg/m2 (90.3% vs. 58.3%). Conclusion: CMF is an effective method to identify SLNs which is safe, efficient, and cost-effective. Furthermore, radiation exposure can be avoided, and it can be performed in institutes without nuclear medicine facilities. CMF achieves a high SLN identification rate, and most of this procedure is feasible under shadowless surgical light conditions. CMF can reliably perform SLN biopsy even in those aged ≥ 60 years and individuals with BMI ≥ 25 kg/m2.

Keywords: sentinel lymph node biopsy, identification rate, indocyanine green (ICG), indigocarmine, fluorescence

Procedia PDF Downloads 149
2532 A Cooperative Signaling Scheme for Global Navigation Satellite Systems

Authors: Keunhong Chae, Seokho Yoon

Abstract:

Recently, the global navigation satellite system (GNSS) such as Galileo and GPS is employing more satellites to provide a higher degree of accuracy for the location service, thus calling for a more efficient signaling scheme among the satellites used in the overall GNSS network. In that the network throughput is improved, the spatial diversity can be one of the efficient signaling schemes; however, it requires multiple antenna that could cause a significant increase in the complexity of the GNSS. Thus, a diversity scheme called the cooperative signaling was proposed, where the virtual multiple-input multiple-output (MIMO) signaling is realized with using only a single antenna in the transmit satellite of interest and with modeling the neighboring satellites as relay nodes. The main drawback of the cooperative signaling is that the relay nodes receive the transmitted signal at different time instants, i.e., they operate in an asynchronous way, and thus, the overall performance of the GNSS network could degrade severely. To tackle the problem, several modified cooperative signaling schemes were proposed; however, all of them are difficult to implement due to a signal decoding at the relay nodes. Although the implementation at the relay nodes could be simpler to some degree by employing the time-reversal and conjugation operations instead of the signal decoding, it would be more efficient if we could implement the operations of the relay nodes at the source node having more resources than the relay nodes. So, in this paper, we propose a novel cooperative signaling scheme, where the data signals are combined in a unique way at the source node, thus obviating the need of the complex operations such as signal decoding, time-reversal and conjugation at the relay nodes. The numerical results confirm that the proposed scheme provides the same performance in the cooperative diversity and the bit error rate (BER) as the conventional scheme, while reducing the complexity at the relay nodes significantly. Acknowledgment: This work was supported by the National GNSS Research Center program of Defense Acquisition Program Administration and Agency for Defense Development.

Keywords: global navigation satellite network, cooperative signaling, data combining, nodes

Procedia PDF Downloads 261
2531 The Attitude of High School Teachers in Saudi Arabia towards Computers: Qualitative Study

Authors: Manal O. Alothman, Judy Robertson

Abstract:

Teachers can play a huge role in encouraging students to use computers and can affect students’ attitudes towards computers. So understanding teachers’ beliefs and their use of computers is an important way to create effective motivational systems for teachers to use computers in the classroom in an effective way.A qualitative study (6 focus group) was carried out among Saudi High school teachers, both male and female, to examine their attitudes towards computers and to find out their computer skills and usage. The study showed a gender difference in that females were less likely to attend computer workshops, females also had less computer skills, and they have more negative attitudes towards computers than males. Also, the study found that low computer skills in the classroom made students unlikely to have the lessons presented using computers. Furthermore, the study found some factors that affected teachers’ attitudes towards computers. These factors were computer experience and confidence as much having skills and good experience in computer use, the role and importance of computers had become in their life and in teaching as well.

Keywords: attitude, education, student, teacher, technology

Procedia PDF Downloads 268
2530 The Investigation of Counselors Attitudes toward Online Counseling upon Taking Clients Perspective

Authors: Omer Ozer, Murat Yikilmaz, Ahmet Altinok, Ferhat Bayolu

Abstract:

There is an increasing number of online counseling services, studies exploring clients’ and counselors’ attitudes toward online counseling services are needed to provide effective and efficient mental health counseling services. The purpose of this study is to investigate counselors’ attitudes toward online counseling in relation to counselors’ genders, their daily usage of computer, their total usage of computer, and their self-efficacy in computer usage. In this study, Personal Information Form, specific items from the Online Counseling Attitudes Scale, and the Face-to-Face Counseling Attitudes Scale were given to 193 counselors to measure attitudes toward online counseling. Data were analyzed by using independent samples t-test and one-way ANOVA. There were no statistically significant differences counselors’ attitudes toward online counseling and counselors’ gender, their daily usage of computer, their total usage of computer, and their self-efficacy in computer usage. The implications of these findings have been discussed in the literature review to provide some suggestions to researchers in the counseling profession.

Keywords: online counseling, counselor, attitude, counseling service

Procedia PDF Downloads 321
2529 The Role of Synthetic Data in Aerial Object Detection

Authors: Ava Dodd, Jonathan Adams

Abstract:

The purpose of this study is to explore the characteristics of developing a machine learning application using synthetic data. The study is structured to develop the application for the purpose of deploying the computer vision model. The findings discuss the realities of attempting to develop a computer vision model for practical purpose, and detail the processes, tools, and techniques that were used to meet accuracy requirements. The research reveals that synthetic data represents another variable that can be adjusted to improve the performance of a computer vision model. Further, a suite of tools and tuning recommendations are provided.

Keywords: computer vision, machine learning, synthetic data, YOLOv4

Procedia PDF Downloads 194
2528 Fog Computing- Network Based Computing

Authors: Navaneeth Krishnan, Chandan N. Bhagwat, Aparajit P. Utpat

Abstract:

Cloud Computing provides us a means to upload data and use applications over the internet. As the number of devices connecting to the cloud grows, there is undue pressure on the cloud infrastructure. Fog computing or Network Based Computing or Edge Computing allows to move a part of the processing in the cloud to the network devices present along the node to the cloud. Therefore the nodes connected to the cloud have a better response time. This paper proposes a method of moving the computation from the cloud to the network by introducing an android like appstore on the networking devices.

Keywords: cloud computing, fog computing, network devices, appstore

Procedia PDF Downloads 347
2527 Analyzing the Street Pattern Characteristics on Young People’s Choice to Walk or Not: A Study Based on Accelerometer and Global Positioning Systems Data

Authors: Ebru Cubukcu, Gozde Eksioglu Cetintahra, Burcin Hepguzel Hatip, Mert Cubukcu

Abstract:

Obesity and overweight cause serious health problems. Public and private organizations aim to encourage walking in various ways in order to cope with the problem of obesity and overweight. This study aims to understand how the spatial characteristics of urban street pattern, connectivity and complexity influence young people’s choice to walk or not. 185 public university students in Izmir, the third largest city in Turkey, participated in the study. Each participant had worn an accelerometer and a global positioning (GPS) device for a week. The accelerometer device records data on the intensity of the participant’s activity at a specified time interval, and the GPS device on the activities’ locations. Combining the two datasets, activity maps are derived. These maps are then used to differentiate the participants’ walk trips and motor vehicle trips. Given that, the frequency of walk and motor vehicle trips are calculated at the street segment level, and the street segments are then categorized into two as ‘preferred by pedestrians’ and ‘preferred by motor vehicles’. Graph Theory-based accessibility indices are calculated to quantify the spatial characteristics of the streets in the sample. Six different indices are used: (I) edge density, (II) edge sinuosity, (III) eta index, (IV) node density, (V) order of a node, and (VI) beta index. T-tests show that the index values for the ‘preferred by pedestrians’ and ‘preferred by motor vehicles’ are significantly different. The findings indicate that the spatial characteristics of the street network have a measurable effect on young people’s choice to walk or not. Policy implications are discussed. This study is funded by the Scientific and Technological Research Council of Turkey, Project No: 116K358.

Keywords: graph theory, walkability, accessibility, street network

Procedia PDF Downloads 185
2526 Spectrum of Dry Eye Disease in Computer Users of Manipur India

Authors: Somorjeet Sharma Shamurailatpam, Rabindra Das, A. Suchitra Devi

Abstract:

Computer and video display users might complain about Asthenopia, burning, dry eyes etc. The management of dry eyes is often not in the lines of severity. Following systematic evaluation and grading, dry eye disease is one condition that can be practiced at all levels of ophthalmic care. In the present study, different spectrum causing dry eye and prevalence of dry eye disease in computer users of Manipur, India are determined with 600 individuals (300 cases and 300 control). Individuals between 15 and 50 years who used computers for more than 3 hrs a day for 1 year or more were included. Tear break up time (TBUT) and Schirmer’s test were conducted. It shows that 33 (20.4%) out of 164 males and 47 (30.3%) out of 136 females have dry eye. Possible explanation for the observed result is discussed.

Keywords: asthenopia, computer vision syndrome, dry eyes, Schirmer's test, TBUT

Procedia PDF Downloads 344
2525 Pathologies in the Left Atrium Reproduced Using a Low-Order Synergistic Numerical Model of the Cardiovascular System

Authors: Nicholas Pearce, Eun-jin Kim

Abstract:

Pathologies of the cardiovascular (CV) system remain a serious and deadly health problem for human society. Computational modelling provides a relatively accessible tool for diagnosis, treatment, and research into CV disorders. However, numerical models of the CV system have largely focused on the function of the ventricles, frequently overlooking the behaviour of the atria. Furthermore, in the study of the pressure-volume relationship of the heart, which is a key diagnosis of cardiac vascular pathologies, previous works often evoke popular yet questionable time-varying elastance (TVE) method that imposes the pressure-volume relationship instead of calculating it consistently. Despite the convenience of the TVE method, there have been various indications of its limitations and the need for checking its validity in different scenarios. A model of the combined left ventricle (LV) and left atrium (LA) is presented, which consistently considers various feedback mechanisms in the heart without having to use the TVE method. Specifically, a synergistic model of the left ventricle is extended and modified to include the function of the LA. The synergy of the original model is preserved by modelling the electro-mechanical and chemical functions of the micro-scale myofiber for the LA and integrating it with the microscale and macro-organ-scale heart dynamics of the left ventricle and CV circulation. The atrioventricular node function is included and forms the conduction pathway for electrical signals between the atria and ventricle. The model reproduces the essential features of LA behaviour, such as the two-phase pressure-volume relationship and the classic figure of eight pressure-volume loops. Using this model, disorders in the internal cardiac electrical signalling are investigated by recreating the mechano-electric feedback (MEF), which is impossible where the time-varying elastance method is used. The effects of AV node block and slow conduction are then investigated in the presence of an atrial arrhythmia. It is found that electrical disorders and arrhythmia in the LA degrade the CV system by reducing the cardiac output, power, and heart rate.

Keywords: cardiovascular system, left atrium, numerical model, MEF

Procedia PDF Downloads 81
2524 A Highly Efficient Broadcast Algorithm for Computer Networks

Authors: Ganesh Nandakumaran, Mehmet Karaata

Abstract:

A wave is a distributed execution, often made up of a broadcast phase followed by a feedback phase, requiring the participation of all the system processes before a particular event called decision is taken. Wave algorithms with one initiator such as the 1-wave algorithm have been shown to be very efficient for broadcasting messages in tree networks. Extensions of this algorithm broadcasting a sequence of waves using a single initiator have been implemented in algorithms such as the m-wave algorithm. However as the network size increases, having a single initiator adversely affects the message delivery times to nodes further away from the initiator. As a remedy, broadcast waves can be allowed to be initiated by multiple initiator nodes distributed across the network to reduce the completion time of broadcasts. These waves initiated by one or more initiator processes form a collection of waves covering the entire network. Solutions to global-snapshots, distributed broadcast and various synchronization problems can be solved efficiently using waves with multiple concurrent initiators. In this paper, we propose the first stabilizing multi-wave sequence algorithm implementing waves started by multiple initiator processes such that every process in the network receives at least one sequence of broadcasts. Due to being stabilizing, the proposed algorithm can withstand transient faults and do not require initialization. We view a fault as a transient fault if it perturbs the configuration of the system but not its program.

Keywords: distributed computing, multi-node broadcast, propagation of information with feedback and cleaning (PFC), stabilization, wave algorithms

Procedia PDF Downloads 472
2523 Justyna Skrzyńska, Zdzisław Kobos, Zbigniew Wochyński

Authors: Vahid Bairami Rad

Abstract:

Due to the tremendous progress in computer technology in the last decades, the capabilities of computers increased enormously and working with a computer became a normal activity for nearly everybody. With all the possibilities a computer can offer, humans and their interaction with computers are now a limiting factor. This gave rise to a lot of research in the field of HCI (human computer interaction) aiming to make interaction easier, more intuitive, and more efficient. To research eye gaze based interfaces it is necessary to understand both sides of the interaction–the human eye and the eye tracker. The first section gives an overview on the anatomy of the eye. The second section accuracy and calibration issue. The subsequent section presents data from a user study where eye movements have been recorded while watching a video and while surfing the Internet. Statistics on the eye movement during these tasks for several individuals provide typical values and ranges for fixation times and saccade lengths and are the foundation for discussions in later chapters. The data also reveal typical limitations of eye trackers.

Keywords: human computer interaction, gaze tracking, calibration, eye movement

Procedia PDF Downloads 507