Search results for: Data stream analysis.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 13651

Search results for: Data stream analysis.

8101 Analysis of Patterns in TV Commercials that Recognize NGO Image

Authors: J. Areerut, F. Samuel

Abstract:

The purpose of this research is to analyze the pattern of television commercials and how they encourage non-governmental organizations to build their image in Thailand. It realizes how public relations can impact an organization's image. It is a truth that bad public relations management can cause hurt a reputation. On the other hand, a very small amount of work in public relations helps your organization to be recognized broadly and eventually accepted even wider. The main idea in this paper is to study and analyze patterns of television commercials that could impact non-governmental organization's images in a greater way. This research uses questionnaires and content analysis to summarize results. The findings show the aspects of how patterns of television commercials that are suited to non-governmental organization work in Thailand. It will be useful for any non-governmental organization that wishes to build their image through television commercials and also for further work based on this research.

Keywords: Television Commercial (TVC), Organization Image, Non-Governmental Organization: NGO, Public Relation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2375
8100 Dynamic Modeling of a Robot for Playing a Curved 3D Percussion Instrument Utilizing a Finite Element Method

Authors: Prakash Persad, Kelvin Loutan, Jr., Trichelle Seepersad

Abstract:

The Finite Element Method is commonly used in the analysis of flexible manipulators to predict elastic displacements and develop joint control schemes for reducing positioning error. In order to preserve simplicity, regular geometries, ideal joints and connections are assumed. This paper presents the dynamic FE analysis of a 4- degrees of freedom open chain manipulator, intended for striking a curved 3D surface percussion musical instrument. This was done utilizing the new MultiBody Dynamics Module in COMSOL, capable of modeling the elastic behavior of a body undergoing rigid body type motion.

Keywords: Dynamic modeling, Entertainment robots, Finite element method, Flexible robot manipulators, Multibody dynamics, Musical robots.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2252
8099 Generic Model for Timetabling Problems by Integer Linear Programming Approach

Authors: N. A. H. Aizam, V. Uvaraja

Abstract:

The agenda of showing the scheduled time for performing certain tasks is known as timetabling. It is widely used in many departments such as transportation, education, and production. Some difficulties arise to ensure all tasks happen in the time and place allocated. Therefore, many researchers invented various programming models to solve the scheduling problems from several fields. However, the studies in developing the general integer programming model for many timetabling problems are still questionable. Meanwhile, this thesis describes about creating a general model which solves different types of timetabling problems by considering the basic constraints. Initially, the common basic constraints from five different fields are selected and analyzed. A general basic integer programming model was created and then verified by using the medium set of data obtained randomly which is much similar to realistic data. The mathematical software, AIMMS with CPLEX as a solver has been used to solve the model. The model obtained is significant in solving many timetabling problems easily since it is modifiable to all types of scheduling problems which have same basic constraints.

Keywords: AIMMS mathematical software, integer linear programming, scheduling problems, timetabling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3023
8098 A Secure Auditing Framework for Load Balancing in Cloud Environment

Authors: R. Geetha, T. Padmavathy

Abstract:

Security audit is an important aspect or feature to be considered in cloud service customer. It is basically a certification process to audit the controls that deliver the security requirements. Security audits are conducted by trained and qualified staffs that belong to an independent auditing organization. Security audits must be carried as a standard of security controls. Proper check to be made that the cloud user has a proper reporting and logging facilities with the customer's system and hence ensuring appropriate business and operational flow of data through cloud service. We propose a cloud-based secure auditing framework, which enables confided in power to safely store their mystery information on the semi-believed cloud specialist co-ops, and specifically share their mystery information with a wide scope of information recipient, to diminish the key administration intricacy for power proprietors and information collectors. Unique in relation to past cloud-based information framework, data proprietors transfer their mystery information into cloud utilizing static and dynamic evaluating plan. Another propelled determination is, if any information beneficiary needs individual record to download, the information collector will send the solicitation to the expert. The specialist proprietor has the Access Control. At the off probability, the businessman must impart the primary record to the knowledge collector, acknowledge statistics beneficiary solicitation. Once the acknowledgement for the records is over, the recipient downloads the first record and this record shifting time with date and downloading time with date are monitored by the inspector. In addition to deduplication concept, diminished cloud memory area using dynamic document distribution has been proposed.

Keywords: Cloud computing, cloud storage auditing, data integrity, key exposure.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1155
8097 Analysis of Detecting Wormhole Attack in Wireless Networks

Authors: Khin Sandar Win

Abstract:

In multi hop wireless systems, such as ad hoc and sensor networks, mobile ad hoc network applications are deployed, security emerges as a central requirement. A particularly devastating attack is known as the wormhole attack, where two or more malicious colluding nodes create a higher level virtual tunnel in the network, which is employed to transport packets between the tunnel end points. These tunnels emulate shorter links in the network. In which adversary records transmitted packets at one location in the network, tunnels them to another location, and retransmits them into the network. The wormhole attack is possible even if the attacker has not compromised any hosts and even if all communication provides authenticity and confidentiality. In this paper, we analyze wormhole attack nature in ad hoc and sensor networks and existing methods of the defending mechanism to detect wormhole attacks without require any specialized hardware. This analysis able to provide in establishing a method to reduce the rate of refresh time and the response time to become more faster.

Keywords: Ad hoc network, Sensor network, Wormhole attack, defending mechanism.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2290
8096 Statistical Wavelet Features, PCA, and SVM Based Approach for EEG Signals Classification

Authors: R. K. Chaurasiya, N. D. Londhe, S. Ghosh

Abstract:

The study of the electrical signals produced by neural activities of human brain is called Electroencephalography. In this paper, we propose an automatic and efficient EEG signal classification approach. The proposed approach is used to classify the EEG signal into two classes: epileptic seizure or not. In the proposed approach, we start with extracting the features by applying Discrete Wavelet Transform (DWT) in order to decompose the EEG signals into sub-bands. These features, extracted from details and approximation coefficients of DWT sub-bands, are used as input to Principal Component Analysis (PCA). The classification is based on reducing the feature dimension using PCA and deriving the supportvectors using Support Vector Machine (SVM). The experimental are performed on real and standard dataset. A very high level of classification accuracy is obtained in the result of classification.

Keywords: Discrete Wavelet Transform, Electroencephalogram, Pattern Recognition, Principal Component Analysis, Support Vector Machine.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3106
8095 Academic Achievement Differences in Grandiose and Vulnerable Narcissists and the Mediating Effects of Self-Esteem and Self-Efficacy

Authors: Amber L. Dummett, Efstathia Tzemou

Abstract:

Narcissism is a personality trait characterised by selfishness, entitlement, and superiority. Narcissism is split into two subtypes, grandiose narcissism (GN) and vulnerable narcissism (VN). Grandiose narcissists are extraverted and arrogant, while vulnerable narcissists are introverted and insecure. This study investigates the psychological mechanisms that lead to differences in academic achievement (AA) between grandiose and vulnerable narcissists, specifically the mediating effects of self-esteem and self-efficacy. While narcissism is considered to be a negative trait, this study considers if better AA is one of them. Moreover, further research into VN is essential to fully compare and contrast it with GN. We hypothesise that grandiose narcissists achieve higher marks due to having high self-esteem which in turn boosts their sense of self-efficacy. In comparison, we hypothesise that vulnerable narcissists underperform due to having low self-esteem which limits their self-efficacy. Two online surveys were distributed to undergraduate university students. The first was a collection of scales measuring the mentioned dimensions, and the second investigated end of year AA. Sequential mediation analyses were conducted using the gathered data. Our analysis shows that neither self-esteem nor self-efficacy mediate the relationship between GN and AA. GN positively predicts self-esteem but has no relationship with self-efficacy. Self-esteem does not mediate the relationship between VN and AA. VN has a negative indirect effect on AA via self-efficacy, and VN negatively predicts self-esteem. Self-efficacy positively predicts AA. GN does not affect AA through the mediation of self-esteem and then self-efficacy, and neither does VN in this way. Overall, having grandiose or vulnerable narcissistic traits does not affect students’ AA. However, being highly efficacious does lead to academic success, therefore, universities should employ methods to improve the self-efficacy of their students.

Keywords: Academic achievement, grandiose narcissism, self-efficacy, self-esteem, vulnerable narcissism.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 437
8094 An Analysis of Variation of Ceiling Height and Window Level for Studio Architecture in Malaysia

Authors: Seyedehzahra Mirrahimi, Nik Lukman Nik Ibrahim, M. Surat

Abstract:

This paper investigated the impact of ceiling height and window head heights variation on daylighting inside architectural teaching studio with a full width window. In architectural education, using the studio is more than normal classroom in most credit hours. Therefore, window position, size and dimension of studio have direct influence on level of daylighting. Daylighting design is a critical factor that improves student learning, concentration and behavior, in addition to these, it also reduces energy consumption. The methodology of analysis involves using Radiance in IES software under overcast and cloudy sky in Malaysia. It has been established that presentation of daylighting of architecture studio can be enhanced by changing the ceiling heights and window level, because, different ceiling heights and window head heights can contribute to different range of daylight levels.

Keywords: Ceiling height, window head height, daylighting, studio architecture, simulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3326
8093 Negative Selection as a Means of Discovering Unknown Temporal Patterns

Authors: Wanli Ma, Dat Tran, Dharmendra Sharma

Abstract:

The temporal nature of negative selection is an under exploited area. In a negative selection system, newly generated antibodies go through a maturing phase, and the survivors of the phase then wait to be activated by the incoming antigens after certain number of matches. These without having enough matches will age and die, while these with enough matches (i.e., being activated) will become active detectors. A currently active detector may also age and die if it cannot find any match in a pre-defined (lengthy) period of time. Therefore, what matters in a negative selection system is the dynamics of the involved parties in the current time window, not the whole time duration, which may be up to eternity. This property has the potential to define the uniqueness of negative selection in comparison with the other approaches. On the other hand, a negative selection system is only trained with “normal" data samples. It has to learn and discover unknown “abnormal" data patterns on the fly by itself. Consequently, it is more appreciate to utilize negation selection as a system for pattern discovery and recognition rather than just pattern recognition. In this paper, we study the potential of using negative selection in discovering unknown temporal patterns.

Keywords: Artificial Immune Systems, ComputationalIntelligence, Negative Selection, Pattern Discovery.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1658
8092 Urban Growth, Sewerage Network and Flooding Risk: Flooding of November 10, 2001 in Algiers

Authors: Boualem El Kechebour, Djilali Benouar

Abstract:

The objective of this work is to present a expertise on flooding hazard analysis and how to reduce the risk. The analysis concerns the disaster induced by the flood on November 10/11, 2001 in the Bab El Oued district of the city of Algiers.The study begins by an expertise of damages in related with the urban environment and the history of the urban growth of the site. After this phase, the work is focalized on the identification of the existing correlations between the development of the town and its vulnerability. The final step consists to elaborate the interpretations on the interactions between the urban growth, the sewerage network and the vulnerability of the urban system.In conclusion, several recommendations are formulated permitting the mitigation of the risk in the future. The principal recommendations concern the new urban operations and the existing urbanized sites.

Keywords: urban growth, sewerage network, vulnerability of town, flooding risk, mitigation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1637
8091 Level Set and Morphological Operation Techniques in Application of Dental Image Segmentation

Authors: Abdolvahab Ehsani Rad, Mohd Shafry Mohd Rahim, Alireza Norouzi

Abstract:

Medical image analysis is one of the great effects of computer image processing. There are several processes to analysis the medical images which the segmentation process is one of the challenging and most important step. In this paper the segmentation method proposed in order to segment the dental radiograph images. Thresholding method has been applied to simplify the images and to morphologically open binary image technique performed to eliminate the unnecessary regions on images. Furthermore, horizontal and vertical integral projection techniques used to extract the each individual tooth from radiograph images. Segmentation process has been done by applying the level set method on each extracted images. Nevertheless, the experiments results by 90% accuracy demonstrate that proposed method achieves high accuracy and promising result.

Keywords: Integral production, level set method, morphological operation, segmentation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4227
8090 Influence of Surface-Treated Coarse Recycled Concrete Aggregate on Compressive Strength of Concrete

Authors: Sallehan Ismail, Mahyuddin Ramli

Abstract:

This paper reports on the influence of surface-treated coarse recycled concrete aggregate (RCA) on developing the compressive strength of concrete. The coarse RCA was initially treated by separately impregnating it in calcium metasilicate (CM) or wollastonite and nanosilica (NS) prepared at various concentrations. The effects of both treatment materials on concrete properties (e.g., slump, density and compressive strength) were evaluated. Scanning electron microscopy (SEM) analysis was performed to examine the microstructure of the resulting concrete. Results show that the effective use of treated coarse RCA significantly enhances the compressive strength of concrete. This result is supported by the SEM analysis, which indicates the formation of a dense interface between the treated coarse RCA and the cement matrix. Coarse RCA impregnated in CM solution results in better concrete strength than NS, and the optimum concentration of CM solution recommended for treated coarse RCA is 10%.

Keywords: Calcium metasilicate, compressive strength, nanosilica, recycled concrete aggregate.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2848
8089 Methodology for Bioenergy Potential and Assessment for Energy Deployment in Rural Vhembe District Areas

Authors: Clement M. Matasane, Mohamed T. Kahn

Abstract:

Biomass resources such as animal waste, agricultural and acro-industrial residues, forestry and woodland waste, and industrial and municipal solid wastes provide alternative means to utilize its untapped potential for biomass/biofuel renewable energy systems. In addition, crop residues (i.e., grain, starch, and energy crops) are commonly available in the district and play an essential role in community farming activities. The remote sensing technology (mappings) and geographic information systems tool will be used to determine the biomass potential in the Vhembe District Municipality. The detailed assessment, estimation, and modeling in quantifying their distribution, abundance, and quality yield an effective and efficient use of their potential. This paper aims to examine the potential and prospects of deploying bioenergy systems in small or micro-systems in the district for community use and applications. This deployment of the biofuels/biomass systems will help communities for sustainable energy supply from their traditional energy use into innovative and suitable methods that improve their livelihood. The study demonstrates the potential applications of Geographical Information Systems (GIS) in spatial mapping analysis, evaluation, modeling, and decision support for easy access to renewable energy systems.

Keywords: Agricultural crops, waste materials, biomass potentials, bioenergy potentials, GIS mappings, environmental data, renewable energy deployment, sustainable energy supply.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 310
8088 Flexible Cities: A Multisided Spatial Application of Tracking Livability of Urban Environment

Authors: Maria Christofi, George Plastiras, Rafaella Elia, Vaggelis Tsiourtis, Theocharis Theocharides, Miltiadis Katsaros

Abstract:

The rapidly expanding urban areas of the world constitute a challenge of how we need to make the transition to "the next urbanization", which will be defined by new analytical tools and new sources of data. This paper is about the production of a spatial application, the ‘FUMapp’, where space and its initiative will be available literally, in meters, but also abstractly, at a sensed level. While existing spatial applications typically focus on illustrations of the urban infrastructure, the suggested application goes beyond the existing: It investigates how our environment's perception adapts to the alterations of the built environment through a dataset construction of biophysical measurements (eye-tracking, heart beating), and physical metrics (spatial characteristics, size of stimuli, rhythm of mobility). It explores the intersections between architecture, cognition, and computing where future design can be improved and identifies the flexibility and livability of the ‘available space’ of specific examined urban paths.

Keywords: Biophysical data, flexibility of urban, livability, next urbanization, spatial application.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1020
8087 The Security Trade-Offs in Resource Constrained Nodes for IoT Application

Authors: Sultan Alharby, Nick Harris, Alex Weddell, Jeff Reeve

Abstract:

The concept of the Internet of Things (IoT) has received much attention over the last five years. It is predicted that the IoT will influence every aspect of our lifestyles in the near future. Wireless Sensor Networks are one of the key enablers of the operation of IoTs, allowing data to be collected from the surrounding environment. However, due to limited resources, nature of deployment and unattended operation, a WSN is vulnerable to various types of attack. Security is paramount for reliable and safe communication between IoT embedded devices, but it does, however, come at a cost to resources. Nodes are usually equipped with small batteries, which makes energy conservation crucial to IoT devices. Nevertheless, security cost in terms of energy consumption has not been studied sufficiently. Previous research has used a security specification of 802.15.4 for IoT applications, but the energy cost of each security level and the impact on quality of services (QoS) parameters remain unknown. This research focuses on the cost of security at the IoT media access control (MAC) layer. It begins by studying the energy consumption of IEEE 802.15.4 security levels, which is followed by an evaluation for the impact of security on data latency and throughput, and then presents the impact of transmission power on security overhead, and finally shows the effects of security on memory footprint. The results show that security overhead in terms of energy consumption with a payload of 24 bytes fluctuates between 31.5% at minimum level over non-secure packets and 60.4% at the top security level of 802.15.4 security specification. Also, it shows that security cost has less impact at longer packet lengths, and more with smaller packet size. In addition, the results depicts a significant impact on data latency and throughput. Overall, maximum authentication length decreases throughput by almost 53%, and encryption and authentication together by almost 62%.

Keywords: Internet of Things, IEEE 802.15.4, security cost evaluation, wireless sensor network, energy consumption.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1484
8086 Changing the Way South Africa Think about Parking Provision at Tertiary Institutions

Authors: M. C. Venter, G. Hitge, S. C. Krygsman, J. Thiart

Abstract:

For decades, South Africa has been planning transportation systems from a supply, rather than a demand side, perspective. In terms of parking, this relates to requiring the minimum parking provision that is enforced by city officials. Newer insight is starting to indicate that South Africa needs to re-think this philosophy in light of a new policy environment that desires a different outcome. Urban policies have shifted from reliance on the private car for access, to employing a wide range of alternative modes. Car dominated travel is influenced by various parameters, of which the availability and location of parking plays a significant role. The question is therefore, what is the right strategy to achieve the desired transport outcomes for SA. The focus of this paper is used to assess this issue with regard to parking provision, and specifically at a tertiary institution. A parking audit was conducted at the Stellenbosch campus of Stellenbosch University, monitoring occupancy at all 60 parking areas, every hour during business hours over a five-day period. The data from this survey was compared with the prescribed number of parking bays according to the Stellenbosch Municipality zoning scheme (requiring a minimum of 0.4 bays per student). The analysis shows that by providing 0.09 bays per student, the maximum total daily occupation of all the parking areas did not exceed an 80% occupation rate. It is concluded that the prevailing parking standards are not supportive of the new urban and transport policy environment, but that it is extremely conservative from a practical demand point of view.

Keywords: Parking provision, parking requirements, travel behaviour, travel demand management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 673
8085 Parametric Modeling Approach for Call Holding Times for IP based Public Safety Networks via EM Algorithm

Authors: Badarch Tuyatsetseg

Abstract:

This paper presents parametric probability density models for call holding times (CHTs) into emergency call center based on the actual data collected for over a week in the public Emergency Information Network (EIN) in Mongolia. When the set of chosen candidates of Gamma distribution family is fitted to the call holding time data, it is observed that the whole area in the CHT empirical histogram is underestimated due to spikes of higher probability and long tails of lower probability in the histogram. Therefore, we provide the Gaussian parametric model of a mixture of lognormal distributions with explicit analytical expressions for the modeling of CHTs of PSNs. Finally, we show that the CHTs for PSNs are fitted reasonably by a mixture of lognormal distributions via the simulation of expectation maximization algorithm. This result is significant as it expresses a useful mathematical tool in an explicit manner of a mixture of lognormal distributions.

Keywords: A mixture of lognormal distributions, modeling call holding times, public safety network.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1642
8084 Fast Generation of High-Performance Driveshafts: A Digital Approach to Automated Linked Topology and Design Optimization

Authors: Willi Zschiebsch, Alrik Dargel, Sebastian Spitzer, Philipp Johst, Robert Böhm, Niels Modler

Abstract:

In this article, we investigate an approach that digitally links individual development process steps by using the drive shaft of an aircraft engine as representative example of a fiber polymer composite. Such high-performance lightweight composite structures have many adjustable parameters that influence the mechanical properties. Only a combination of optimal parameter values can lead to energy efficient lightweight structures. The development tools required for the Engineering Design Process (EDP) are often isolated solutions and their compatibility with each other is limited. A digital framework is presented in this study, which allows individual specialised tools to be linked via the generated data in such a way that automated optimization across programs becomes possible. This is demonstrated using the example of linking geometry generation with numerical structural analysis. The proposed digital framework for automated design optimization demonstrates the feasibility of developing a complete digital approach to design optimization. The methodology shows promising potential for achieving optimal solutions in terms of mass, material utilization, eigenfrequency and deformation under lateral load with less development effort. The development of such a framework is an important step towards promoting a more efficient design approach that can lead to stable and balanced results.

Keywords: Digital Linked Process, composite, CFRP, multi-objective, EDP, NSGA-2, NSGA-3, TPE.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 143
8083 Trabecular Bone Radiograph Characterization Using Fractal, Multifractal Analysis and SVM Classifier

Authors: I. Slim, H. Akkari, A. Ben Abdallah, I. Bhouri, M. Hedi Bedoui

Abstract:

Osteoporosis is a common disease characterized by low bone mass and deterioration of micro-architectural bone tissue, which provokes an increased risk of fracture. This work treats the texture characterization of trabecular bone radiographs. The aim was to analyze according to clinical research a group of 174 subjects: 87 osteoporotic patients (OP) with various bone fracture types and 87 control cases (CC). To characterize osteoporosis, Fractal and MultiFractal (MF) methods were applied to images for features (attributes) extraction. In order to improve the results, a new method of MF spectrum based on the q-stucture function calculation was proposed and a combination of Fractal and MF attributes was used. The Support Vector Machines (SVM) was applied as a classifier to distinguish between OP patients and CC subjects. The features fusion (fractal and MF) allowed a good discrimination between the two groups with an accuracy rate of 96.22%.

Keywords: Fractal, micro-architecture analysis, multifractal, SVM, osteoporosis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 975
8082 Human Walking Vertical Force and Vertical Vibration of Pedestrian Bridge Induced by Its Higher Components

Authors: M. Yoneda

Abstract:

The purpose of this study is to identify human walking vertical force by using FFT power spectrum density from the experimental acceleration data of the human body. An experiment on human walking is carried out on a stationary floor especially paying attention to higher components of dynamic vertical walking force. Based on measured acceleration data of the human lumbar part, not only in-phase component with frequency of 2fw, 3fw, but also in-opposite-phase component with frequency of 0.5 fw, 1.5 fw, 2.5 fw where fw is the walking rate is observed. The vertical vibration of pedestrian bridge induced by higher components of human walking vertical force is also discussed in this paper. A full scale measurement for the existing pedestrian bridge with center span length of 33 m is carried out focusing on the resonance phenomenon due to higher components of human walking vertical force. Dynamic response characteristics excited by these vertical higher components of human walking are revealed from the dynamic design viewpoint of pedestrian bridge.

Keywords: Simplified method, Human walking vertical force, Higher component, Pedestrian bridge vibration.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1809
8081 Improving 99mTc-tetrofosmin Myocardial Perfusion Images by Time Subtraction Technique

Authors: Yasuyuki Takahashi, Hayato Ishimura, Masao Miyagawa, Teruhito Mochizuki

Abstract:

Quantitative measurement of myocardium perfusion is possible with single photon emission computed tomography (SPECT) using a semiconductor detector. However, accumulation of 99mTc-tetrofosmin in the liver may make it difficult to assess that accurately in the inferior myocardium. Our idea is to reduce the high accumulation in the liver by using dynamic SPECT imaging and a technique called time subtraction. We evaluated the performance of a new SPECT system with a cadmium-zinc-telluride solid-state semi- conductor detector (Discovery NM 530c; GE Healthcare). Our system acquired list-mode raw data over 10 minutes for a typical patient. From the data, ten SPECT images were reconstructed, one for every minute of acquired data. Reconstruction with the semiconductor detector was based on an implementation of a 3-D iterative Bayesian reconstruction algorithm. We studied 20 patients with coronary artery disease (mean age 75.4 ± 12.1 years; range 42-86; 16 males and 4 females). In each subject, 259 MBq of 99mTc-tetrofosmin was injected intravenously. We performed both a phantom and a clinical study using dynamic SPECT. An approximation to a liver-only image is obtained by reconstructing an image from the early projections during which time the liver accumulation dominates (0.5~2.5 minutes SPECT image-5~10 minutes SPECT image). The extracted liver-only image is then subtracted from a later SPECT image that shows both the liver and the myocardial uptake (5~10 minutes SPECT image-liver-only image). The time subtraction of liver was possible in both a phantom and the clinical study. The visualization of the inferior myocardium was improved. In past reports, higher accumulation in the myocardium due to the overlap of the liver is un-diagnosable. Using our time subtraction method, the image quality of the 99mTc-tetorofosmin myocardial SPECT image is considerably improved.

Keywords: 99mTc-tetrofosmin, dynamic SPECT, time subtraction, semiconductor detector.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1027
8080 A Damage Level Assessment Model for Extra High Voltage Transmission Towers

Authors: Huan-Chieh Chiu, Hung-Shuo Wu, Chien-Hao Wang, Yu-Cheng Yang, Ching-Ya Tseng, Joe-Air Jiang

Abstract:

Power failure resulting from tower collapse due to violent seismic events might bring enormous and inestimable losses. The Chi-Chi earthquake, for example, strongly struck Taiwan and caused huge damage to the power system on September 21, 1999. Nearly 10% of extra high voltage (EHV) transmission towers were damaged in the earthquake. Therefore, seismic hazards of EHV transmission towers should be monitored and evaluated. The ultimate goal of this study is to establish a damage level assessment model for EHV transmission towers. The data of earthquakes provided by Taiwan Central Weather Bureau serve as a reference and then lay the foundation for earthquake simulations and analyses afterward. Some parameters related to the damage level of each point of an EHV tower are simulated and analyzed by the data from monitoring stations once an earthquake occurs. Through the Fourier transform, the seismic wave is then analyzed and transformed into different wave frequencies, and the data would be shown through a response spectrum. With this method, the seismic frequency which damages EHV towers the most is clearly identified. An estimation model is built to determine the damage level caused by a future seismic event. Finally, instead of relying on visual observation done by inspectors, the proposed model can provide a power company with the damage information of a transmission tower. Using the model, manpower required by visual observation can be reduced, and the accuracy of the damage level estimation can be substantially improved. Such a model is greatly useful for health and construction monitoring because of the advantages of long-term evaluation of structural characteristics and long-term damage detection.

Keywords: Smart grid, EHV transmission tower, response spectrum, damage level monitoring.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1059
8079 Diagnosis of Multivariate Process via Nonlinear Kernel Method Combined with Qualitative Representation of Fault Patterns

Authors: Hyun-Woo Cho

Abstract:

The fault detection and diagnosis of complicated production processes is one of essential tasks needed to run the process safely with good final product quality. Unexpected events occurred in the process may have a serious impact on the process. In this work, triangular representation of process measurement data obtained in an on-line basis is evaluated using simulation process. The effect of using linear and nonlinear reduced spaces is also tested. Their diagnosis performance was demonstrated using multivariate fault data. It has shown that the nonlinear technique based diagnosis method produced more reliable results and outperforms linear method. The use of appropriate reduced space yielded better diagnosis performance. The presented diagnosis framework is different from existing ones in that it attempts to extract the fault pattern in the reduced space, not in the original process variable space. The use of reduced model space helps to mitigate the sensitivity of the fault pattern to noise.

Keywords: Real-time Fault diagnosis, triangular representation of patterns in reduced spaces, Nonlinear kernel technique, multivariate statistical modeling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1595
8078 Modeling Default Probabilities of the Chosen Czech Banks in the Time of the Financial Crisis

Authors: Petr Gurný

Abstract:

One of the most important tasks in the risk management is the correct determination of probability of default (PD) of particular financial subjects. In this paper a possibility of determination of financial institution’s PD according to the creditscoring models is discussed. The paper is divided into the two parts. The first part is devoted to the estimation of the three different models (based on the linear discriminant analysis, logit regression and probit regression) from the sample of almost three hundred US commercial banks. Afterwards these models are compared and verified on the control sample with the view to choose the best one. The second part of the paper is aimed at the application of the chosen model on the portfolio of three key Czech banks to estimate their present financial stability. However, it is not less important to be able to estimate the evolution of PD in the future. For this reason, the second task in this paper is to estimate the probability distribution of the future PD for the Czech banks. So, there are sampled randomly the values of particular indicators and estimated the PDs’ distribution, while it’s assumed that the indicators are distributed according to the multidimensional subordinated Lévy model (Variance Gamma model and Normal Inverse Gaussian model, particularly). Although the obtained results show that all banks are relatively healthy, there is still high chance that “a financial crisis” will occur, at least in terms of probability. This is indicated by estimation of the various quantiles in the estimated distributions. Finally, it should be noted that the applicability of the estimated model (with respect to the used data) is limited to the recessionary phase of the financial market.

Keywords: Credit-scoring Models, Multidimensional Subordinated Lévy Model, Probability of Default.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1914
8077 Hydrodynamic Analysis with Heat Transfer in Solid Gas Fluidized Bed Reactor for Solar Thermal Applications

Authors: Sam Rasoulzadeh, Atefeh Mousavi

Abstract:

Fluidized bed reactors are known as highly exothermic and endothermic according to uniformity in temperature as a safe and effective mean for catalytic reactors. In these reactors, a wide range of catalyst particles can be used and by using a continuous operation proceed to produce in succession. Providing optimal conditions for the operation of these types of reactors will prevent the exorbitant costs necessary to carry out laboratory work. In this regard, a hydrodynamic analysis was carried out with heat transfer in the solid-gas fluidized bed reactor for solar thermal applications. The results showed that in the fluid flow the input of the reactor has a lower temperature than the outlet, and when the fluid is passing from the reactor, the heat transfer happens between cylinder and solar panel and fluid. It increases the fluid temperature in the outlet pump and also the kinetic energy of the fluid has been raised in the outlet areas.

Keywords: Heat transfer, solar reactor, fluidized bed reactor, CFD.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 648
8076 Internal Force State Recognition of Jiujiang Bridge Based on Cable Force-displacement Relationship

Authors: Weifeng Wang, Guoqing Huang, Xianwei Zeng

Abstract:

The nearly 21-year-old Jiujiang Bridge, which is suffering from uneven line shape, constant great downwarping of the main beam and cracking of the box girder, needs reinforcement and cable adjustment. It has undergone cable adjustment for twice with incomplete data. Therefore, the initial internal force state of the Jiujiang Bridge is identified as the key for the cable adjustment project. Based on parameter identification by means of static force test data, this paper suggests determining the initial internal force state of the cable-stayed bridge according to the cable force-displacement relationship parameter identification method. That is, upon measuring the displacement and the change in cable forces for twice, one can identify the parameters concerned by means of optimization. This method is applied to the cable adjustment, replacement and reinforcement project for the Jiujiang Bridge as a guidance for the cable adjustment and reinforcement project of the bridge.

Keywords: Cable-stayed bridge, cable force-displacement, parameter identification, internal force state

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1534
8075 Design of Nonlinear Robust Control in a Class of Structurally Stable Functions

Authors: V. Ten

Abstract:

An approach of design of stable of control systems with ultimately wide ranges of uncertainly disturbed parameters is offered. The method relies on using of nonlinear structurally stable functions from catastrophe theory as controllers. Theoretical part presents an analysis of designed nonlinear second-order control systems. As more important the integrators in series, canonical controllable form and Jordan forms are considered. The analysis resumes that due to added controllers systems become stable and insensitive to any disturbance of parameters. Experimental part presents MATLAB simulation of design of control systems of epidemic spread, aircrafts angular motion and submarine depth. The results of simulation confirm the efficiency of offered method of design. KeywordsCatastrophes, robust control, simulation, uncertain parameters.

Keywords: Catastrophes, robust control, simulation, uncertain parameters.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1287
8074 Assessment and Uncertainty Analysis of ROSA/LSTF Test on Pressurized Water Reactor 1.9% Vessel Upper Head Small-Break Loss-of-Coolant Accident

Authors: Takeshi Takeda

Abstract:

An experiment utilizing the ROSA/LSTF (rig of safety assessment/large-scale test facility) simulated a 1.9% vessel upper head small-break loss-of-coolant accident with an accident management (AM) measure under the total failure of high-pressure injection system of emergency core cooling system in a pressurized water reactor. Steam generator (SG) secondary-side depressurization on the AM measure was started by fully opening relief valves in both SGs when the maximum core exit temperature rose to 623 K. A large increase took place in the cladding surface temperature of simulated fuel rods on account of a late and slow response of core exit thermocouples during core boil-off. The author analyzed the LSTF test by reference to the matrix of an integral effect test for the validation of a thermal-hydraulic system code. Problems remained in predicting the primary coolant distribution and the core exit temperature with the RELAP5/MOD3.3 code. The uncertainty analysis results of the RELAP5 code confirmed that the sample size with respect to the order statistics influences the value of peak cladding temperature with a 95% probability at a 95% confidence level, and the Spearman’s rank correlation coefficient.

Keywords: LSTF, LOCA, uncertainty analysis, RELAP5.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 720
8073 Four Phase Methodology for Developing Secure Software

Authors: Carlos Gonzalez-Flores, Ernesto Liñan-García

Abstract:

A simple and robust approach for developing secure software. A Four Phase methodology consists in developing the non-secure software in phase one, and for the next three phases, one phase for each of the secure developing types (i.e. self-protected software, secure code transformation, and the secure shield). Our methodology requires first the determination and understanding of the type of security level needed for the software. The methodology proposes the use of several teams to accomplish this task. One Software Engineering Developing Team, a Compiler Team, a Specification and Requirements Testing Team, and for each of the secure software developing types: three teams of Secure Software Developing, three teams of Code Breakers, and three teams of Intrusion Analysis. These teams will interact among each other and make decisions to provide a secure software code protected against a required level of intruder.

Keywords: Secure Software, Four Phase Methodology, Software Engineering, Code Breakers, Intrusion Analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1827
8072 Rank-Based Chain-Mode Ensemble for Binary Classification

Authors: Chongya Song, Kang Yen, Alexander Pons, Jin Liu

Abstract:

In the field of machine learning, the ensemble has been employed as a common methodology to improve the performance upon multiple base classifiers. However, the true predictions are often canceled out by the false ones during consensus due to a phenomenon called “curse of correlation” which is represented as the strong interferences among the predictions produced by the base classifiers. In addition, the existing practices are still not able to effectively mitigate the problem of imbalanced classification. Based on the analysis on our experiment results, we conclude that the two problems are caused by some inherent deficiencies in the approach of consensus. Therefore, we create an enhanced ensemble algorithm which adopts a designed rank-based chain-mode consensus to overcome the two problems. In order to evaluate the proposed ensemble algorithm, we employ a well-known benchmark data set NSL-KDD (the improved version of dataset KDDCup99 produced by University of New Brunswick) to make comparisons between the proposed and 8 common ensemble algorithms. Particularly, each compared ensemble classifier uses the same 22 base classifiers, so that the differences in terms of the improvements toward the accuracy and reliability upon the base classifiers can be truly revealed. As a result, the proposed rank-based chain-mode consensus is proved to be a more effective ensemble solution than the traditional consensus approach, which outperforms the 8 ensemble algorithms by 20% on almost all compared metrices which include accuracy, precision, recall, F1-score and area under receiver operating characteristic curve.

Keywords: Consensus, curse of correlation, imbalanced classification, rank-based chain-mode ensemble.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 722