Search results for: Complexity measurement
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1853

Search results for: Complexity measurement

293 Dynamic Threshold Adjustment Approach For Neural Networks

Authors: Hamza A. Ali, Waleed A. J. Rasheed

Abstract:

The use of neural networks for recognition application is generally constrained by their inherent parameters inflexibility after the training phase. This means no adaptation is accommodated for input variations that have any influence on the network parameters. Attempts were made in this work to design a neural network that includes an additional mechanism that adjusts the threshold values according to the input pattern variations. The new approach is based on splitting the whole network into two subnets; main traditional net and a supportive net. The first deals with the required output of trained patterns with predefined settings, while the second tolerates output generation dynamically with tuning capability for any newly applied input. This tuning comes in the form of an adjustment to the threshold values. Two levels of supportive net were studied; one implements an extended additional layer with adjustable neuronal threshold setting mechanism, while the second implements an auxiliary net with traditional architecture performs dynamic adjustment to the threshold value of the main net that is constructed in dual-layer architecture. Experiment results and analysis of the proposed designs have given quite satisfactory conducts. The supportive layer approach achieved over 90% recognition rate, while the multiple network technique shows more effective and acceptable level of recognition. However, this is achieved at the price of network complexity and computation time. Recognition generalization may be also improved by accommodating capabilities involving all the innate structures in conjugation with Intelligence abilities with the needs of further advanced learning phases.

Keywords: Classification, Recognition, Neural Networks, Pattern Recognition, Generalization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1587
292 Investigating Daylight Quality in Malaysian Government Office Buildings Through Daylight Factor and Surface Luminance

Authors: Mohd Zin Kandar, Mohd Sabere Sulaiman, Yong Razidah Rashid, Dilshan Remaz Ossen, Aminatuzuhariah MAbdullah, Lim Yaik Wah, Mansour Nikpour

Abstract:

In recent years, there has been an increasing interest in using daylight to save energy in buildings. In tropical regions, daylighting is always an energy saver. On the other hand, daylight provides visual comfort. According to standards, it shows that many criteria should be taken into consideration in order to have daylight utilization and visual comfort. The current standard in Malaysia, MS 1525 does not provide sufficient guideline. Hence, more research is needed on daylight performance. If architects do not consider daylight design, it not only causes inconvenience in working spaces but also causes more energy consumption as well as environmental pollution. This research had surveyed daylight performance in 5 selected office buildings from different area of Malaysian through experimental method. Several parameters of daylight quality such as daylight factor, surface luminance and surface luminance ratio were measured in different rooms in each building. The result of this research demonstrated that most of the buildings were not designed for daylight utilization. Therefore, it is very important that architects follow the daylight design recommendation to reduce consumption of electric power for artificial lighting while the sufficient quality of daylight is available.

Keywords: Daylight factor, Field measurement, Daylighting quality, Tropical

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3418
291 The Ecological Role of Loligo forbesii in the Moray Firth Ecosystem, Northeast Scotland

Authors: Godwin A. Otogo, Sansanee Wangvoralak, Graham J. Pierce, Lee C. Hastie, Beth Scott

Abstract:

The squid Loligo forbesii is suspected to be an important species in marine food webs, as it can strongly impact its prey and be impacted upon by predation, competition, fishing and/or climate variability. To quantify these impacts in the food web, the measurement of its trophic position and ecological role within well-studied ecosystems is essential. An Ecopath model was balanced and run for the Moray Firth ecosystem and was used to investigate the significance of this squid’s trophic roles. The network analysis routine included in Ecopath with Ecosim (EwE) was used to estimate trophic interaction, system indicators (health condition and developmental stage) and food web features. Results indicated that within the Moray Firth squid occupy a top trophic position in the food web and also a major prey item for many other species. Results from Omnivory Index (OI) showed that squid is a generalized feeder transferring energy across wide trophic levels and is more important as a predator than that as a prey in the Moray Firth ecosystem. The results highlight the importance of taking squid into account in the management of Europe’s living marine resources.

Keywords: Ecopath, Loligo forbesii, moray firth, squid, trophic-level.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1323
290 Use of Fuzzy Logic in the Corporate Reputation Assessment: Stock Market Investors’ Perspective

Authors: Tomasz L. Nawrocki, Danuta Szwajca

Abstract:

The growing importance of reputation in building enterprise value and achieving long-term competitive advantage creates the need for its measurement and evaluation for the management purposes (effective reputation and its risk management). The paper presents practical application of self-developed corporate reputation assessment model from the viewpoint of stock market investors. The model has a pioneer character and example analysis performed for selected industry is a form of specific test for this tool. In the proposed solution, three aspects - informational, financial and development, as well as social ones - were considered. It was also assumed that the individual sub-criteria will be based on public sources of information, and as the calculation apparatus, capable of obtaining synthetic final assessment, fuzzy logic will be used. The main reason for developing this model was to fulfill the gap in the scope of synthetic measure of corporate reputation that would provide higher degree of objectivity by relying on "hard" (not from surveys) and publicly available data. It should be also noted that results obtained on the basis of proposed corporate reputation assessment method give possibilities of various internal as well as inter-branch comparisons and analysis of corporate reputation impact.

Keywords: Corporate reputation, fuzzy logic, fuzzy model, stock market investors.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1324
289 A Comparative Analysis Approach Based on Fuzzy AHP, TOPSIS and PROMETHEE for the Selection Problem of GSCM Solutions

Authors: Omar Boutkhoum, Mohamed Hanine, Abdessadek Bendarag

Abstract:

Sustainable economic growth is nowadays driving firms to extend toward the adoption of many green supply chain management (GSCM) solutions. However, the evaluation and selection of these solutions is a matter of concern that needs very serious decisions, involving complexity owing to the presence of various associated factors. To resolve this problem, a comparative analysis approach based on multi-criteria decision-making methods is proposed for adequate evaluation of sustainable supply chain management solutions. In the present paper, we propose an integrated decision-making model based on FAHP (Fuzzy Analytic Hierarchy Process), TOPSIS (Technique for Order of Preference by Similarity to Ideal Solution) and PROMETHEE (Preference Ranking Organisation METHod for Enrichment Evaluations) to contribute to a better understanding and development of new sustainable strategies for industrial organizations. Due to the varied importance of the selected criteria, FAHP is used to identify the evaluation criteria and assign the importance weights for each criterion, while TOPSIS and PROMETHEE methods employ these weighted criteria as inputs to evaluate and rank the alternatives. The main objective is to provide a comparative analysis based on TOPSIS and PROMETHEE processes to help make sound and reasoned decisions related to the selection problem of GSCM solution.

Keywords: GSCM solutions, multi-criteria analysis, FAHP, TOPSIS, PROMETHEE, decision support system.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 883
288 Activity Recognition by Smartphone Accelerometer Data Using Ensemble Learning Methods

Authors: Eu Tteum Ha, Kwang Ryel Ryu

Abstract:

As smartphones are equipped with various sensors, there have been many studies focused on using these sensors to create valuable applications. Human activity recognition is one such application motivated by various welfare applications, such as the support for the elderly, measurement of calorie consumption, lifestyle and exercise patterns analyses, and so on. One of the challenges one faces when using smartphone sensors for activity recognition is that the number of sensors should be minimized to save battery power. In this paper, we show that a fairly accurate classifier can be built that can distinguish ten different activities by using only a single sensor data, i.e., the smartphone accelerometer data. The approach that we adopt to deal with this twelve-class problem uses various methods. The features used for classifying these activities include not only the magnitude of acceleration vector at each time point, but also the maximum, the minimum, and the standard deviation of vector magnitude within a time window. The experiments compared the performance of four kinds of basic multi-class classifiers and the performance of four kinds of ensemble learning methods based on three kinds of basic multi-class classifiers. The results show that while the method with the highest accuracy is ECOC based on Random forest.

Keywords: Ensemble learning, activity recognition, smartphone accelerometer.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2116
287 Study on the Impact of Size and Position of the Shear Field in Determining the Shear Modulus of Glulam Beam Using Photogrammetry Approach

Authors: Niaz Gharavi, Hexin Zhang

Abstract:

The shear modulus of a timber beam can be determined using torsion test or shear field test method. The shear field test method is based on shear distortion measurement of the beam at the zone with the constant transverse load in the standardized four-point bending test. The current code of practice advises using two metallic arms act as an instrument to measure the diagonal displacement of the constructing square. The size and the position of the constructing square might influence the shear modulus determination. This study aimed to investigate the size and the position effect of the square in the shear field test method. A binocular stereo vision system has been employed to determine the 3D displacement of a grid of target points. Six glue laminated beams were produced and tested. Analysis of Variance (ANOVA) was performed on the acquired data to evaluate the significance of the size effect and the position effect of the square. The results have shown that the size of the square has a noticeable influence on the value of shear modulus, while, the position of the square within the area with the constant shear force does not affect the measured mean shear modulus.

Keywords: Shear field test method, structural-sized test, shear modulus of Glulam beam, photogrammetry approach.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 949
286 Pleurotus Ostreatus for Durability Test of Rubber and Sengon Woods using Indonesian National Standard and Japanese Standard Methods

Authors: Elis N. Herliyana , Kunio Tsunoda, Yusuf S. Hadi, Arinana, Dewi A. Natalia

Abstract:

This study aims to determine the level of resistance of Hevea brasiliensis and Paraserianthes falcataria (synonym: Falcataria molucana) against wood rot fungi Pleurotus ostreatus based on Indonesian standard SNI 01.7207-2006 and Japanese standard JIS K 1571-2004. The variables measured were visual appearance and weight loss percentage of wood based on longitudinal and cross section fiber directions of rubber wood and sengon wood. Measurement of oven dry weight loss of wood samples performed after 12 weeks incubation. Replication performed was 10 times at each treatment combination. The results based on SNI 01.7207-2006, weight loss value of H. brasiliensis and P. falcataria wood with fiber direction longitudinal were 23,12 and 22,25% respectively and cross section were 20,77 and 18,76% respectively, and all were classified to resistance class IV (no resistance). The results based on JIS K 1571-2004, weight loss value of both woods with fiber direction cross section were 10,95 and 14,20% respectively.

Keywords: H. brasiliensis wood, Natural durability, P. falcataria wood, P. ostreatus.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3021
285 Systems Engineering Management Using Transdisciplinary Quality System Development Lifecycle Model

Authors: Mohamed Asaad Abdelrazek, Amir Taher El-Sheikh, M. Zayan, A.M. Elhady

Abstract:

The successful realization of complex systems is dependent not only on the technology issues and the process for implementing them, but on the management issues as well. Managing the systems development lifecycle requires technical management. Systems engineering management is the technical management. Systems engineering management is accomplished by incorporating many activities. The three major activities are development phasing, systems engineering process and lifecycle integration. Systems engineering management activities are performed across the system development lifecycle. Due to the ever-increasing complexity of systems as well the difficulty of managing and tracking the development activities, new ways to achieve systems engineering management activities are required. This paper presents a systematic approach used as a design management tool applied across systems engineering management roles. In this approach, Transdisciplinary System Development Lifecycle (TSDL) Model has been modified and integrated with Quality Function Deployment. Hereinafter, the name of the systematic approach is the Transdisciplinary Quality System Development Lifecycle (TQSDL) Model. The QFD translates the voice of customers (VOC) into measurable technical characteristics. The modified TSDL model is based on Axiomatic Design developed by Suh which is applicable to all designs: products, processes, systems and organizations. The TQSDL model aims to provide a robust structure and systematic thinking to support the implementation of systems engineering management roles. This approach ensures that the customer requirements are fulfilled as well as satisfies all the systems engineering manager roles and activities.

Keywords: Axiomatic design, quality function deployment, systems engineering management, system development lifecycle.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1685
284 Experimental Investigation on Residual Stresses in Welded Medium-Walled I-shaped Sections Fabricated from Q460GJ Structural Steel Plates

Authors: Qian Zhu, Shidong Nie, Bo Yang, Gang Xiong, Guoxin Dai

Abstract:

GJ steel is a new type of high-performance structural steel which has been increasingly adopted in practical engineering. Q460GJ structural steel has a nominal yield strength of 460 MPa, which does not decrease significantly with the increase of steel plate thickness like normal structural steel. Thus, Q460GJ structural steel is normally used in medium-walled welded sections. However, research works on the residual stress in GJ steel members are few though it is one of the vital factors that can affect the member and structural behavior. This article aims to investigate the residual stresses in welded I-shaped sections fabricated from Q460GJ structural steel plates by experimental tests. A total of four full scale welded medium-walled I-shaped sections were tested by sectioning method. Both circular curve correction method and straightening measurement method were adopted in this study to obtain the final magnitude and distribution of the longitudinal residual stresses. In addition, this paper also explores the interaction between flanges and webs. And based on the statistical evaluation of the experimental data, a multilayer residual stress model is proposed.

Keywords: Q460GJ structural steel, residual stresses, sectioning method, Welded medium-walled I-shaped sections.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1007
283 Adaptive Fuzzy Control of Stewart Platform under Actuator Saturation

Authors: Dongsu Wu, Hongbin Gu, Peng Li

Abstract:

A novel adaptive fuzzy trajectory tracking algorithm of Stewart platform based motion platform is proposed to compensate path deviation and degradation of controller-s performance due to actuator torque limit. The algorithm can be divided into two parts: the real-time trajectory shaping part and the joint space adaptive fuzzy controller part. For a reference trajectory in task space whenever any of the actuators is saturated, the desired acceleration of the reference trajectory is modified on-line by using dynamic model of motion platform. Meanwhile an additional action with respect to the difference between the nominal and modified trajectories is utilized in the non-saturated region of actuators to reduce the path error. Using modified trajectory as input, the joint space controller incorporates compute torque controller, leg velocity observer and fuzzy disturbance observer with saturation compensation. It can ensure stability and tracking performance of controller in present of external disturbance and position only measurement. Simulation results verify the effectiveness of proposed control scheme.

Keywords: Actuator saturation, adaptive fuzzy control, Stewartplatform, trajectory shaping, flight simulator

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1982
282 The Communication Library DIALOG for iFDAQ of the COMPASS Experiment

Authors: Y. Bai, M. Bodlak, V. Frolov, S. Huber, V. Jary, I. Konorov, D. Levit, J. Novy, D. Steffen, O. Subrt, M. Virius

Abstract:

Modern experiments in high energy physics impose great demands on the reliability, the efficiency, and the data rate of Data Acquisition Systems (DAQ). This contribution focuses on the development and deployment of the new communication library DIALOG for the intelligent, FPGA-based Data Acquisition System (iFDAQ) of the COMPASS experiment at CERN. The iFDAQ utilizing a hardware event builder is designed to be able to readout data at the maximum rate of the experiment. The DIALOG library is a communication system both for distributed and mixed environments, it provides a network transparent inter-process communication layer. Using the high-performance and modern C++ framework Qt and its Qt Network API, the DIALOG library presents an alternative to the previously used DIM library. The DIALOG library was fully incorporated to all processes in the iFDAQ during the run 2016. From the software point of view, it might be considered as a significant improvement of iFDAQ in comparison with the previous run. To extend the possibilities of debugging, the online monitoring of communication among processes via DIALOG GUI is a desirable feature. In the paper, we present the DIALOG library from several insights and discuss it in a detailed way. Moreover, the efficiency measurement and comparison with the DIM library with respect to the iFDAQ requirements is provided.

Keywords: Data acquisition system, DIALOG library, DIM library, FPGA, Qt framework, TCP/IP.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1066
281 Motivating Factors of Couple Involvement in Copreneurship Businesses in Malaysia

Authors: Norasmah Othman, Suzana Mohamed, Salpiah Suradi

Abstract:

Copreneurship is a term used to describe the business pattern of operations run by married couples who share commitment, goals, and responsibilities in handling a business. Research conducted overseas showed that copreneurship business activities grew quickly and played a role in elevating families’ and nations’ socio-economic standards. In Malaysia, copreneurship has long been cultivated by spouses. Thus, this study aimed to explore the factors that motivate married partners to start a copreneurship business, and who is the dominant partner in the management of this business. The study participants are four entrepreneurial couples who are SME business operators selected through purposive sampling. In-depth interviews and direct observation were used as methods of measurement for triangulation of qualitative data in this study. The findings of the interviews were administered using NVivo 8.0 software. The result shows that freedom is a key factor that drives entrepreneurs to set up copreneurship businesses, and that the husband dominates the management aspects of the business. The study gives an overview of the parties involved in entrepreneurship to provide understanding of the copreneurship concept as it is practiced. This study provides academic value by creating understanding of the importance of a harmonious family institution specifically for forming entrepreneurs in the familial environment in Malaysia.

Keywords: Copreneurs, copreneurship, business management, enterprise.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2662
280 Co-payment Strategies for Chronic Medications: A Qualitative and Comparative Analysis at European Level

Authors: Pedro M. Abreu, Bruno R. Mendes

Abstract:

The management of pharmacotherapy and the process of dispensing medicines is becoming critical in clinical pharmacy due to the increase of incidence and prevalence of chronic diseases, the complexity and customization of therapeutic regimens, the introduction of innovative and more expensive medicines, the unbalanced relation between expenditure and revenue as well as due to the lack of rationalization associated with medication use. For these reasons, co-payments emerged in Europe in the 70s and have been applied over the past few years in healthcare. Co-payments lead to a rationing and rationalization of user’s access under healthcare services and products, and simultaneously, to a qualification and improvement of the services and products for the end-user. This analysis, under hospital practices particularly and co-payment strategies in general, was carried out on all the European regions and identified four reference countries, that apply repeatedly this tool and with different approaches. The structure, content and adaptation of European co-payments were analyzed through 7 qualitative attributes and 19 performance indicators, and the results expressed in a scorecard, allowing to conclude that the German models (total score of 68,2% and 63,6% in both elected co-payments) can collect more compliance and effectiveness, the English models (total score of 50%) can be more accessible, and the French models (total score of 50%) can be more adequate to the socio-economic and legal framework. Other European models did not show the same quality and/or performance, so were not taken as a standard in the future design of co-payments strategies. In this sense, we can see in the co-payments a strategy not only to moderate the consumption of healthcare products and services, but especially to improve them, as well as a strategy to increment the value that the end-user assigns to these services and products, such as medicines.

Keywords: Clinical pharmacy, co-payments, healthcare, medicines.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1302
279 Optimization of Thermopile Sensor Performance of Polycrystalline Silicon Film

Authors: Li Long, Thomas Ortlepp

Abstract:

A theoretical model for the optimization of thermopile sensor performance is developed for thermoelectric-based infrared radiation detection. It is shown that the performance of polycrystalline silicon film thermopile sensor can be optimized according to the thermoelectric quality factor, sensor layer structure factor and sensor layout shape factor. Based on the properties of electrons, phonons, grain boundaries and their interactions, the thermoelectric quality factor of polycrystalline silicon is analyzed with the relaxation time approximation of Boltzmann transport equation. The model includes the effects of grain structure, grain boundary trap properties and doping concentration. The layer structure factor of sensor is analyzed with respect to infrared absorption coefficient. The effect of layout design is characterized with the shape factor, which is calculated for different sensor designs. Double layer polycrystalline silicon thermopile infrared sensors on suspended support membrane have been designed and fabricated with a CMOS-compatible process. The theoretical approach is confirmed with measurement results.

Keywords: Polycrystalline silicon film, relaxation time approximation, specific detectivity, thermal conductivity, thermopile infrared sensor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 154
278 Information Technology for Business Process Management in Insurance Companies

Authors: Vesna Bosilj Vukšić, Darija Ivandić Vidović, Ljubica Milanović Glavan

Abstract:

Information technology plays an irreplaceable role in introducing and improving business process orientation in a company. It enables implementation of the theoretical concept, measurement of results achieved and undertaking corrective measures aimed at improvements. Information technology is a key concept in the development and implementation of the business process management systems as it establishes a connection to business operations. Both in the literature and practice, insurance companies are often seen as highly process oriented due to the nature of their business and focus on customers. They are also considered leaders in using information technology for business process management. The research conducted aimed to investigate whether the perceived leadership status of insurance companies is well deserved, i.e. to establish the level of process orientation and explore the practice of information technology use in insurance companies in the region. The main instrument for primary data collection within this research was an electronic survey questionnaire sent to the management of insurance companies in the Republic of Croatia, Bosnia and Herzegovina, Slovenia, Serbia and Macedonia. The conducted research has shown that insurance companies have a satisfactory level of process orientation, but that there is also a huge potential for improvement, especially in the segment of information technology and its connection to business processes.

Keywords: Business processes management, process orientation, information technology, insurance companies.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2411
277 The Application of Queuing Theory in Multi-Stage Production Lines

Authors: Hani Shafeek, Muhammed Marsudi

Abstract:

The purpose of this work is examining the multiproduct multi-stage in a battery production line. To improve the performances of an assembly production line by determine the efficiency of each workstation. Data collected from every workstation. The data are throughput rate, number of operator, and number of parts that arrive and leaves during part processing. Data for the number of parts that arrives and leaves are collected at least at the amount of ten samples to make the data is possible to be analyzed by Chi-Squared Goodness Test and queuing theory. Measures of this model served as the comparison with the standard data available in the company. Validation of the task time value resulted by comparing it with the task time value based on the company database. Some performance factors for the multi-product multi-stage in a battery production line in this work are shown. The efficiency in each workstation was also shown. Total production time to produce each part can be determined by adding the total task time in each workstation. To reduce the queuing time and increase the efficiency based on the analysis any probably improvement should be done. One probably action is by increasing the number of operators how manually operate this workstation.

Keywords: Production line, manufacturing, performance measurement, queuing theory.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3092
276 Design of Liquids Mixing Control System using Fuzzy Time Control Discrete Event Model for Industrial Applications

Authors: M.Saleem Khan, Khaled Benkrid

Abstract:

This paper presents a time control liquids mixing system in the tanks as an application of fuzzy time control discrete model. The system is designed for a wide range of industrial applications. The simulation design of control system has three inputs: volume, viscosity, and selection of product, along with the three external control adjustments for the system calibration or to take over the control of the system autonomously in local or distributed environment. There are four controlling elements: rotatory motor, grinding motor, heating and cooling units, and valves selection, each with time frame limit. The system consists of three controlled variables measurement through its sensing mechanism for feed back control. This design also facilitates the liquids mixing system to grind certain materials in tanks and mix with fluids under required temperature controlled environment to achieve certain viscous level. Design of: fuzzifier, inference engine, rule base, deffuzifiers, and discrete event control system, is discussed. Time control fuzzy rules are formulated, applied and tested using MATLAB simulation for the system.

Keywords: Fuzzy time control, industrial application and timecontrol systems, adjustment of Fuzzy system, liquids mixing system, design of fuzzy time control DEV system.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2490
275 Performance Analysis of Digital Signal Processors Using SMV Benchmark

Authors: Erh-Wen Hu, Cyril S. Ku, Andrew T. Russo, Bogong Su, Jian Wang

Abstract:

Unlike general-purpose processors, digital signal processors (DSP processors) are strongly application-dependent. To meet the needs for diverse applications, a wide variety of DSP processors based on different architectures ranging from the traditional to VLIW have been introduced to the market over the years. The functionality, performance, and cost of these processors vary over a wide range. In order to select a processor that meets the design criteria for an application, processor performance is usually the major concern for digital signal processing (DSP) application developers. Performance data are also essential for the designers of DSP processors to improve their design. Consequently, several DSP performance benchmarks have been proposed over the past decade or so. However, none of these benchmarks seem to have included recent new DSP applications. In this paper, we use a new benchmark that we recently developed to compare the performance of popular DSP processors from Texas Instruments and StarCore. The new benchmark is based on the Selectable Mode Vocoder (SMV), a speech-coding program from the recent third generation (3G) wireless voice applications. All benchmark kernels are compiled by the compilers of the respective DSP processors and run on their simulators. Weighted arithmetic mean of clock cycles and arithmetic mean of code size are used to compare the performance of five DSP processors. In addition, we studied how the performance of a processor is affected by code structure, features of processor architecture and optimization of compiler. The extensive experimental data gathered, analyzed, and presented in this paper should be helpful for DSP processor and compiler designers to meet their specific design goals.

Keywords: digital signal processors, DSP benchmark, instruction level parallelism, modified cyclomatic complexity, performance analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1553
274 A Review of Lortie’s Schoolteacher

Authors: Tsai-Hsiu Lin

Abstract:

Dan C. Lortie’s Schoolteacher: A sociological study is one of the best works on the sociology of teaching since W. Waller’s classic study. It is a book worthy of review. Following the tradition of symbolic interactionists, Lortie demonstrated the qualities who studied the occupation of teaching. Using several methods to gather effective data, Lortie has portrayed the ethos of the teaching profession. Therefore, the work is an important book on the teaching profession and teacher culture. Though outstanding, Lortie’s work is also flawed in that his perspectives and methodology were adopted largely from symbolic interactionism. First, Lortie in his work analyzed many points regarding teacher culture; for example, he was interested in exploring “sentiment,” “cathexis,” and “ethos.” Thus, he was more a psychologist than a sociologist. Second, symbolic interactionism led him to discern the teacher culture from a micro view, thereby missing the structural aspects. For example, he did not fully discuss the issue of gender and he ignored the issue of race. Finally, following the qualitative sociological tradition, Lortie employed many qualitative methods to gather data but only foucused on obtaining and presenting interview data. Moreover, he used measurement methods that were too simplistic for analyzing quantitative data fully.

Keywords: Lortie’s Schooltacher, Symbolic interactionism, teacher culture, teaching profession.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4172
273 Closed form Delay Model for on-Chip VLSIRLCG Interconnects for Ramp Input for Different Damping Conditions

Authors: Susmita Sahoo, Madhumanti Datta, Rajib Kar

Abstract:

Fast delay estimation methods, as opposed to simulation techniques, are needed for incremental performance driven layout synthesis. On-chip inductive effects are becoming predominant in deep submicron interconnects due to increasing clock speed and circuit complexity. Inductance causes noise in signal waveforms, which can adversely affect the performance of the circuit and signal integrity. Several approaches have been put forward which consider the inductance for on-chip interconnect modelling. But for even much higher frequency, of the order of few GHz, the shunt dielectric lossy component has become comparable to that of other electrical parameters for high speed VLSI design. In order to cope up with this effect, on-chip interconnect has to be modelled as distributed RLCG line. Elmore delay based methods, although efficient, cannot accurately estimate the delay for RLCG interconnect line. In this paper, an accurate analytical delay model has been derived, based on first and second moments of RLCG interconnection lines. The proposed model considers both the effect of inductance and conductance matrices. We have performed the simulation in 0.18μm technology node and an error of as low as less as 5% has been achieved with the proposed model when compared to SPICE. The importance of the conductance matrices in interconnect modelling has also been discussed and it is shown that if G is neglected for interconnect line modelling, then it will result an delay error of as high as 6% when compared to SPICE.

Keywords: Delay Modelling; On-Chip Interconnect; RLCGInterconnect; Ramp Input; Damping; VLSI

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1993
272 Recognition Machine (RM) for On-line and Isolated Flight Deck Officer (FDO) Gestures

Authors: Deniz T. Sodiri, Venkat V S S Sastry

Abstract:

The paper presents an on-line recognition machine (RM) for continuous/isolated, dynamic and static gestures that arise in Flight Deck Officer (FDO) training. RM is based on generic pattern recognition framework. Gestures are represented as templates using summary statistics. The proposed recognition algorithm exploits temporal and spatial characteristics of gestures via dynamic programming and Markovian process. The algorithm predicts corresponding index of incremental input data in the templates in an on-line mode. Accumulated consistency in the sequence of prediction provides a similarity measurement (Score) between input data and the templates. The algorithm provides an intuitive mechanism for automatic detection of start/end frames of continuous gestures. In the present paper, we consider isolated gestures. The performance of RM is evaluated using four datasets - artificial (W TTest), hand motion (Yang) and FDO (tracker, vision-based ). RM achieves comparable results which are in agreement with other on-line and off-line algorithms such as hidden Markov model (HMM) and dynamic time warping (DTW). The proposed algorithm has the additional advantage of providing timely feedback for training purposes.

Keywords: On-line Recognition Algorithm, IsolatedDynamic/Static Gesture Recognition, On-line Markovian/DynamicProgramming, Training in Virtual Environments.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1421
271 Impact of Ship Traffic to PM2.5 and Particle Number Concentrations in Three Port-Cities of the Adriatic/Ionian Area

Authors: Daniele Contini, Antonio Donateo, Andrea Gambaro, Athanasios Argiriou, Dimitrios Melas, Daniela Cesari, Anastasia Poupkou, Athanasios Karagiannidis, Apostolos Tsakis, Eva Merico, Rita Cesari, Adelaide Dinoi

Abstract:

Emissions of atmospheric pollutants from ships and harbour activities are a growing concern at international level given their potential impacts on air quality and climate. These close-to-land emissions have potential impact on local communities in terms of air quality and health. Recent studies show that the impact of maritime traffic to atmospheric particulate matter concentrations in several coastal urban areas is comparable with the impact of road traffic of a medium size town. However, several different approaches have been used for these estimates making difficult a direct comparison of results. In this work, an integrated approach based on emission inventories and dedicated measurement campaigns has been applied to give a comparable estimate of the impact of maritime traffic to PM2.5 and particle number concentrations in three major harbours of the Adriatic/Ionian Seas. The influences of local meteorology and of the logistic layout of the harbours are discussed.

Keywords: Ship emissions, PM2.5, particle number concentrations, impact of shipping to atmospheric aerosol.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2345
270 Reducing Variation of Dyeing Process in Textile Manufacturing Industry

Authors: M. Zeydan, G. Toğa

Abstract:

This study deals with a multi-criteria optimization problem which has been transformed into a single objective optimization problem using Response Surface Methodology (RSM), Artificial Neural Network (ANN) and Grey Relational Analyses (GRA) approach. Grey-RSM and Grey-ANN are hybrid techniques which can be used for solving multi-criteria optimization problem. There have been two main purposes of this research as follows. 1. To determine optimum and robust fiber dyeing process conditions by using RSM and ANN based on GRA, 2. To obtain the best suitable model by comparing models developed by different methodologies. The design variables for fiber dyeing process in textile are temperature, time, softener, anti-static, material quantity, pH, retarder, and dispergator. The quality characteristics to be evaluated are nominal color consistency of fiber, maximum strength of fiber, minimum color of dyeing solution. GRA-RSM with exact level value, GRA-RSM with interval level value and GRA-ANN models were compared based on GRA output value and MSE (Mean Square Error) performance measurement of outputs with each other. As a result, GRA-ANN with interval value model seems to be suitable reducing the variation of dyeing process for GRA output value of the model.

Keywords: Artificial Neural Network, Grey Relational Analysis, Optimization, Response Surface Methodology

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3498
269 Hydraulic Unbalance in Oil Injected Twin Rotary Screw Compressor Vibration Analysis (A Case History Related to Iran Oil Industries)

Authors: Omid A. Zargar

Abstract:

Vibration analysis of screw compressors is one of the most challenging cases in preventive maintenance. This kind of equipment considered as vibration bad actor facilities in industrial plants. On line condition monitoring systems developed too much in recent years. The high frequency vibration of ball bearings, gears, male and female caused complex fast Fourier transform (FFT) and time wave form (TWF) in screw compressors. The male and female randomly are sent to balance shop for balancing operation. This kind of operation usually caused some bending in rotors during the process that could cause further machining in such equipment. This kind of machining operation increased the vibration analysis complexity beside some process characteristic abnormality like inlet and out let pressure and temperature. In this paper mechanical principal and different type of screw compressors explained. Besides, some new condition monitoring systems and techniques for screw compressors discussed. Finally, one of the common behavior of oil injected twin rotary screw compressors called hydraulic unbalance that usually occurred after machining operation of male or female and have some specific characteristics in FFT and TWF discussed in details through a case history related to Iran oil industries.

Keywords: Vibration analysis, twin screw compressor, oil injected screw compressor, time wave form (TWF), fast Fourier transform (FFT), Hydraulic unbalance and rotor unbalance.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4531
268 Optimal Location of Multi Type Facts Devices for Multiple Contingencies Using Particle Swarm Optimization

Authors: S. Sutha, N. Kamaraj

Abstract:

In deregulated operating regime power system security is an issue that needs due thoughtfulness from researchers in the horizon of unbundling of generation and transmission. Electric power systems are exposed to various contingencies. Network contingencies often contribute to overloading of branches, violation of voltages and also leading to problems of security/stability. To maintain the security of the systems, it is desirable to estimate the effect of contingencies and pertinent control measurement can be taken on to improve the system security. This paper presents the application of particle swarm optimization algorithm to find the optimal location of multi type FACTS devices in a power system in order to eliminate or alleviate the line over loads. The optimizations are performed on the parameters, namely the location of the devices, their types, their settings and installation cost of FACTS devices for single and multiple contingencies. TCSC, SVC and UPFC are considered and modeled for steady state analysis. The selection of UPFC and TCSC suitable location uses the criteria on the basis of improved system security. The effectiveness of the proposed method is tested for IEEE 6 bus and IEEE 30 bus test systems.

Keywords: Contingency Severity Index, Particle Swarm Optimization, Performance Index, Static Security Assessment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2720
267 Hierarchies Based On the Number of Cooperating Systems of Finite Automata on Four-Dimensional Input Tapes

Authors: Makoto Sakamoto, Yasuo Uchida, Makoto Nagatomo, Takao Ito, Tsunehiro Yoshinaga, Satoshi Ikeda, Masahiro Yokomichi, Hiroshi Furutani

Abstract:

In theoretical computer science, the Turing machine has played a number of important roles in understanding and exploiting basic concepts and mechanisms in computing and information processing [20]. It is a simple mathematical model of computers [9]. After that, M.Blum and C.Hewitt first proposed two-dimensional automata as a computational model of two-dimensional pattern processing, and investigated their pattern recognition abilities in 1967 [7]. Since then, a lot of researchers in this field have been investigating many properties about automata on a two- or three-dimensional tape. On the other hand, the question of whether processing fourdimensional digital patterns is much more difficult than two- or threedimensional ones is of great interest from the theoretical and practical standpoints. Thus, the study of four-dimensional automata as a computasional model of four-dimensional pattern processing has been meaningful [8]-[19],[21]. This paper introduces a cooperating system of four-dimensional finite automata as one model of four-dimensional automata. A cooperating system of four-dimensional finite automata consists of a finite number of four-dimensional finite automata and a four-dimensional input tape where these finite automata work independently (in parallel). Those finite automata whose input heads scan the same cell of the input tape can communicate with each other, that is, every finite automaton is allowed to know the internal states of other finite automata on the same cell it is scanning at the moment. In this paper, we mainly investigate some accepting powers of a cooperating system of eight- or seven-way four-dimensional finite automata. The seven-way four-dimensional finite automaton is an eight-way four-dimensional finite automaton whose input head can move east, west, south, north, up, down, or in the fu-ture, but not in the past on a four-dimensional input tape.

Keywords: computational complexity, cooperating system, finite automaton, four-dimension, hierarchy, multihead.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1833
266 Titania and Cu-Titania Composite Layer on Graphite Substrate as Negative Electrode for Li-Ion Battery

Authors: Fitria Rahmawati, Nuryani, Liviana Wijayanti

Abstract:

This research study the application of the immobilized TiO2 layer and Cu-TiO2 layer on graphite substrate as a negative electrode or anode for Li-ion battery. The titania layer was produced through chemical bath deposition method, meanwhile Cu particles were deposited electrochemically. A material can be used as an electrode as it has capability to intercalates Li ions into its crystal structure. The Li intercalation into TiO2/Graphite and Cu- TiO2/Graphite were analyzed from the changes of its XRD pattern after it was used as electrode during discharging process. The XRD patterns were refined by Le Bail method in order to determine the crystal structure of the prepared materials. A specific capacity and the cycle ability measurement were carried out to study the performance of the prepared materials as negative electrode of the Li-ion battery. The specific capacity was measured during discharging process from fully charged until the cut off voltage. A 300 was used as a load. The result shows that the specific capacity of Li-ion battery with TiO2/Graphite as negative electrode is 230.87 ± 1.70mAh.g-1 which is higher than the specific capacity of Li-ion battery with pure graphite as negative electrode, i.e 140.75 ±0.46mAh.g-1. Meanwhile deposition of Cu onto TiO2 layer does not increase the specific capacity, and the value even lower than the battery with TiO2/Graphite as electrode. The cycle ability of the prepared battery is only two cycles, due to the Li ribbon which was used as cathode became fragile and easily broken.

Keywords: Cu-TiO2, electrode, graphite substrate, Li-ion battery, TiO2 layer.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1912
265 New Method for Determining the Distribution of Birefringence and Linear Dichroism in Polymer Materials Based On Polarization-Holographic Grating

Authors: Barbara Kilosanidze, George Kakauridze, Levan Nadareishvili, Yuri Mshvenieradze

Abstract:

A new method for determining the distribution of birefringence and linear dichroism in optical polymer materials is presented. The method is based on the use of polarizationholographic diffraction grating that forms an orthogonal circular basis in the process of diffraction of probing laser beam on the grating. The intensities ratio of the orders of diffraction on this grating enables the value of birefringence and linear dichroism in the sample to be determined. The distribution of birefringence in the sample is determined by scanning with a circularly polarized beam with a wavelength far from the absorption band of the material. If the scanning is carried out by probing beam with the wavelength near to a maximum of the absorption band of the chromophore then the distribution of linear dichroism can be determined. An appropriate theoretical model of this method is presented. A laboratory setup was created for the proposed method. An optical scheme of the laboratory setup is presented. The results of measurement in polymer films with two-dimensional gradient distribution of birefringence and linear dichroism are discussed.

Keywords: Birefringence, graded oriented polymers, linear dichroism, optical polymers, optical anisotropy, polarization-holographic grating,

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1471
264 Cumulative Learning based on Dynamic Clustering of Hierarchical Production Rules(HPRs)

Authors: Kamal K.Bharadwaj, Rekha Kandwal

Abstract:

An important structuring mechanism for knowledge bases is building clusters based on the content of their knowledge objects. The objects are clustered based on the principle of maximizing the intraclass similarity and minimizing the interclass similarity. Clustering can also facilitate taxonomy formation, that is, the organization of observations into a hierarchy of classes that group similar events together. Hierarchical representation allows us to easily manage the complexity of knowledge, to view the knowledge at different levels of details, and to focus our attention on the interesting aspects only. One of such efficient and easy to understand systems is Hierarchical Production rule (HPRs) system. A HPR, a standard production rule augmented with generality and specificity information, is of the following form Decision If < condition> Generality Specificity . HPRs systems are capable of handling taxonomical structures inherent in the knowledge about the real world. In this paper, a set of related HPRs is called a cluster and is represented by a HPR-tree. This paper discusses an algorithm based on cumulative learning scenario for dynamic structuring of clusters. The proposed scheme incrementally incorporates new knowledge into the set of clusters from the previous episodes and also maintains summary of clusters as Synopsis to be used in the future episodes. Examples are given to demonstrate the behaviour of the proposed scheme. The suggested incremental structuring of clusters would be useful in mining data streams.

Keywords: Cumulative learning, clustering, data mining, hierarchical production rules.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1393