Search results for: quality approaches
1585 Extensiveness and Effectiveness of Corporate Governance Regulations in South-Eastern Europe
Authors: Ana Bobirca, Paul-Gabriel Miclaus
Abstract:
The purpose of the article is to illustrate the main characteristics of the corporate governance challenge facing the countries of South-Eastern Europe (SEE) and to subsequently determine and assess the extensiveness and effectiveness of corporate governance regulations in these countries. Therefore, we start with an overview on the subject of the key problems of corporate governance in transition. We then address the issue of corporate governance measurement for SEE countries. To this end, we include a review of the methodological framework for determining both the extensiveness and the effectiveness of corporate governance legislation. We then focus on the actual analysis of the quality of corporate governance codes, as well as of legal institutions effectiveness and provide a measure of corporate governance in Romania and other SEE emerging markets. The paper concludes by emphasizing the corporate governance enforcement gap and by identifying research issues that require further study.Keywords: corporate governance, effectiveness, extensiveness, South-Eastern Europe
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16701584 Spatial Distribution of Cd, Zn and Hg in Groundwater at Rayong Province, Thailand
Authors: T. Makkasap, T. Satapanajaru
Abstract:
The objective of this study was to evaluate the distribution patterns of Cd, Zn and Hg in groundwater by geospatial interpolation. The study was performed at Rayong province in the eastern part of Thailand, with high agricultural and industrial activities. Groundwater samples were collected twice a year from 31 tubewells around this area. Inductively Coupled Plasma-Atomic Emission Spectrometer (ICP-AES) was used to measure the concentrations of Cd, Zn, and Hg in groundwater samples. The results demonstrated that concentrations of Cd, Zn and Hg range from 0.000-0.297 mg/L (x = 0.021±0.033 mg/L), 0.022-33.236 mg/L (x = 4.214±4.766 mg/L) and 0.000-0.289 mg/L (x = 0.023±0.034 mg/L), respectively. Most of the heavy metals concentrations were exceeded groundwater quality standards as specified in the Ministry of Natural Resources and Environment, Thailand. The trend distribution of heavy metals were high concentrations at the southeastern part of the area that especially vulnerable to heavy metals and other contaminants.Keywords: Groundwater, Heavy metals, Kriging, Rayong, Spatial distribution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19531583 Performance Analysis of a Hybrid DF-AF Hybrid RF/FSO System under Gamma Gamma Atmospheric Turbulence Channel Using MPPM Modulation
Authors: Hechmi Saidi, Noureddine Hamdi
Abstract:
The performance of hybrid amplify and forward - decode and forward (AF-DF) hybrid radio frequency/free space optical (RF/FSO) communication system, that adopts M-ary pulse position modulation (MPPM) techniques, is analyzed. Both exact and approximate symbol-error rates (SERs) are derived. The random variations of the received optical irradiance, produced by the atmospheric turbulence, is modeled by the gamma-gamma (GG) statistical distribution. A closed-form expression for the probability density function (PDF) is derived for the whole above system is obtained. Thanks to the use of hybrid AF-DF hybrid RF/FSO configuration and MPPM, the effects of atmospheric turbulence is mitigated; hence the capacity of combating atmospheric turbulence and the transmissitted signal quality are improved.Keywords: FSO, RF, hybrid, AF, DF, SER, SNR, GG channel.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10701582 K-Means for Spherical Clusters with Large Variance in Sizes
Authors: A. M. Fahim, G. Saake, A. M. Salem, F. A. Torkey, M. A. Ramadan
Abstract:
Data clustering is an important data exploration technique with many applications in data mining. The k-means algorithm is well known for its efficiency in clustering large data sets. However, this algorithm is suitable for spherical shaped clusters of similar sizes and densities. The quality of the resulting clusters decreases when the data set contains spherical shaped with large variance in sizes. In this paper, we introduce a competent procedure to overcome this problem. The proposed method is based on shifting the center of the large cluster toward the small cluster, and recomputing the membership of small cluster points, the experimental results reveal that the proposed algorithm produces satisfactory results.Keywords: K-Means, Data Clustering, Cluster Analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32851581 Analyzing and Formulation of Product Lead Time
Authors: B. Fahimnia, L.H.S. Luong, B. Motevallian, R. M. Marian, M. M. Esmaeil
Abstract:
Product Lead Time (PLT) is the period of time from receiving a customer's order to delivering the final product. PLT is an indicator of the manufacturing controllability, efficiency and performance. Due to the explosion in the rate of technological innovations and the rapid changes in the nature of manufacturing processes, manufacturing firms can bring the new products to market quicker only if they can reduce their PLT and speed up the rate at which they can design, plan, control, and manufacture. Although there is a substantial body of research on manufacturing relating to cost and quality issues, there is no much specific research conducted in relation to the formulation of PLT, despite its significance and importance. This paper analyzes and formulates PLT which can be used as a guideline for achieving the shorter PLT. Further more this paper identifies the causes of delay and factors that contributes to the increased product lead-time.Keywords: Manufacturing Control, Manufacturing Lead Time, Manufacturing Planning, Product Design, and Product Lead Time.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17691580 The Impact of Regulatory Changes on the Development of Mobile Medical Apps
Abstract:
Mobile applications are being used to perform a wide variety of tasks in day-to-day life, ranging from checking email to controlling your home heating. Application developers have recognized the potential to transform a smart device into a medical device, by using a mobile medical application i.e. a mobile phone or a tablet. When initially conceived these mobile medical applications performed basic functions e.g. BMI calculator, accessing reference material etc.; however, increasing complexity offers clinicians and patients a range of functionality. As this complexity and functionality increases, so too does the potential risk associated with using such an application. Examples include any applications that provide the ability to inflate and deflate blood pressure cuffs, as well as applications that use patient-specific parameters and calculate dosage or create a dosage plan for radiation therapy. If an unapproved mobile medical application is marketed by a medical device organization, then they face significant penalties such as receiving an FDA warning letter to cease the prohibited activity, fines and possibility of facing a criminal conviction. Regulatory bodies have finalized guidance intended for mobile application developers to establish if their applications are subject to regulatory scrutiny. However, regulatory controls appear contradictory with the approaches taken by mobile application developers who generally work with short development cycles and very little documentation and as such, there is the potential to stifle further improvements due to these regulations. The research presented as part of this paper details how by adopting development techniques, such as agile software development, mobile medical application developers can meet regulatory requirements whilst still fostering innovation.
Keywords: Medical, mobile, applications, software Engineering, FDA, standards, regulations, agile.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20711579 Automatic Staging and Subtype Determination for Non-Small Cell Lung Carcinoma Using PET Image Texture Analysis
Authors: Seyhan Karaçavuş, Bülent Yılmaz, Ömer Kayaaltı, Semra İçer, Arzu Taşdemir, Oğuzhan Ayyıldız, Kübra Eset, Eser Kaya
Abstract:
In this study, our goal was to perform tumor staging and subtype determination automatically using different texture analysis approaches for a very common cancer type, i.e., non-small cell lung carcinoma (NSCLC). Especially, we introduced a texture analysis approach, called Law’s texture filter, to be used in this context for the first time. The 18F-FDG PET images of 42 patients with NSCLC were evaluated. The number of patients for each tumor stage, i.e., I-II, III or IV, was 14. The patients had ~45% adenocarcinoma (ADC) and ~55% squamous cell carcinoma (SqCCs). MATLAB technical computing language was employed in the extraction of 51 features by using first order statistics (FOS), gray-level co-occurrence matrix (GLCM), gray-level run-length matrix (GLRLM), and Laws’ texture filters. The feature selection method employed was the sequential forward selection (SFS). Selected textural features were used in the automatic classification by k-nearest neighbors (k-NN) and support vector machines (SVM). In the automatic classification of tumor stage, the accuracy was approximately 59.5% with k-NN classifier (k=3) and 69% with SVM (with one versus one paradigm), using 5 features. In the automatic classification of tumor subtype, the accuracy was around 92.7% with SVM one vs. one. Texture analysis of FDG-PET images might be used, in addition to metabolic parameters as an objective tool to assess tumor histopathological characteristics and in automatic classification of tumor stage and subtype.Keywords: Cancer stage, cancer cell type, non-small cell lung carcinoma, PET, texture analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9851578 Thresholding Approach for Automatic Detection of Pseudomonas aeruginosa Biofilms from Fluorescence in situ Hybridization Images
Authors: Zonglin Yang, Tatsuya Akiyama, Kerry S. Williamson, Michael J. Franklin, Thiruvarangan Ramaraj
Abstract:
Pseudomonas aeruginosa is an opportunistic pathogen that forms surface-associated microbial communities (biofilms) on artificial implant devices and on human tissue. Biofilm infections are difficult to treat with antibiotics, in part, because the bacteria in biofilms are physiologically heterogeneous. One measure of biological heterogeneity in a population of cells is to quantify the cellular concentrations of ribosomes, which can be probed with fluorescently labeled nucleic acids. The fluorescent signal intensity following fluorescence in situ hybridization (FISH) analysis correlates to the cellular level of ribosomes. The goals here are to provide computationally and statistically robust approaches to automatically quantify cellular heterogeneity in biofilms from a large library of epifluorescent microscopy FISH images. In this work, the initial steps were developed toward these goals by developing an automated biofilm detection approach for use with FISH images. The approach allows rapid identification of biofilm regions from FISH images that are counterstained with fluorescent dyes. This methodology provides advances over other computational methods, allowing subtraction of spurious signals and non-biological fluorescent substrata. This method will be a robust and user-friendly approach which will enable users to semi-automatically detect biofilm boundaries and extract intensity values from fluorescent images for quantitative analysis of biofilm heterogeneity.
Keywords: Image informatics, Pseudomonas aeruginosa, biofilm, FISH, computer vision, data visualization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11951577 Detecting Abnormal ECG Signals Utilising Wavelet Transform and Standard Deviation
Authors: Dejan Stantic, Jun Jo
Abstract:
ECG contains very important clinical information about the cardiac activities of the heart. Often the ECG signal needs to be captured for a long period of time in order to identify abnormalities in certain situations. Such signal apart of a large volume often is characterised by low quality due to the noise and other influences. In order to extract features in the ECG signal with time-varying characteristics at first need to be preprocessed with the best parameters. Also, it is useful to identify specific parts of the long lasting signal which have certain abnormalities and to direct the practitioner to those parts of the signal. In this work we present a method based on wavelet transform, standard deviation and variable threshold which achieves 100% accuracy in identifying the ECG signal peaks and heartbeat as well as identifying the standard deviation, providing a quick reference to abnormalities.
Keywords: Electrocardiogram-ECG, Arrhythmia, Signal Processing, Wavelet Transform, Standard Deviation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29141576 Characterization of the Airtightness Level in School Classrooms in Mediterranean Climate
Authors: Miguel A. Campano, Jesica Fernández-Agüera, Samuel Domínguez-Amarillo, Juan J. Sendra
Abstract:
An analysis of the air tightness level is performed on a representative sample of school classrooms in Southern Spain, which allows knowing the infiltration level of these classrooms, mainly through its envelope, which can affect both energy demand and occupant's thermal comfort. By using a pressurization/depressurization equipment (Blower-Door test), a characterization of 45 multipurpose classrooms have been performed in nine non-university educational institutions of the main climate zones of Southern Spain. In spite of having two doors and a high ratio between glass surface and outer surface, it is possible to see in these classrooms that there is an adequate level of airtightness, since all the n50 values obtained are lower than 9.0 ACH, with an average value around 7.0 ACH.Keywords: Air infiltration, energy efficiency, school buildings, thermal comfort, indoor air quality, ventilation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18551575 Bridging the Mental Gap between Convolution Approach and Compartmental Modeling in Functional Imaging: Typical Embedding of an Open Two-Compartment Model into the Systems Theory Approach of Indicator Dilution Theory
Authors: Gesine Hellwig
Abstract:
Functional imaging procedures for the non-invasive assessment of tissue microcirculation are highly requested, but require a mathematical approach describing the trans- and intercapillary passage of tracer particles. Up to now, two theoretical, for the moment different concepts have been established for tracer kinetic modeling of contrast agent transport in tissues: pharmacokinetic compartment models, which are usually written as coupled differential equations, and the indicator dilution theory, which can be generalized in accordance with the theory of lineartime- invariant (LTI) systems by using a convolution approach. Based on mathematical considerations, it can be shown that also in the case of an open two-compartment model well-known from functional imaging, the concentration-time course in tissue is given by a convolution, which allows a separation of the arterial input function from a system function being the impulse response function, summarizing the available information on tissue microcirculation. Due to this reason, it is possible to integrate the open two-compartment model into the system-theoretic concept of indicator dilution theory (IDT) and thus results known from IDT remain valid for the compartment approach. According to the long number of applications of compartmental analysis, even for a more general context similar solutions of the so-called forward problem can already be found in the extensively available appropriate literature of the seventies and early eighties. Nevertheless, to this day, within the field of biomedical imaging – not from the mathematical point of view – there seems to be a trench between both approaches, which the author would like to get over by exemplary analysis of the well-known model.
Keywords: Functional imaging, Tracer kinetic modeling, LTIsystem, Indicator dilution theory / convolution approach, Two-Compartment model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14211574 A World Map of Seabed Sediment Based on 50 Years of Knowledge
Authors: T. Garlan, I. Gabelotaud, S. Lucas, E. Marchès
Abstract:
Production of a global sedimentological seabed map has been initiated in 1995 to provide the necessary tool for searches of aircraft and boats lost at sea, to give sedimentary information for nautical charts, and to provide input data for acoustic propagation modelling. This original approach had already been initiated one century ago when the French hydrographic service and the University of Nancy had produced maps of the distribution of marine sediments of the French coasts and then sediment maps of the continental shelves of Europe and North America. The current map of the sediment of oceans presented was initiated with a UNESCO's general map of the deep ocean floor. This map was adapted using a unique sediment classification to present all types of sediments: from beaches to the deep seabed and from glacial deposits to tropical sediments. In order to allow good visualization and to be adapted to the different applications, only the granularity of sediments is represented. The published seabed maps are studied, if they present an interest, the nature of the seabed is extracted from them, the sediment classification is transcribed and the resulted map is integrated in the world map. Data come also from interpretations of Multibeam Echo Sounder (MES) imagery of large hydrographic surveys of deep-ocean. These allow a very high-quality mapping of areas that until then were represented as homogeneous. The third and principal source of data comes from the integration of regional maps produced specifically for this project. These regional maps are carried out using all the bathymetric and sedimentary data of a region. This step makes it possible to produce a regional synthesis map, with the realization of generalizations in the case of over-precise data. 86 regional maps of the Atlantic Ocean, the Mediterranean Sea, and the Indian Ocean have been produced and integrated into the world sedimentary map. This work is permanent and permits a digital version every two years, with the integration of some new maps. This article describes the choices made in terms of sediment classification, the scale of source data and the zonation of the variability of the quality. This map is the final step in a system comprising the Shom Sedimentary Database, enriched by more than one million punctual and surface items of data, and four series of coastal seabed maps at 1:10,000, 1:50,000, 1:200,000 and 1:1,000,000. This step by step approach makes it possible to take into account the progresses in knowledge made in the field of seabed characterization during the last decades. Thus, the arrival of new classification systems for seafloor has improved the recent seabed maps, and the compilation of these new maps with those previously published allows a gradual enrichment of the world sedimentary map. But there is still a lot of work to enhance some regions, which are still based on data acquired more than half a century ago.
Keywords: Marine sedimentology, seabed map, sediment classification, World Ocean.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10491573 Investigation on Novel Based Metaheuristic Algorithms for Combinatorial Optimization Problems in Ad Hoc Networks
Authors: C. Rajan, N. Shanthi, C. Rasi Priya, K. Geetha
Abstract:
Routing in MANET is extremely challenging because of MANETs dynamic features, its limited bandwidth, frequent topology changes caused by node mobility and power energy consumption. In order to efficiently transmit data to destinations, the applicable routing algorithms must be implemented in mobile ad-hoc networks. Thus we can increase the efficiency of the routing by satisfying the Quality of Service (QoS) parameters by developing routing algorithms for MANETs. The algorithms that are inspired by the principles of natural biological evolution and distributed collective behavior of social colonies have shown excellence in dealing with complex optimization problems and are becoming more popular. This paper presents a survey on few meta-heuristic algorithms and naturally-inspired algorithms.
Keywords: Ant colony optimization, genetic algorithm, Naturally-inspired algorithms and particle swarm optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27081572 Precision Identification of Nonlinear Damping Parameter for a Miniature Moving-Coil Transducer
Authors: Yu-Ting Tsai, Yu-da Lee, Jin H. Huang
Abstract:
The nonlinear damping behavior is usually ignored in the design of a miniature moving-coil loudspeaker. But when the loudspeaker operated in air, the damping parameter varies with the voice-coil displacement corresponding due to viscous air flow. The present paper presents an identification model as inverse problem to identify the nonlinear damping parameter in the lumped parameter model for the loudspeaker. Theoretical results for the nonlinear damping are verified by using laser displacement measurement scanner. These results indicate that the damping parameter has the greatly different nonlinearity between in air and vacuum. It is believed that the results of the present work can be applied in diagnosis and sound quality improvement of a miniature loudspeaker.
Keywords: Miniature loudspeaker, non-linear damping, system identification, Lumped parameter model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19861571 Instance-Based Ontology Matching Using Different Kinds of Formalism
Authors: Katrin Zaiß, Tim Schlüter, Stefan Conrad
Abstract:
Ontology Matching is a task needed in various applica-tions, for example for comparison or merging purposes. In literature,many algorithms solving the matching problem can be found, butmost of them do not consider instances at all. Mappings are deter-mined by calculating the string-similarity of labels, by recognizinglinguistic word relations (synonyms, subsumptions etc.) or by ana-lyzing the (graph) structure. Due to the facts that instances are oftenmodeled within the ontology and that the set of instances describesthe meaning of the concepts better than their meta information,instances should definitely be incorporated into the matching process.In this paper several novel instance-based matching algorithms arepresented which enhance the quality of matching results obtainedwith common concept-based methods. Different kinds of formalismsare use to classify concepts on account of their instances and finallyto compare the concepts directly.KeywordsInstances, Ontology Matching, Semantic Web
Keywords: Instances, Ontology Matching, Semantic Web
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15291570 A Dynamic Model for a Drill in the Drilling Process
Authors: Bo Wun Huang, Ah Der Lin, Yung Chuan Chen, Jao Hwa Kuang
Abstract:
The dynamic model of a drill in drilling process was proposed and investigated in this study. To assure a good drilling quality, the vibration variation on the drill tips during high speed drilling is needed to be investigated. A pre-twisted beam is used to simulate the drill. The moving Winkler-Type elastic foundation is used to characterize the tip boundary variation in drilling. Due to the variation of the drill depth, a time dependent dynamic model for the drill is proposed. Results simulated from this proposed model indicate that an abrupt natural frequencies drop are experienced as the drill tip tough the workpiece, and a severe vibration is induced. The effects of parameters, e.g. drilling speed, depth, drill size and thrust force on the drill tip responses studied.Keywords: Drilling, vibration of drill, twisted beam
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15201569 The Current Home Hemodialysis Practices and Patients’ Safety Related Factors: A Case Study from Germany
Authors: Ilyas Khan. Liliane Pintelon, Harry Martin, Michael Shömig
Abstract:
The increasing costs of healthcare on one hand, and the rise in aging population and associated chronic disease, on the other hand, are putting increasing burden on the current health care system in many Western countries. For instance, chronic kidney disease (CKD) is a common disease and in Europe, the cost of renal replacement therapy (RRT) is very significant to the total health care cost. However, the recent advancement in healthcare technology, provide the opportunity to treat patients at home in their own comfort. It is evident that home healthcare offers numerous advantages apparently, low costs and high patients’ quality of life. Despite these advantages, the intake of home hemodialysis (HHD) therapy is still low in particular in Germany. Many factors are accounted for the low number of HHD intake. However, this paper is focusing on patients’ safety-related factors of current HHD practices in Germany. The aim of this paper is to analyze the current HHD practices in Germany and to identify risks related factors if any exist. A case study has been conducted in a dialysis center which consists of four dialysis centers in the south of Germany. In total, these dialysis centers have 350 chronic dialysis patients, of which, four patients are on HHD. The centers have 126 staff which includes six nephrologists and 120 other staff i.e. nurses and administration. The results of the study revealed several risk-related factors. Most importantly, these centers do not offer allied health services at the pre-dialysis stage, the HHD training did not have an established curriculum; however, they have just recently developed the first version. Only a soft copy of the machine manual is offered to patients. Surprisingly, the management was not aware of any standard available for home assessment and installation. The home assessment is done by a third party (i.e. the machines and equipment provider) and they may not consider the hygienic quality of the patient’s home. The type of machine provided to patients at home is similar to the one in the center. The model may not be suitable at home because of its size and complexity. Even though portable hemodialysis machines, which are specially designed for home use, are available in the market such as the NxStage series. Besides the type of machine, no assistance is offered for space management at home in particular for placing the machine. Moreover, the centers do not offer remote assistance to patients and their carer at home. However, telephonic assistance is available. Furthermore, no alternative is offered if a carer is not available. In addition, the centers are lacking medical staff including nephrologists and renal nurses.
Keywords: Home hemodialysis, home hemodialysis practices, patients’ related risks in the current home hemodialysis practices, patient safety in home hemodialysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8171568 A Stochastic Approach to Extreme Wind Speeds Conditions on a Small Axial Wind Turbine
Authors: Nkongho Ayuketang Arreyndip, Ebobenow Joseph
Abstract:
In this paper, to model a real life wind turbine, a probabilistic approach is proposed to model the dynamics of the blade elements of a small axial wind turbine under extreme stochastic wind speeds conditions. It was found that the power and the torque probability density functions even-dough decreases at these extreme wind speeds but are not infinite. Moreover, we also fund that it is possible to stabilize the power coefficient (stabilizing the output power)above rated wind speeds by turning some control parameters. This method helps to explain the effect of turbulence on the quality and quantity of the harness power and aerodynamic torque.Keywords: Probability, Stochastic, Probability density function, Turbulence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17511567 A New Technique for Progressive ECG Transmission using Discrete Radon Transform
Authors: Amine Naït-Ali
Abstract:
The aim of this paper is to present a new method which can be used for progressive transmission of electrocardiogram (ECG). The idea consists in transforming any ECG signal to an image, containing one beat in each row. In the first step, the beats are synchronized in order to reduce the high frequencies due to inter-beat transitions. The obtained image is then transformed using a discrete version of Radon Transform (DRT). Hence, transmitting the ECG, leads to transmit the most significant energy of the transformed image in Radon domain. For decoding purpose, the receptor needs to use the inverse Radon Transform as well as the two synchronization frames. The presented protocol can be adapted for lossy to lossless compression systems. In lossy mode we show that the compression ratio can be multiplied by an average factor of 2 for an acceptable quality of reconstructed signal. These results have been obtained on real signals from MIT database.Keywords: Discrete Radon Transform, ECG compression, synchronization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14311566 Conceptual Design of a Wi-Fi and GPS Based Robotic Library Using an Intelligent System
Authors: M. S. Sreejith, Steffy Joy, Abhishesh Pal, Beom-Sahng Ryuh, V. R. Sanal Kumar
Abstract:
In this paper, an attempt has been made for the design of a robotic library using an intelligent system. The robot works on the ARM microprocessor, motor driver circuit with 5 degrees of freedom with Wi-Fi and GPS based communication protocol. The authenticity of the library books is controlled by RFID. The proposed robotic library system is facilitated with embedded system and ARM. In this library issuance system, the previous potential readers’ authentic review reports have been taken into consideration for recommending suitable books to the deserving new users and the issuance of books or periodicals is based on the users’ decision. We have conjectured that the Wi-Fi based robotic library management system would allow fast transaction of books issuance and it also produces quality readers.Keywords: GPS based based robotic library, library management system, robotic library, Wi-Fi library.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22921565 Distinctive Features of Legal Relations in the Area of Subsoil Use, Renewal and Protection in Ukraine
Authors: N. Maksimentseva
Abstract:
The issue of public administration in subsoil use, renewal and protection is of high importance for Ukraine since it is strongly linked to energy security of the state as well as it shall facilitate the people of Ukraine to efficiently implement its propitiatory rights towards natural resources and redistribution of national wealth. As it is stipulated in the Article 11 of the Subsoil Code of Ukraine (the Code) the authorities that administer the industry are limited to central executive bodies and local governments. In particular, it is stipulated in the Code that the Ukraine’s Cabinet of Ministers carries out public administration in geological exploration, production and protection of subsoil. Other state bodies of public administration include central public authority responsible for state environmental protection policies; central public authority in charge of implementation of state geological exploration and efficient subsoil use policies; central authority in charge of state health and safety control policies. There are also public authorities in the Autonomous Republic of Crimea; local executive bodies and other state authorities and local self-government authorities in compliance with laws of Ukraine. This article is devoted to the analysis of the legal relations in the area of public administration of subsoil use, renewal and protection in Ukraine. The main approaches to study the essence of legal relations in the named area as well as its tasks, functions and methods are analyzed. It is concluded in this article that legal relationship in the field of public administration of subsoil use, renewal and protection is characterized by specifics of its task (development of natural resources).
Keywords: Legal relations, public administration, Subsoil Code of Ukraine, subsoil use, renewal and protection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10981564 Energetic Considerations for Sputter Deposition Processes
Authors: Dirk Hegemann, Martin Amberg
Abstract:
Sputter deposition processes, especially for sputtering from metal targets, are well investigated. For practical reasons, i.e. for industrial processes, energetic considerations for sputter deposition are useful in order to optimize the sputtering process. In particular, for substrates at floating conditions it is required to obtain energetic conditions during film growth that enables sufficient dense metal films of good quality. The influence of ion energies, energy density and momentum transfer is thus examined both for sputtering at the target as well as during film growth. Different regimes dominated by ion energy, energy density and momentum transfer were identified by using different plasma sources and by varying power input, pressure and bias voltage.
Keywords: Energy density, film growth, momentum transfer, sputtering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24501563 Optimized Weight Vector for QoS Aware Web Service Selection Algorithm Using Particle Swarm Optimization
Authors: N. Arulanand, P. M. Ananth
Abstract:
Quality of Service (QoS) attributes as part of the service description is an important factor for service attribute. It is not easy to exactly quantify the weight of each QoS conditions since human judgments based on their preference causes vagueness. As web services selection requires optimization, evolutionary computing based on heuristics to select an optimal solution is adopted. In this work, the evolutionary computing technique Particle Swarm Optimization (PSO) is used for selecting a suitable web services based on the user’s weightage of each QoS values by optimizing the QoS weight vector and thereby finding the best weight vectors for best services that is being selected. Finally the results are compared and analyzed using static inertia weight and deterministic inertia weight of PSO.Keywords: QoS, Optimization, Particle Swarm Optimization (PSO), weight vector, web services, web service selection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20201562 Enhancing Hand Efficiency of Smart Glass Cleaning Robot through Generative Design Module
Authors: Pankaj Gupta, Amit Kumar Srivastava, Nitesh Pandey
Abstract:
This article explores the domain of generative design in order to enhance the development of robot designs for innovative and efficient maintenance approaches for tall buildings. This study aims to optimize the design of robotic hands by focusing on minimizing mass and volume while ensuring they can withstand the specified pressure with equal strength. The research procedure is structured and systematic. The purpose of optimization is to enhance the efficiency of the robot and reduce the manufacturing expenses. The project seeks to investigate the application of generative design in order to optimize products. Autodesk Fusion 360 offers the capability to immediately apply the generative design functionality to the solid model. The effort involved creating a solid model of the Smart Glass Cleaning Robot and optimizing one of its components, the Hand, using generative techniques. The article has thoroughly examined the designs, outcomes, and procedure. These loads serve as a benchmark for creating designs that can endure the necessary level of pressure and preserve their structural integrity. The efficacy of the generative design process is contingent upon the selection of materials, as different materials possess distinct physical attributes. The study utilizes five different materials, namely Steel, Stainless Steel, Titanium, Aluminum, and CFRP (Carbon Fiber Reinforced Polymer), in order to investigate a range of design possibilities.
Keywords: Generative design, mass and volume optimization, material strength analysis, generative design, smart glass cleaning robot.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2211561 Calculus Logarithmic Function for Image Encryption
Authors: Adil AL-Rammahi
Abstract:
When we prefer to make the data secure from various attacks and fore integrity of data, we must encrypt the data before it is transmitted or stored. This paper introduces a new effective and lossless image encryption algorithm using a natural logarithmic function. The new algorithm encrypts an image through a three stage process. In the first stage, a reference natural logarithmic function is generated as the foundation for the encryption image. The image numeral matrix is then analyzed to five integer numbers, and then the numbers’ positions are transformed to matrices. The advantages of this method is useful for efficiently encrypting a variety of digital images, such as binary images, gray images, and RGB images without any quality loss. The principles of the presented scheme could be applied to provide complexity and then security for a variety of data systems such as image and others.
Keywords: Linear Systems, Image Encryption, Calculus.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24051560 Improvement of Bit-Error-Rate in Optical Fiber Receivers
Authors: Hadj Bourdoucen, Amer Alhabsi
Abstract:
In this paper, a post processing scheme is suggested for improvement of Bit Error-Rate (BER) in optical fiber transmission receivers. The developed scheme has been tested on optical fiber systems operating with a non-return-to-zero (NRZ) format at transmission rates of up to 10Gbps. The transmission system considered is based on well known transmitters and receivers blocks operating at wavelengths in the region of 1550 nm using a standard single mode fiber. Performance of improved detected signals has been evaluated via the analysis of quality factor and computed bit error rates. Numerical simulations have shown a noticeable improvement of the system BER after implementation of the suggested post processing operation on the detected electrical signals.Keywords: BER improvement, Optical fiber, transmissionperformance, NRZ.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21551559 Reversible Watermarking for H.264/AVC Videos
Authors: Yih-Chuan Lin, Jung-Hong Li
Abstract:
In this paper, we propose a reversible watermarking scheme based on histogram shifting (HS) to embed watermark bits into the H.264/AVC standard videos by modifying the last nonzero level in the context adaptive variable length coding (CAVLC) domain. The proposed method collects all of the last nonzero coefficients (or called last level coefficient) of 4×4 sub-macro blocks in a macro block and utilizes predictions for the current last level from the neighbor block-s last levels to embed watermark bits. The feature of the proposed method is low computational and has the ability of reversible recovery. The experimental results have demonstrated that our proposed scheme has acceptable degradation on video quality and output bit-rate for most test videos.Keywords: Reversible data hiding, H.264/AVC standard, CAVLC, Histogram shifting
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20361558 Multi-Agent Systems Applied in the Modeling and Simulation of Biological Problems: A Case Study in Protein Folding
Authors: Pedro Pablo González Pérez, Hiram I. Beltrán, Arturo Rojo-Domínguez, Máximo EduardoSánchez Gutiérrez
Abstract:
Multi-agent system approach has proven to be an effective and appropriate abstraction level to construct whole models of a diversity of biological problems, integrating aspects which can be found both in "micro" and "macro" approaches when modeling this type of phenomena. Taking into account these considerations, this paper presents the important computational characteristics to be gathered into a novel bioinformatics framework built upon a multiagent architecture. The version of the tool presented herein allows studying and exploring complex problems belonging principally to structural biology, such as protein folding. The bioinformatics framework is used as a virtual laboratory to explore a minimalist model of protein folding as a test case. In order to show the laboratory concept of the platform as well as its flexibility and adaptability, we studied the folding of two particular sequences, one of 45-mer and another of 64-mer, both described by an HP model (only hydrophobic and polar residues) and coarse grained 2D-square lattice. According to the discussion section of this piece of work, these two sequences were chosen as breaking points towards the platform, in order to determine the tools to be created or improved in such a way to overcome the needs of a particular computation and analysis of a given tough sequence. The backwards philosophy herein is that the continuous studying of sequences provides itself important points to be added into the platform, to any time improve its efficiency, as is demonstrated herein.
Keywords: multi-agent systems, blackboard-based agent architecture, bioinformatics framework, virtual laboratory, protein folding.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22111557 Input Data Balancing in a Neural Network PM-10 Forecasting System
Authors: Suk-Hyun Yu, Heeyong Kwon
Abstract:
Recently PM-10 has become a social and global issue. It is one of major air pollutants which affect human health. Therefore, it needs to be forecasted rapidly and precisely. However, PM-10 comes from various emission sources, and its level of concentration is largely dependent on meteorological and geographical factors of local and global region, so the forecasting of PM-10 concentration is very difficult. Neural network model can be used in the case. But, there are few cases of high concentration PM-10. It makes the learning of the neural network model difficult. In this paper, we suggest a simple input balancing method when the data distribution is uneven. It is based on the probability of appearance of the data. Experimental results show that the input balancing makes the neural networks’ learning easy and improves the forecasting rates.
Keywords: AI, air quality prediction, neural networks, pattern recognition, PM-10.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8291556 Performance Analysis of IDMA Scheme Using Quasi-Cyclic Low Density Parity Check Codes
Authors: Anurag Saxena, Alkesh Agrawal, Dinesh Kumar
Abstract:
The next generation mobile communication systems i.e. fourth generation (4G) was developed to accommodate the quality of service and required data rate. This project focuses on multiple access technique proposed in 4G communication systems. It is attempted to demonstrate the IDMA (Interleave Division Multiple Access) technology. The basic principle of IDMA is that interleaver is different for each user whereas CDMA employs different signatures. IDMA inherits many advantages of CDMA such as robust against fading, easy cell planning; dynamic channel sharing and IDMA increase the spectral efficiency and reduce the receiver complexity. In this, performance of IDMA is analyzed using QC-LDPC coding scheme further it is compared with LDPC coding and at last BER is calculated and plotted in MATLAB.
Keywords: 4G, QC-LDPC, CDMA, IDMA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1996