Search results for: first order plusdead time process
10229 Robust Image Transmission Over Time-varying Channels using Hierarchical Joint Source Channel Coding
Authors: Hatem. Elmeddeb, Noureddine, Hamdi, Ammar. Bouallègue
Abstract:
In this paper, a joint source-channel coding (JSCC) scheme for time-varying channels is presented. The proposed scheme uses hierarchical framework for both source encoder and transmission via QAM modulation. Hierarchical joint source channel codes with hierarchical QAM constellations are designed to track the channel variations which yields to a higher throughput by adapting certain parameters of the receiver to the channel variation. We consider the problem of still image transmission over time-varying channels with channel state information (CSI) available at 1) receiver only and 2) both transmitter and receiver being informed about the state of the channel. We describe an algorithm that optimizes hierarchical source codebooks by minimizing the distortion due to source quantizer and channel impairments. Simulation results, based on image representation, show that, the proposed hierarchical system outperforms the conventional schemes based on a single-modulator and channel optimized source coding.Keywords: Channel-optimized VQ (COVQ), joint optimization, QAM, hierarchical systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 142510228 Examining Effects of Electronic Market Functions on Decrease in Product Unit Cost and Response Time to Customer
Authors: Maziyar Nouraee
Abstract:
Electronic markets in recent decades contribute remarkably in business transactions. Many organizations consider traditional ways of trade non-economical and therefore they do trade only through electronic markets. There are different categorizations of electronic markets functions. In one classification, functions of electronic markets are categorized into classes as information, transactions, and value added. In the present paper, effects of the three classes on the two major elements of the supply chain management are measured. The two elements are decrease in the product unit cost and reduction in response time to the customer. The results of the current research show that among nine minor elements related to the three classes of electronic markets functions, six factors and three factors influence on reduction of the product unit cost and reduction of response time to the customer, respectively.
Keywords: Electronic Commerce, Electronic Market, B2B Trade, Supply Chain Management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 200910227 Enhancing Performance of Bluetooth Piconets Using Priority Scheduling and Exponential Back-Off Mechanism
Authors: Dharmendra Chourishi “Maitraya”, Sridevi Seshadri
Abstract:
Bluetooth is a personal wireless communication technology and is being applied in many scenarios. It is an emerging standard for short range, low cost, low power wireless access technology. Current existing MAC (Medium Access Control) scheduling schemes only provide best-effort service for all masterslave connections. It is very challenging to provide QoS (Quality of Service) support for different connections due to the feature of Master Driven TDD (Time Division Duplex). However, there is no solution available to support both delay and bandwidth guarantees required by real time applications. This paper addresses the issue of how to enhance QoS support in a Bluetooth piconet. The Bluetooth specification proposes a Round Robin scheduler as possible solution for scheduling the transmissions in a Bluetooth Piconet. We propose an algorithm which will reduce the bandwidth waste and enhance the efficiency of network. We define token counters to estimate traffic of real-time slaves. To increase bandwidth utilization, a back-off mechanism is then presented for best-effort slaves to decrease the frequency of polling idle slaves. Simulation results demonstrate that our scheme achieves better performance over the Round Robin scheduling.Keywords: Piconet, Medium Access Control, Polling algorithm, Scheduling, QoS, Time Division Duplex (TDD).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 170010226 Multidimensional and Data Mining Analysis for Property Investment Risk Analysis
Authors: Nur Atiqah Rochin Demong, Jie Lu, Farookh Khadeer Hussain
Abstract:
Property investment in the real estate industry has a high risk due to the uncertainty factors that will affect the decisions made and high cost. Analytic hierarchy process has existed for some time in which referred to an expert-s opinion to measure the uncertainty of the risk factors for the risk analysis. Therefore, different level of experts- experiences will create different opinion and lead to the conflict among the experts in the field. The objective of this paper is to propose a new technique to measure the uncertainty of the risk factors based on multidimensional data model and data mining techniques as deterministic approach. The propose technique consist of a basic framework which includes four modules: user, technology, end-user access tools and applications. The property investment risk analysis defines as a micro level analysis as the features of the property will be considered in the analysis in this paper.Keywords: Uncertainty factors, data mining, multidimensional data model, risk analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 292210225 Lyapunov Type Inequalities for Fractional Impulsive Hamiltonian Systems
Authors: Kazem Ghanbari, Yousef Gholami
Abstract:
This paper deals with study about fractional order impulsive Hamiltonian systems and fractional impulsive Sturm-Liouville type problems derived from these systems. The main purpose of this paper devotes to obtain so called Lyapunov type inequalities for mentioned problems. Also, in view point on applicability of obtained inequalities, some qualitative properties such as stability, disconjugacy, nonexistence and oscillatory behaviour of fractional Hamiltonian systems and fractional Sturm-Liouville type problems under impulsive conditions will be derived. At the end, we want to point out that for studying fractional order Hamiltonian systems, we will apply recently introduced fractional Conformable operators.Keywords: Fractional derivatives and integrals, Hamiltonian system, Lyapunov type inequalities, stability, disconjugacy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 152610224 Optimising Business Rules in the Services Sector
Authors: Alan Dormer
Abstract:
Business rules are widely used within the services sector. They provide consistency and allow relatively unskilled staff to process complex transactions correctly. But there are many examples where the rules themselves have an impact on the costs and profits of an organisation. Financial services, transport and human services are areas where the rules themselves can impact the bottom line in a predictable way. If this is the case, how can we find that set of rules that maximise profit, performance or customer service, or any other key performance indicators? The manufacturing, energy and process industries have embraced mathematical optimisation techniques to improve efficiency, increase production and so on. This paper explores several real world (but simplified) problems in the services sector and shows how business rules can be optimised. It also examines the similarities and differences between the service and other sectors, and how optimisation techniques could be used to deliver similar benefits.Keywords: Business rules, services, optimisation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 162110223 Probabilistic Characteristics of older PR Frames in the Mid-America Earthquake Region
Authors: Do-Hwan Kim, Roberto Leon
Abstract:
Probabilistic characteristics of seismic responses of the Partially Restrained connection rotation (PRCR) and panel zone deformation (PZD) installed in older steel moment frames were investigated in accordance with statistical inference in decision-making process. The 4, 6 and 8 story older steel moment frames with clip angle and T-stub connections were designed and analyzed using 2%/50yrs ground motions in four cities of the Mid-America earthquake region. The probability density function and cumulative distribution function of PRCR and PZD were determined by the goodness-of-fit tests based on probabilistic parameters measured from the results of the nonlinear time-history analyses. The obtained probabilistic parameters and distributions can be used to find out what performance level mainly PR connections and panel zones satisfy and how many PR connections and panel zones experience a serious damage under the Mid-America ground motions.Keywords: Mid-America earthquake, Panel zone, PR connection, Probabilistic characteristics, seismic performance
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 141210222 Towards Better Understanding of the Concept of Tacit Knowledge – A Cognitive Approach
Authors: Ilkka J. Virtanen
Abstract:
Tacit knowledge has been one of the most discussed and contradictory concepts in the field of knowledge management since the mid 1990s. The concept is used relatively vaguely to refer to any type of information that is difficult to articulate, which has led to discussions about the original meaning of the concept (adopted from Polanyi-s philosophy) and the nature of tacit knowing. It is proposed that the subject should be approached from the perspective of cognitive science in order to connect tacit knowledge to empirically studied cognitive phenomena. Some of the most important examples of tacit knowing presented by Polanyi are analyzed in order to trace the cognitive mechanisms of tacit knowing and to promote better understanding of the nature of tacit knowledge. The cognitive approach to Polanyi-s theory reveals that the tacit/explicit typology of knowledge often presented in the knowledge management literature is not only artificial but totally opposite approach compared to Polanyi-s thinking.Keywords: Cognitive science, explicit knowledge, knowledgemanagement, tacit knowledge.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 245710221 Framework for Delivery Reliability in European Machinery and Equipment Industry
Authors: G. Schuh, A. Kampker, A. Hoeschen, T. Jasinski
Abstract:
Today-s manufacturing companies are facing multiple and dynamic customer-supplier-relationships embedded in nonhierarchical production networks. This complex environment leads to problems with delivery reliability and wasteful turbulences throughout the entire network. This paper describes an operational model based on a theoretical framework which improves delivery reliability of each individual customer-supplier-relationship within non-hierarchical production networks of the European machinery and equipment industry. By developing a non-centralized coordination mechanism based on determining the value of delivery reliability and derivation of an incentive system for suppliers the number of in time deliveries can be increased and thus the turbulences in the production network smoothened. Comparable to an electronic stock exchange the coordination mechanism will transform the manual and nontransparent process of determining penalties for delivery delays into an automated and transparent market mechanism creating delivery reliability.
Keywords: delivery reliability, machinery and equipmentindustry, non-hierarchical production networks, supply chainmanagement
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 155910220 Tools and Techniques in Risk Assessment in Public Risk Management Organisations
Authors: Atousa Khodadadyan, Gabe Mythen, Hirbod Assa, Beverley Bishop
Abstract:
Risk assessment and the knowledge provided through this process is a crucial part of any decision-making process in the management of risks and uncertainties. Failure in assessment of risks can cause inadequacy in the entire process of risk management, which in turn can lead to failure in achieving organisational objectives as well as having significant damaging consequences on populations affected by the potential risks being assessed. The choice of tools and techniques in risk assessment can influence the degree and scope of decision-making and subsequently the risk response strategy. There are various available qualitative and quantitative tools and techniques that are deployed within the broad process of risk assessment. The sheer diversity of tools and techniques available to practitioners makes it difficult for organisations to consistently employ the most appropriate methods. This tools and techniques adaptation is rendered more difficult in public risk regulation organisations due to the sensitive and complex nature of their activities. This is particularly the case in areas relating to the environment, food, and human health and safety, when organisational goals are tied up with societal, political and individuals’ goals at national and international levels. Hence, recognising, analysing and evaluating different decision support tools and techniques employed in assessing risks in public risk management organisations was considered. This research is part of a mixed method study which aimed to examine the perception of risk assessment and the extent to which organisations practise risk assessment’ tools and techniques. The study adopted a semi-structured questionnaire with qualitative and quantitative data analysis to include a range of public risk regulation organisations from the UK, Germany, France, Belgium and the Netherlands. The results indicated the public risk management organisations mainly use diverse tools and techniques in the risk assessment process. The primary hazard analysis; brainstorming; hazard analysis and critical control points were described as the most practiced risk identification techniques. Within qualitative and quantitative risk analysis, the participants named the expert judgement, risk probability and impact assessment, sensitivity analysis and data gathering and representation as the most practised techniques.
Keywords: Decision-making, public risk management organisations, risk assessment, tools and techniques.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 164510219 Identification of Differentially Expressed Gene(DEG) in Atherosclerotic Lesion by Annealing Control Primer (ACP)-Based Genefishing™ PCR
Authors: M. Maimunah, G. A. Froemming, H. Nawawi, M. I. Nafeeza, O. Effat, M. Y. Rosmadi, M. S. Mohamed Saifulaman
Abstract:
Atherosclerosis was identified as a chronic inflammatory process resulting from interactions between plasma lipoproteins, cellular components (monocyte, macrophages, T lymphocytes, endothelial cells and smooth muscle cells) and the extracellular matrix of the arterial wall. Several types of genes were known to express during formation of atherosclerosis. This study is carried out to identify unknown differentially expressed gene (DEG) in atherogenesis. Rabbit’s aorta tissues were stained by H&E for histomorphology. GeneFishing™ PCR analysis was performed from total RNA extracted from the aorta tissues. The DNA fragment from DEG was cloned, sequenced and validated by Real-time PCR. Histomorphology showed intimal thickening in the aorta. DEG detected from ACP-41 was identified as cathepsin B gene and showed upregulation at week-8 and week-12 of atherogenesis. Therefore, ACP-based GeneFishing™ PCR facilitated identification of cathepsin B gene which was differentially expressed during development of atherosclerosis.
Keywords: Atherosclerosis, GeneFishing™ PCR, cathepsin B gene.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 195610218 Design and Fabrication of Micro-Bubble Oxygenator
Authors: Chiang-Ho Cheng, An-Shik Yang, Hong-Yih Cheng
Abstract:
This paper applies the MEMS technology to design and fabricate a micro-bubble generator by a piezoelectric actuator. Coupled with a nickel nozzle plate, an annular piezoelectric ceramic was utilized as the primary structure of the generator. In operations, the piezoelectric element deforms transversely under an electric field applied across the thickness of the generator. The surface of the nozzle plate can expand or contract because of the induction of radial strain, resulting in the whole structure to bend, and successively transport oxygen micro-bubbles into the blood flow for enhancing the oxygen content in blood. In the tests, a high magnification microscope and a high speed CCD camera were employed to photograph the time evolution of meniscus shape of gaseous bubbles dispensed from the micro-bubble generator for flow visualization. This investigation thus explored the bubble formation process including the influences of inlet gas pressure along with driving voltage and resonance frequency on the formed bubble extent.
Keywords: Micro-bubble, nozzle, oxygenator, piezoelectric.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 135410217 The Reach of Shopping Center Layout Form on U Subway - Based On Kernel Density Estimate
Authors: Wen Liu
Abstract:
With the rapid progress of modern cities, the railway construction must be developing quickly in China.As a typical high-density country, shopping center on the subway should be one important factor during the process of urban development. The paper discusses the influence of the layout of shopping center on the subway, and put it in the time and space’s axis of Shanghai urban development. We usethe digital technology to establish the database of relevant information. And then get the change role about shopping center on subway in Shanghaiby the Kernel density estimate.The result shows the development of shopping center on subway has a relationship with local economic strength, population size, policysupport, and city construction. And the suburbanization trend of shopping center would be increasingly significant.By this case research, we could see the Kernel density estimate is an efficient analysis method on the spatial layout. It could reveal the characters of layout form of shopping center on subway in essence. And it can also be applied to the other research of space form.
Keywords: Shanghai, Shopping center on the subway, Layout form, The Kernel density estimate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 178710216 Enhanced Shell Sorting Algorithm
Authors: Basit Shahzad, Muhammad Tanvir Afzal
Abstract:
Many algorithms are available for sorting the unordered elements. Most important of them are Bubble sort, Heap sort, Insertion sort and Shell sort. These algorithms have their own pros and cons. Shell Sort which is an enhanced version of insertion sort, reduces the number of swaps of the elements being sorted to minimize the complexity and time as compared to insertion sort. Shell sort improves the efficiency of insertion sort by quickly shifting values to their destination. Average sort time is O(n1.25), while worst-case time is O(n1.5). It performs certain iterations. In each iteration it swaps some elements of the array in such a way that in last iteration when the value of h is one, the number of swaps will be reduced. Donald L. Shell invented a formula to calculate the value of ?h?. this work focuses to identify some improvement in the conventional Shell sort algorithm. ''Enhanced Shell Sort algorithm'' is an improvement in the algorithm to calculate the value of 'h'. It has been observed that by applying this algorithm, number of swaps can be reduced up to 60 percent as compared to the existing algorithm. In some other cases this enhancement was found faster than the existing algorithms available.Keywords: Algorithm, Computation, Shell, Sorting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 313610215 Determination of Water Pollution and Water Quality with Decision Trees
Authors: Çiğdem Bakır, Mecit Yüzkat
Abstract:
With the increasing emphasis on water quality worldwide, the search for and expanding the market for new and intelligent monitoring systems has increased. The current method is the laboratory process, where samples are taken from bodies of water, and tests are carried out in laboratories. This method is time-consuming, a waste of manpower and uneconomical. To solve this problem, we used machine learning methods to detect water pollution in our study. We created decision trees with the Orange3 software used in the study and tried to determine all the factors that cause water pollution. An automatic prediction model based on water quality was developed by taking many model inputs such as water temperature, pH, transparency, conductivity, dissolved oxygen, and ammonia nitrogen with machine learning methods. The proposed approach consists of three stages: Preprocessing of the data used, feature detection and classification. We tried to determine the success of our study with different accuracy metrics and the results were presented comparatively. In addition, we achieved approximately 98% success with the decision tree.
Keywords: Decision tree, water quality, water pollution, machine learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26010214 Predictive Semi-Empirical NOx Model for Diesel Engine
Authors: Saurabh Sharma, Yong Sun, Bruce Vernham
Abstract:
Accurate prediction of NOx emission is a continuous challenge in the field of diesel engine-out emission modeling. Performing experiments for each conditions and scenario cost significant amount of money and man hours, therefore model-based development strategy has been implemented in order to solve that issue. NOx formation is highly dependent on the burn gas temperature and the O2 concentration inside the cylinder. The current empirical models are developed by calibrating the parameters representing the engine operating conditions with respect to the measured NOx. This makes the prediction of purely empirical models limited to the region where it has been calibrated. An alternative solution to that is presented in this paper, which focus on the utilization of in-cylinder combustion parameters to form a predictive semi-empirical NOx model. The result of this work is shown by developing a fast and predictive NOx model by using the physical parameters and empirical correlation. The model is developed based on the steady state data collected at entire operating region of the engine and the predictive combustion model, which is developed in Gamma Technology (GT)-Power by using Direct Injected (DI)-Pulse combustion object. In this approach, temperature in both burned and unburnt zone is considered during the combustion period i.e. from Intake Valve Closing (IVC) to Exhaust Valve Opening (EVO). Also, the oxygen concentration consumed in burnt zone and trapped fuel mass is also considered while developing the reported model. Several statistical methods are used to construct the model, including individual machine learning methods and ensemble machine learning methods. A detailed validation of the model on multiple diesel engines is reported in this work. Substantial numbers of cases are tested for different engine configurations over a large span of speed and load points. Different sweeps of operating conditions such as Exhaust Gas Recirculation (EGR), injection timing and Variable Valve Timing (VVT) are also considered for the validation. Model shows a very good predictability and robustness at both sea level and altitude condition with different ambient conditions. The various advantages such as high accuracy and robustness at different operating conditions, low computational time and lower number of data points requires for the calibration establishes the platform where the model-based approach can be used for the engine calibration and development process. Moreover, the focus of this work is towards establishing a framework for the future model development for other various targets such as soot, Combustion Noise Level (CNL), NO2/NOx ratio etc.
Keywords: Diesel engine, machine learning, NOx emission, semi-empirical.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 85510213 Design and Fabrication of an Electrostatically Actuated Parallel-Plate Mirror by 3D-Printer
Authors: J. Mizuno, S. Takahashi
Abstract:
In this paper, design and fabrication of an actuated parallel-plate mirror based on a 3D-printer is described. The mirror and electrode layers are fabricated separately and assembled thereafter. The alignment is performed by dowel pin-hole pairs fabricated on the respective layers. The electrodes are formed on the surface of the electrode layer by Au ion sputtering using a suitable mask, which is also fabricated by a 3D-printer.For grounding the mirror layer, except the contact area with the electrode paths, all the surface is Au ion sputtered. 3D-printers are widely used for creating 3D models or mock-ups. The authors have recently proposed that these models can perform electromechanical functions such as actuators by suitably masking them followed by metallization process. Since the smallest possible fabrication size is in the order of sub-millimeters, these electromechanical devices are named by the authors as SMEMS (Sub-Milli Electro-Mechanical Systems) devices. The proposed mirror described in this paper which consists of parallel-plate electrostatic actuators is also one type of SMEMS devices. In addition, SMEMS is totally environment-clean compared to MEMS (Micro Electro-Mechanical Systems) fabrication processes because any hazardous chemicals or gases are utilized.
Keywords: MEMS, parallel-plate mirror, SMEMS, 3D-printer.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 181010212 Locating Center Points for Radial Basis Function Networks Using Instance Reduction Techniques
Authors: Rana Yousef, Khalil el Hindi
Abstract:
The behavior of Radial Basis Function (RBF) Networks greatly depends on how the center points of the basis functions are selected. In this work we investigate the use of instance reduction techniques, originally developed to reduce the storage requirements of instance based learners, for this purpose. Five Instance-Based Reduction Techniques were used to determine the set of center points, and RBF networks were trained using these sets of centers. The performance of the RBF networks is studied in terms of classification accuracy and training time. The results obtained were compared with two Radial Basis Function Networks: RBF networks that use all instances of the training set as center points (RBF-ALL) and Probabilistic Neural Networks (PNN). The former achieves high classification accuracies and the latter requires smaller training time. Results showed that RBF networks trained using sets of centers located by noise-filtering techniques (ALLKNN and ENN) rather than pure reduction techniques produce the best results in terms of classification accuracy. The results show that these networks require smaller training time than that of RBF-ALL and higher classification accuracy than that of PNN. Thus, using ALLKNN and ENN to select center points gives better combination of classification accuracy and training time. Our experiments also show that using the reduced sets to train the networks is beneficial especially in the presence of noise in the original training sets.
Keywords: Radial basis function networks, Instance-based reduction, PNN.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 168710211 Development of a Simulator for Explaining Organic Chemical Reactions Based on Qualitative Process Theory
Authors: Alicia Y. C. Tang, Rukaini Hj. Abdullah, Sharifuddin M. Zain
Abstract:
This paper discusses the development of a qualitative simulator (abbreviated QRiOM) for predicting the behaviour of organic chemical reactions. The simulation technique is based on the qualitative process theory (QPT) ontology. The modelling constructs of QPT embody notions of causality which can be used to explain the behaviour of a chemical system. The major theme of this work is that, in a qualitative simulation environment, students are able to articulate his/her knowledge through the inspection of explanations generated by software. The implementation languages are Java and Prolog. The software produces explanation in various forms that stresses on the causal theories in the chemical system which can be effectively used to support learning.Keywords: Chemical reactions, explanation, qualitative processtheory, simulation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 156610210 On the Optimality Assessment of Nanoparticle Size Spectrometry and Its Association to the Entropy Concept
Authors: A. Shaygani, R. Saifi, M. S. Saidi, M. Sani
Abstract:
Particle size distribution, the most important characteristics of aerosols, is obtained through electrical characterization techniques. The dynamics of charged nanoparticles under the influence of electric field in Electrical Mobility Spectrometer (EMS) reveals the size distribution of these particles. The accuracy of this measurement is influenced by flow conditions, geometry, electric field and particle charging process, therefore by the transfer function (transfer matrix) of the instrument. In this work, a wire-cylinder corona charger was designed and the combined fielddiffusion charging process of injected poly-disperse aerosol particles was numerically simulated as a prerequisite for the study of a multichannel EMS. The result, a cloud of particles with no uniform charge distribution, was introduced to the EMS. The flow pattern and electric field in the EMS were simulated using Computational Fluid Dynamics (CFD) to obtain particle trajectories in the device and therefore to calculate the reported signal by each electrometer. According to the output signals (resulted from bombardment of particles and transferring their charges as currents), we proposed a modification to the size of detecting rings (which are connected to electrometers) in order to evaluate particle size distributions more accurately. Based on the capability of the system to transfer information contents about size distribution of the injected particles, we proposed a benchmark for the assessment of optimality of the design. This method applies the concept of Von Neumann entropy and borrows the definition of entropy from information theory (Shannon entropy) to measure optimality. Entropy, according to the Shannon entropy, is the ''average amount of information contained in an event, sample or character extracted from a data stream''. Evaluating the responses (signals) which were obtained via various configurations of detecting rings, the best configuration which gave the best predictions about the size distributions of injected particles, was the modified configuration. It was also the one that had the maximum amount of entropy. A reasonable consistency was also observed between the accuracy of the predictions and the entropy content of each configuration. In this method, entropy is extracted from the transfer matrix of the instrument for each configuration. Ultimately, various clouds of particles were introduced to the simulations and predicted size distributions were compared to the exact size distributions.Keywords: Aerosol Nano-Particle, CFD, Electrical Mobility Spectrometer, Von Neumann entropy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 185910209 Preparation and Characterization of Newly Developed Trabecular Structures in Titanium Alloy to Optimize Osteointegration
Authors: M. Regis, E. Marin, S. Fusi, M. Pressacco, L. Fedrizzi
Abstract:
Electron Beam Melting (EBM) process was used to prepare porous scaffolds with controlled porosity to ensure optimal levels of osteointegration for different trabeculae sizes. Morphological characterization by means of SEM analyses was carried out to assess pore dimensions; tensile, compression and adhesion tests have been carried out to determine the mechanical behavior. The results indicate that EBM process allows the creation of regular and repeatable porous scaffolds. Mechanical properties greatly depend on pore dimension and on bulk-pore ratio. Adhesion resistance meets the normative requirements, and the overall performance of the produced structures is compatible with potential orthopaedic applications.
Keywords: Additive manufacturing, orthopaedic implants, osteointegration, trabecular structures
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 231610208 Tracing Syrian Refugees Urban Mobilities: The Case of Egypt and Canada
Authors: N. Elgendy, N. Hussein
Abstract:
The current Syrian crisis has caused unprecedented practices of global mobility. The process of forced eviction and the resettlement of refugees could be seen through the insights of the “new mobilities paradigm”. The mobility of refugees in terms of meaning and practice is a subject that calls for further studies. There is a need for the development of an approach to human mobility to understand a practice that is turning into a phenomenon in the 21st century. This paper aims at studying, from a qualitative point of view, the process of movement within the six constituents of mobility defined as the first phase of the journey of a refugee. The second phase would include the process of settling in and re-defining the host country as new “home” to refugees. The change in the refugee state of mind and crossing the physical and mental borders from a “foreigner” to a citizen is encouraged by both the governmental policies and the local communities’ efforts to embrace these newcomers. The paper would focus on these policies of social and economic integration. The concept of integration connotes the idea that refugees would enjoy the opportunities, rights and services available to the citizens of the refugee’s new community. So, this paper examines this concept through showcasing the two hosting countries of Canada and Egypt, as they provide two contrasting situations in terms of cultural, geographical, economic and political backgrounds. The analysis would highlight the specific policies defined towards the refugees including the mass communication, media calls, and access to employment. This research is part of a qualitative research project on the process of Urban Mobility practiced by the Syrian Refugees, drawing on conversational interviews with new-settlers who have moved to the different hosting countries, from their home in Syria. It explores these immigrants’ practical and emotional relationships with the process of movement and settlement. It uses the conversational interviews as a tool to document analysis and draw relationships in an attempt to establish an understanding of the factors that contribute to the new-settlers feeling of home and integration within the new community.Keywords: Mobility, refugees, home, integration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 130110207 Improved Processing Speed for Text Watermarking Algorithm in Color Images
Authors: Hamza A. Al-Sewadi, Akram N. A. Aldakari
Abstract:
Copyright protection and ownership proof of digital multimedia are achieved nowadays by digital watermarking techniques. A text watermarking algorithm for protecting the property rights and ownership judgment of color images is proposed in this paper. Embedding is achieved by inserting texts elements randomly into the color image as noise. The YIQ image processing model is found to be faster than other image processing methods, and hence, it is adopted for the embedding process. An optional choice of encrypting the text watermark before embedding is also suggested (in case required by some applications), where, the text can is encrypted using any enciphering technique adding more difficulty to hackers. Experiments resulted in embedding speed improvement of more than double the speed of other considered systems (such as least significant bit method, and separate color code methods), and a fairly acceptable level of peak signal to noise ratio (PSNR) with low mean square error values for watermarking purposes.
Keywords: Steganography, watermarking, private keys, time complexity measurements.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 81610206 Prediction of the Thermal Parameters of a High-Temperature Metallurgical Reactor Using Inverse Heat Transfer
Authors: Mohamed Hafid, Marcel Lacroix
Abstract:
This study presents an inverse analysis for predicting the thermal conductivities and the heat flux of a high-temperature metallurgical reactor simultaneously. Once these thermal parameters are predicted, the time-varying thickness of the protective phase-change bank that covers the inside surface of the brick walls of a metallurgical reactor can be calculated. The enthalpy method is used to solve the melting/solidification process of the protective bank. The inverse model rests on the Levenberg-Marquardt Method (LMM) combined with the Broyden method (BM). A statistical analysis for the thermal parameter estimation is carried out. The effect of the position of the temperature sensors, total number of measurements and measurement noise on the accuracy of inverse predictions is investigated. Recommendations are made concerning the location of temperature sensors.
Keywords: Inverse heat transfer, phase change, metallurgical reactor, Levenberg–Marquardt method, Broyden method, bank thickness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 169310205 Modeling of a Vehicle Wheel System Having a Built-in Suspension Structure Consisted of Radially Deployed Colloidal Spokes between Hub and Rim
Authors: Barenten Suciu
Abstract:
In this work, by replacing the traditional solid spokes with colloidal spokes, a vehicle wheel with a built-in suspension structure is proposed. Following the background and description of the wheel system, firstly, a vibration model of the wheel equipped with colloidal spokes is proposed, and based on such model the equivalent damping coefficients and spring constants are identified. Then, a modified model of a quarter-vehicle moving on a rough pavement is proposed in order to estimate the transmissibility of vibration from the road roughness to vehicle body. In the end, the optimal design of the colloidal spokes and the optimum number of colloidal spokes are decided in order to minimize the transmissibility of vibration, i.e., to maximize the ride comfort of the vehicle.
Keywords: Built-in suspension, colloidal spoke, intrinsic spring, vibration analysis, wheel.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 193610204 Promoting Gender Equality within Islamic Tradition via Contextualist Approach
Authors: Ali Akbar
Abstract:
The importance of advancing women’s rights is closely intertwined with the development of civil society and the institutionalization of democracy in Middle Eastern countries. There is indeed an intimate relationship between the process of democratization and promoting gender equality, since democracy necessitates equality between men and women. In order to advance the issue of gender equality, what is required is a solid theoretical framework which has its roots in the reexamination of pre-modern interpretation of certain Qurʾānic passages that seem to have given men more rights than it gives women. This paper suggests that those Muslim scholars who adopt a contextualist approach to the Qurʾānic text and its interpretation provide a solid theoretical background for improving women’s rights. Indeed, the aim of the paper is to discuss how the contextualist approach to the Qurʾānic text and its interpretation given by a number of prominent scholars is capable of promoting the issue of gender equality. The paper concludes that since (1) much of the gender inequality found in the primary sources of Islam as well as pre-modern Muslim writings is rooted in the natural cultural norms and standards of early Islamic societies and (2) since the context of today’s world is so different from that of the pre-modern era, the proposed models provide a solid theoretical framework for promoting women’s rights and gender equality.Keywords: Contextualism, Gender equality, Islam, Women’s rights.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 176010203 A Comparative Study on Fuzzy and Neuro-Fuzzy Enabled Cluster Based Routing Protocols for Wireless Sensor Networks
Authors: Y. Harold Robinson, E. Golden Julie
Abstract:
Dynamic Routing in Wireless Sensor Networks (WSNs) has played a significant task in research for the recent years. Energy consumption and data delivery in time are the major parameters with the usage of sensor nodes that are significant criteria for these networks. The location of sensor nodes must not be prearranged. Clustering in WSN is a key methodology which is used to enlarge the life-time of a sensor network. It consists of numerous real-time applications. The features of WSNs are minimized the consumption of energy. Soft computing techniques can be included to accomplish improved performance. This paper surveys the modern trends in routing enclose fuzzy logic and Neuro-fuzzy logic based on the clustering techniques and implements a comparative study of the numerous related methodologies.Keywords: Wireless sensor networks, clustering, fuzzy logic, neuro-fuzzy logic, energy efficiency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 98910202 Modeling and Investigation of Elongation in Free Explosive Forming of Aluminum Alloy Plate
Authors: R. Alipour, F.Najarian
Abstract:
Because of high ductility, aluminum alloys, have been widely used as an important base of metal forming industries. But the main week point of these alloys is their low strength so in forming them with conventional methods like deep drawing, hydro forming, etc have been always faced with problems like fracture during of forming process. Because of this, recently using of explosive forming method for forming of these plates has been recommended. In this paper free explosive forming of A2024 aluminum alloy is numerically simulated and during it, explosion wave propagation process is studied. Consequences of this simulation can be effective in prediction of quality of production. These consequences are compared with an experimental test and show the superiority of this method to similar methods like hydro forming and deep drawing.
Keywords: Free explosive forming, CEL, Johnson cook.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 230810201 Performance Analysis Model Development for Mae Moh Coal-Fired Power Plant
Authors: Thitiporn Supasri, Natanee Vorayos, Piriya Thongchiew
Abstract:
Electrification is a complex process and governed by various parameters. Modeling of power plant’s target efficiency or target heat rate is often formulated and compared with the actual values. This comparison not only implies the performance of the power plant but also reflects the energy losses possibly inherited in some of related equipment and processes. The current modeling of Coal-fired Mae Moh power plant was formulated at the first commissioning. Some of equipments were replaced due to its life time. Relatively outdated for 20 years, the utilization of the model is not accomplished. This work has focused on the development of the performance analysis model of aforementioned power plant according to the most updated and current working conditions. The model is more appropriate and shows accuracy in its analysis. Losses are detected and measures are introduced such that reduction in energy consumption, related cost, and also environment impacts can be anticipated.
Keywords: Performance analysis model, Power plant modeling, Target heat rate, Target efficiency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 233010200 A Modified Speech Enhancement Using Adaptive Gain Equalizer with Non linear Spectral Subtraction for Robust Speech Recognition
Authors: C. Ganesh Babu, P. T. Vanathi
Abstract:
In this paper we present an enhanced noise reduction method for robust speech recognition using Adaptive Gain Equalizer with Non linear Spectral Subtraction. In Adaptive Gain Equalizer method (AGE), the input signal is divided into a number of subbands that are individually weighed in time domain, in accordance to the short time Signal-to-Noise Ratio (SNR) in each subband estimation at every time instant. Instead of focusing on suppression the noise on speech enhancement is focused. When analysis was done under various noise conditions for speech recognition, it was found that Adaptive Gain Equalizer method algorithm has an obvious failing point for a SNR of -5 dB, with inadequate levels of noise suppression for SNR less than this point. This work proposes the implementation of AGE when coupled with Non linear Spectral Subtraction (AGE-NSS) for robust speech recognition. The experimental result shows that out AGE-NSS performs the AGE when SNR drops below -5db level.
Keywords: Adaptive Gain Equalizer, Non Linear Spectral Subtraction, Speech Enhancement, and Speech Recognition.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1703