Search results for: business process model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 11916

Search results for: business process model

5976 Effects of the Coagulation Bath and Reduction Process on SO2 Adsorption Capacity of Graphene Oxide Fiber

Authors: Özge Alptoğa, Nuray Uçar, Nilgün Karatepe Yavuz, Ayşen Önen

Abstract:

Sulfur dioxide (SO2) is a very toxic air pollutant gas and it causes the greenhouse effect, photochemical smog, and acid rain, which threaten human health severely. Thus, the capture of SO2 gas is very important for the environment. Graphene which is two-dimensional material has excellent mechanical, chemical, thermal properties, and many application areas such as energy storage devices, gas adsorption, sensing devices, and optical electronics. Further, graphene oxide (GO) is examined as a good adsorbent because of its important features such as functional groups (epoxy, carboxyl and hydroxyl) on the surface and layered structure. The SO2 adsorption properties of the fibers are usually investigated on carbon fibers. In this study, potential adsorption capacity of GO fibers was researched. GO dispersion was first obtained with Hummers’ method from graphite, and then GO fibers were obtained via wet spinning process. These fibers were converted into a disc shape, dried, and then subjected to SO2 gas adsorption test. The SO2 gas adsorption capacity of GO fiber discs was investigated in the fields of utilization of different coagulation baths and reduction by hydrazine hydrate. As coagulation baths, single and triple baths were used. In single bath, only ethanol and CaCl2 (calcium chloride) salt were added. In triple bath, each bath has a different concentration of water/ethanol and CaCl2 salt, and the disc obtained from triple bath has been called as reference disk. The fibers which were produced with single bath were flexible and rough, and the analyses show that they had higher SO2 adsorption capacity than triple bath fibers (reference disk). However, the reduction process did not increase the adsorption capacity, because the SEM images showed that the layers and uniform structure in the fiber form were damaged, and reduction decreased the functional groups which SO2 will be attached. Scanning Electron Microscopy (SEM), Fourier Transform Infrared Spectroscopy (FTIR), X-Ray Diffraction (XRD) analyzes were performed on the fibers and discs, and the effects on the results were interpreted. In the future applications of the study, it is aimed that subjects such as pH and additives will be examined.

Keywords: Coagulation bath, graphene oxide fiber, reduction, SO2 gas adsorption.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1154
5975 Social Dimension of Air Transport Sustainable Development

Authors: Dimitrios J. Dimitriou, Maria F. Sartzetaki

Abstract:

Air Transport links markets and individuals, making regions more competitive and promoting social and economic development. The assessment of social contribution is the key objective of this paper, focusing on the definition of the components of social dimension and welfare metrics in the national scale. According to a top-down approach, the key dimensions that affect the social welfare are presented. Conventional wisdom is to provide estimations on added value to social issues caused by the air transport development and present the methodology framework for measuring the contribution of transport development in social value chain. Greece is the case study of this paper, providing results from the contribution of air transport infrastructures in national welfare. The application key findings are essential for managers and decision makers to support actions and plans towards economic recovery of an economy presenting strong seasonal characteristics (because of tourism) and suffering from recession.

Keywords: Air transport, social dimension, social coherence, resilient business development, socioeconomic impact.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 931
5974 Opportunities and Options for Government to Promote Corporate Social Responsibility in the Czech Republic

Authors: Pavel Adámek

Abstract:

The concept of corporate social responsibility (CSR) in the Czech Republic has evolved notably during the last few years and an issue that started as an interest- and motive-based activity for businesses is becoming more commonplace. Governments have a role to play in ensuring that corporations behave according to the rules and norms of society and can legislate, foster, collaborate with businesses and endorse good practice in order to facilitate the development of CSR. The purpose of this paper is to examine the opportunities and options of CSR in government policy and research its relevance to a business sector. An increasing number of companies is engaging in responsible activities, the public awareness of CSR is rising, and customers are giving higher importance to CSR of companies in their choice. By drawing on existing CSR approach in Czech and understanding of CSR are demonstrated. The paper provides an overview, more detailed government approach of CSR.

Keywords: Approach, corporate social responsibility, government policy, instruments.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2209
5973 Transport of Analytes under Mixed Electroosmotic and Pressure Driven Flow of Power Law Fluid

Authors: Naren Bag, S. Bhattacharyya, Partha P. Gopmandal

Abstract:

In this study, we have analyzed the transport of analytes under a two dimensional steady incompressible flow of power-law fluids through rectangular nanochannel. A mathematical model based on the Cauchy momentum-Nernst-Planck-Poisson equations is considered to study the combined effect of mixed electroosmotic (EO) and pressure driven (PD) flow. The coupled governing equations are solved numerically by finite volume method. We have studied extensively the effect of key parameters, e.g., flow behavior index, concentration of the electrolyte, surface potential, imposed pressure gradient and imposed electric field strength on the net average flow across the channel. In addition to study the effect of mixed EOF and PD on the analyte distribution across the channel, we consider a nonlinear model based on general convective-diffusion-electromigration equation. We have also presented the retention factor for various values of electrolyte concentration and flow behavior index.

Keywords: Electric double layer, finite volume method, flow behavior index, mixed electroosmotic/pressure driven flow, Non-Newtonian power-law fluids, numerical simulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1186
5972 A Unified Approach for Naval Telecommunication Architectures

Authors: Y. Lacroix, J.-F. Malbranque

Abstract:

We present a chronological evolution for naval telecommunication networks. We distinguish periods: with or without multiplexers, with switch systems, with federative systems, with medium switching, and with medium switching with wireless networks. This highlights the introduction of new layers and technology in the architecture. These architectures are presented using layer models of transmission, in a unified way, which enables us to integrate pre-existing models. A ship of a naval fleet has internal communications (i.e. applications' networks of the edge) and external communications (i.e. the use of the means of transmission between edges). We propose architectures, deduced from the layer model, which are the point of convergence between the networks on board and the HF, UHF radio, and satellite resources. This modelling allows to consider end-to-end naval communications, and in a more global way, that is from the user on board towards the user on shore, including transmission and networks on the shore side. The new architectures need take care of quality of services for end-to-end communications, the more remote control develops a lot and will do so in the future. Naval telecommunications will be more and more complex and will use more and more advanced technologies, it will thus be necessary to establish clear global communication schemes to grant consistency of the architectures. Our latest model has been implemented in a military naval situation, and serves as the basic architecture for the RIFAN2 network.

Keywords: Equilibrium beach profile, eastern tombolo of Giens, potential function, erosion.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 848
5971 A New Approach to Face Recognition Using Dual Dimension Reduction

Authors: M. Almas Anjum, M. Younus Javed, A. Basit

Abstract:

In this paper a new approach to face recognition is presented that achieves double dimension reduction, making the system computationally efficient with better recognition results and out perform common DCT technique of face recognition. In pattern recognition techniques, discriminative information of image increases with increase in resolution to a certain extent, consequently face recognition results change with change in face image resolution and provide optimal results when arriving at a certain resolution level. In the proposed model of face recognition, initially image decimation algorithm is applied on face image for dimension reduction to a certain resolution level which provides best recognition results. Due to increased computational speed and feature extraction potential of Discrete Cosine Transform (DCT), it is applied on face image. A subset of coefficients of DCT from low to mid frequencies that represent the face adequately and provides best recognition results is retained. A tradeoff between decimation factor, number of DCT coefficients retained and recognition rate with minimum computation is obtained. Preprocessing of the image is carried out to increase its robustness against variations in poses and illumination level. This new model has been tested on different databases which include ORL , Yale and EME color database.

Keywords: Biometrics, DCT, Face Recognition, Illumination, Computation, Feature extraction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1668
5970 Using Blockchain Technology to Extend the Vendor Managed Inventory for Sustainability

Authors: Elham Ahmadi, Roshaali Khaturia, Pardis Sahraei, Mohammad Niyayesh, Omid Fatahi Valilai

Abstract:

Nowadays, Information Technology (IT) is changing the way traditional enterprise management concepts work. One of the most dominant IT achievements is the Blockchain Technology. This technology enables the distributed collaboration of stakeholders for their interactions while fulfilling the security and consensus rules among them. This paper has focused on the application of Blockchain technology to enhance one of traditional inventory management models. The Vendor Managed Inventory (VMI) has been considered one of the most efficient mechanisms for vendor inventory planning by the suppliers. While VMI has brought competitive advantages for many industries, however its centralized mechanism limits the collaboration of a pool of suppliers and vendors simultaneously. This paper has studied the recent research for VMI application in industries and also has investigated the applications of Blockchain technology for decentralized collaboration of stakeholders. Focusing on sustainability issue for total supply chain consisting suppliers and vendors, it has proposed a Blockchain based VMI conceptual model. The different capabilities of this model for enabling the collaboration of stakeholders while maintaining the competitive advantages and sustainability issues have been discussed.

Keywords: Vendor Managed Inventory, Blockchain Technology, supply chain planning, sustainability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 778
5969 Markov Game Controller Design Algorithms

Authors: Rajneesh Sharma, M. Gopal

Abstract:

Markov games are a generalization of Markov decision process to a multi-agent setting. Two-player zero-sum Markov game framework offers an effective platform for designing robust controllers. This paper presents two novel controller design algorithms that use ideas from game-theory literature to produce reliable controllers that are able to maintain performance in presence of noise and parameter variations. A more widely used approach for controller design is the H∞ optimal control, which suffers from high computational demand and at times, may be infeasible. Our approach generates an optimal control policy for the agent (controller) via a simple Linear Program enabling the controller to learn about the unknown environment. The controller is facing an unknown environment, and in our formulation this environment corresponds to the behavior rules of the noise modeled as the opponent. Proposed controller architectures attempt to improve controller reliability by a gradual mixing of algorithmic approaches drawn from the game theory literature and the Minimax-Q Markov game solution approach, in a reinforcement-learning framework. We test the proposed algorithms on a simulated Inverted Pendulum Swing-up task and compare its performance against standard Q learning.

Keywords: Reinforcement learning, Markov Decision Process, Matrix Games, Markov Games, Smooth Fictitious play, Controller, Inverted Pendulum.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1508
5968 An Investigation into Ozone Concentration at Urban and Rural Monitoring Stations in Malaysia

Authors: Negar Banan, Mohd Talib Latif

Abstract:

This study investigated the relationship between urban and rural ozone concentrations and quantified the extent to which ambient rural conditions and the concentrations of other pollutants can be used to predict urban ozone concentrations. The study describes the variations of ozone in weekday and weekends as well as the daily maximum recorded at selected monitoring stations. The results showed that Putrajaya station had the highest concentrations of O3 on weekend due the titration of NO during the weekday. Additionally, Jerantut had the lowest average concentration with a reading value high on Wednesdays. The comparisons of average and maximum concentrations of ozone for the three stations showed that the strongest significant correlation is recorded in Jerantut station with the value R2= 0.769. Ozone concentrations originating from a neighbouring urban site form a better predictor to the urban ozone concentrations than widespread rural ozone at some levels of temporal averaging. It is found that in urban and rural of Malaysian peninsular, the concentration of ozone depends on the concentration of NOx and seasonal meteorological factors. The HYSPLIT Model (the northeast monsoon) showed that the wind direction can also influence the concentration of ozone in the atmosphere in the studied areas.

Keywords: Ozone, Hysplit model, Weekend effect, Daily Average and Daily maximum, Malaysia

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2117
5967 Development of Cellulose Panels with Porous Structure for Sustainable Building Insulation

Authors: P. Garbagnoli, M. Musitelli, B. Del Curto, MP. Pedeferri

Abstract:

The study and development of an innovative material for building insulation is really important for a sustainable society in order to improve comfort and reducing energy consumption. The aim of this work is the development of insulating panels for sustainable buildings based on an innovative material made by cardboard and Phase Change Materials (PCMs). The research has consisted in laboratory tests whose purpose has been the obtaining of the required properties for insulation panels: lightweight, porous structures and mechanical resistance. PCMs have been used for many years in the building industry as smart insulation technology because of their properties of storage and release high quantity of latent heat at useful specific temperatures [1]- [2]. The integration of PCMs into cellulose matrix during the waste paper recycling process has been developed in order to obtain a composite material. Experiments on the productive process for the realization of insulating panels were done in order to make the new material suitable for building application. The addition of rising agents demonstrated the possibility to obtain a lighter structure with better insulation properties. Several tests were conducted to verify the new panel properties. The results obtained have shown the possibility to realize an innovative and sustainable material suitable to replace insulating panels currently used.

Keywords: Sustainability, recycling, waste cardboard, PCM, cladding system, insulating materials.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2269
5966 Influence of a High-Resolution Land Cover Classification on Air Quality Modelling

Authors: C. Silveira, A. Ascenso, J. Ferreira, A. I. Miranda, P. Tuccella, G. Curci

Abstract:

Poor air quality is one of the main environmental causes of premature deaths worldwide, and mainly in cities, where the majority of the population lives. It is a consequence of successive land cover (LC) and use changes, as a result of the intensification of human activities. Knowing these landscape modifications in a comprehensive spatiotemporal dimension is, therefore, essential for understanding variations in air pollutant concentrations. In this sense, the use of air quality models is very useful to simulate the physical and chemical processes that affect the dispersion and reaction of chemical species into the atmosphere. However, the modelling performance should always be evaluated since the resolution of the input datasets largely dictates the reliability of the air quality outcomes. Among these data, the updated LC is an important parameter to be considered in atmospheric models, since it takes into account the Earth’s surface changes due to natural and anthropic actions, and regulates the exchanges of fluxes (emissions, heat, moisture, etc.) between the soil and the air. This work aims to evaluate the performance of the Weather Research and Forecasting model coupled with Chemistry (WRF-Chem), when different LC classifications are used as an input. The influence of two LC classifications was tested: i) the 24-classes USGS (United States Geological Survey) LC database included by default in the model, and the ii) CLC (Corine Land Cover) and specific high-resolution LC data for Portugal, reclassified according to the new USGS nomenclature (33-classes). Two distinct WRF-Chem simulations were carried out to assess the influence of the LC on air quality over Europe and Portugal, as a case study, for the year 2015, using the nesting technique over three simulation domains (25 km2, 5 km2 and 1 km2 horizontal resolution). Based on the 33-classes LC approach, particular emphasis was attributed to Portugal, given the detail and higher LC spatial resolution (100 m x 100 m) than the CLC data (5000 m x 5000 m). As regards to the air quality, only the LC impacts on tropospheric ozone concentrations were evaluated, because ozone pollution episodes typically occur in Portugal, in particular during the spring/summer, and there are few research works relating to this pollutant with LC changes. The WRF-Chem results were validated by season and station typology using background measurements from the Portuguese air quality monitoring network. As expected, a better model performance was achieved in rural stations: moderate correlation (0.4 – 0.7), BIAS (10 – 21µg.m-3) and RMSE (20 – 30 µg.m-3), and where higher average ozone concentrations were estimated. Comparing both simulations, small differences grounded on the Leaf Area Index and air temperature values were found, although the high-resolution LC approach shows a slight enhancement in the model evaluation. This highlights the role of the LC on the exchange of atmospheric fluxes, and stresses the need to consider a high-resolution LC characterization combined with other detailed model inputs, such as the emission inventory, to improve air quality assessment.

Keywords: Land cover, tropospheric ozone, WRF-Chem, air quality assessment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 773
5965 An Efficient Biometric Cryptosystem using Autocorrelators

Authors: R. Bremananth, A. Chitra

Abstract:

Cryptography provides the secure manner of information transmission over the insecure channel. It authenticates messages based on the key but not on the user. It requires a lengthy key to encrypt and decrypt the sending and receiving the messages, respectively. But these keys can be guessed or cracked. Moreover, Maintaining and sharing lengthy, random keys in enciphering and deciphering process is the critical problem in the cryptography system. A new approach is described for generating a crypto key, which is acquired from a person-s iris pattern. In the biometric field, template created by the biometric algorithm can only be authenticated with the same person. Among the biometric templates, iris features can efficiently be distinguished with individuals and produces less false positives in the larger population. This type of iris code distribution provides merely less intra-class variability that aids the cryptosystem to confidently decrypt messages with an exact matching of iris pattern. In this proposed approach, the iris features are extracted using multi resolution wavelets. It produces 135-bit iris codes from each subject and is used for encrypting/decrypting the messages. The autocorrelators are used to recall original messages from the partially corrupted data produced by the decryption process. It intends to resolve the repudiation and key management problems. Results were analyzed in both conventional iris cryptography system (CIC) and non-repudiation iris cryptography system (NRIC). It shows that this new approach provides considerably high authentication in enciphering and deciphering processes.

Keywords: Autocorrelators, biometrics cryptography, irispatterns, wavelets.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1511
5964 Glorification Trap in Combating Human Trafficking in Indonesia: An Application of Three-Dimensional Model of Anti-Trafficking Policy

Authors: M. Kosandi, V. Susanti, N. I. Subono, E. Kartini

Abstract:

This paper discusses the risk of glorification trap in combating human trafficking, as it is shown in the case of Indonesia. Based on a research on Indonesian combat against trafficking in 2017-2018, this paper shows the tendency of misinterpretation and misapplication of the Indonesian anti-trafficking law into misusing the law for glorification, to create an image of certain extent of achievement in combating human trafficking. The objective of this paper is to explain the persistent occurrence of human trafficking crimes despite the significant progress of anti-trafficking efforts of Indonesian government. The research was conducted in 2017-2018 by qualitative approach through observation, depth interviews, discourse analysis, and document study, applying the three-dimensional model for analyzing human trafficking in the source country. This paper argues that the drive for glorification of achievement in the combat against trafficking has trapped Indonesian government in the loop of misinterpretation, misapplication, and misuse of the anti-trafficking law. In return, the so-called crime against humanity remains high and tends to increase in Indonesia.

Keywords: Human trafficking, anti-trafficking policy, transnational crime, source country, glorification trap.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 936
5963 Fast Adjustable Threshold for Uniform Neural Network Quantization

Authors: Alexander Goncharenko, Andrey Denisov, Sergey Alyamkin, Evgeny Terentev

Abstract:

The neural network quantization is highly desired procedure to perform before running neural networks on mobile devices. Quantization without fine-tuning leads to accuracy drop of the model, whereas commonly used training with quantization is done on the full set of the labeled data and therefore is both time- and resource-consuming. Real life applications require simplification and acceleration of quantization procedure that will maintain accuracy of full-precision neural network, especially for modern mobile neural network architectures like Mobilenet-v1, MobileNet-v2 and MNAS. Here we present a method to significantly optimize training with quantization procedure by introducing the trained scale factors for discretization thresholds that are separate for each filter. Using the proposed technique, we quantize the modern mobile architectures of neural networks with the set of train data of only ∼ 10% of the total ImageNet 2012 sample. Such reduction of train dataset size and small number of trainable parameters allow to fine-tune the network for several hours while maintaining the high accuracy of quantized model (accuracy drop was less than 0.5%). Ready-for-use models and code are available in the GitHub repository.

Keywords: Distillation, machine learning, neural networks, quantization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 711
5962 Flow Analysis of Viscous Nanofluid Due to Rotating Rigid Disk with Navier’s Slip: A Numerical Study

Authors: Khalil Ur Rehman, M. Y. Malik, Usman Ali

Abstract:

In this paper, the problem proposed by Von Karman is treated in the attendance of additional flow field effects when the liquid is spaced above the rotating rigid disk. To be more specific, a purely viscous fluid flow yield by rotating rigid disk with Navier’s condition is considered in both magnetohydrodynamic and hydrodynamic frames. The rotating flow regime is manifested with heat source/sink and chemically reactive species. Moreover, the features of thermophoresis and Brownian motion are reported by considering nanofluid model. The flow field formulation is obtained mathematically in terms of high order differential equations. The reduced system of equations is solved numerically through self-coded computational algorithm. The pertinent outcomes are discussed systematically and provided through graphical and tabular practices. A simultaneous way of study makes this attempt attractive in this sense that the article contains dual framework and validation of results with existing work confirms the execution of self-coded algorithm for fluid flow regime over a rotating rigid disk.

Keywords: Nanoparticles, Newtonian fluid model, chemical reaction, heat source/sink.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 973
5961 Effect of Bentonite on the Rheological Behavior of Cement Grout in Presence of Superplasticizer

Authors: K. Benyounes, A. Benmounah

Abstract:

Cement-based grouts has been used successfully to repair cracks in many concrete structures such as bridges, tunnels, buildings and to consolidate soils or rock foundations. In the present study the rheological characterization of cement grout with water/binder ratio (W/B) is fixed at 0.5. The effect of the replacement of cement by bentonite (2 to 10% wt) in presence of superplasticizer (0.5% wt) was investigated. Several rheological tests were carried out by using controlled-stress rheometer equipped with vane geometry in temperature of 20°C. To highlight the influence of bentonite and superplasticizer on the rheological behavior of grout cement, various flow tests in a range of shear rate from 0 to 200 s-1 were observed. Cement grout showed a non-Newtonian viscosity behavior at all concentrations of bentonite. Three parameter model Herschel- Bulkley was chosen for fitting of experimental data. Based on the values of correlation coefficients of the estimated parameters, The Herschel-Bulkley law model well described the rheological behavior of the grouts. Test results showed that the dosage of bentonite increases the viscosity and yield stress of the system and introduces more thixotropy. While the addition of both bentonite and superplasticizer with cement grout improve significantly the fluidity and reduced the yield stress due to the action of dispersion of SP.

Keywords: Cement grout, bentonite, superplasticizer, viscosity, yield stress.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3547
5960 Evaluating the Effect of Domestic Price on Rice Production in an African Setting: A Typical Evidence of the Sierra Leone Case

Authors: Alhaji M. H. Conteh, Xiangbin Yan, Alfred V Gborie

Abstract:

Rice, which is the staple food in Sierra Leone, is consumed on a daily basis. It is the most imperative food crop extensively grown by farmers across all ecologies in the country. Though much attention is now given to rice grain production through the small holder commercialization programme (SHCP), however, no attention has been given in investigating the limitations faced by rice producers. This paper will contribute to attempts to overcome the development challenges caused by food insecurity. The objective of this paper is thus, to analysis the relationship between rice production and the domestic retail price of rice. The study employed a log linear model in which, the quantity of rice produced is the dependent variable, quantity of rice imported, price of imported rice and price of domestic rice as explanatory variables. Findings showed that, locally produced rice is even more expensive than the imported rice per ton, and almost all the inhabitants in the capital city which hosts about 65% of the entire population of the country favor imported rice, as it is free from stones with other impurities. On the other hand, to control price and simultaneously increase rice production, the government should purchase the rice from the farmers and then sell to private retailers.

Keywords: Domestic price of rice, Econometric model, Rice production, Sierra Leone.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2464
5959 Comparison between the Efficiency of Heterojunction Thin Film InGaP\GaAs\Ge and InGaP\GaAs Solar Cell

Authors: F. Djaafar, B. Hadri, G. Bachir

Abstract:

This paper presents the design parameters for a thin film 3J InGaP/GaAs/Ge solar cell with a simulated maximum efficiency of 32.11% using Tcad Silvaco. Design parameters include the doping concentration, molar fraction, layers’ thickness and tunnel junction characteristics. An initial dual junction InGaP/GaAs model of a previous published heterojunction cell was simulated in Tcad Silvaco to accurately predict solar cell performance. To improve the solar cell’s performance, we have fixed meshing, material properties, models and numerical methods. However, thickness and layer doping concentration were taken as variables. We, first simulate the InGaP\GaAs dual junction cell by changing the doping concentrations and thicknesses which showed an increase in efficiency. Next, a triple junction InGaP/GaAs/Ge cell was modeled by adding a Ge layer to the previous dual junction InGaP/GaAs model with an InGaP /GaAs tunnel junction.

Keywords: Heterojunction, modeling, simulation, thin film, Tcad Silvaco.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1222
5958 A New Framework and a Model for Product Development with an Application in the Telecommunications Services Sector

Authors: Ghada A. El Khayat

Abstract:

This paper argues that a product development exercise involves in addition to the conventional stages, several decisions regarding other aspects. These aspects should be addressed simultaneously in order to develop a product that responds to the customer needs and that helps realize objectives of the stakeholders in terms of profitability, market share and the like. We present a framework that encompasses these different development dimensions. The framework shows that a product development methodology such as the Quality Function Deployment (QFD) is the basic tool which allows definition of the target specifications of a new product. Creativity is the first dimension that enables the development exercise to live and end successfully. A number of group processes need to be followed by the development team in order to ensure enough creativity and innovation. Secondly, packaging is considered to be an important extension of the product. Branding strategies, quality and standardization requirements, identification technologies, design technologies, production technologies and costing and pricing are also integral parts to the development exercise. These dimensions constitute the proposed framework. The paper also presents a mathematical model used to calculate the design targets based on the target costing principle. The framework is used to study a case of a new product development in the telecommunications services sector.

Keywords: Product Development Framework, Quality FunctionDeployment, Mathematical Models, Telecommunications.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1542
5957 Game based Learning to Enhance Cognitive and Physical Capabilities of Elderly People: Concepts and Requirements

Authors: Aurelie Aurilla Bechina Arntzen

Abstract:

The last decade has seen an early majority of people The last decade, the role of the of the information communication technologies has increased in improving the social and business life of people. Today, it is recognized that game could contribute to enhance virtual rehabilitation by better engaging patients. Our research study aims to develop a game based system enhancing cognitive and physical capabilities of elderly people. To this end, the project aims to develop a low cost hand held system based on existing game such as Wii, PSP, or Xbox. This paper discusses the concepts and requirements for developing such game for elderly people. Based on the requirement elicitation, we intend to develop a prototype related to sport and dance activities.

Keywords: Elderly people, Game based learning system, Health systems, rehabilitation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2497
5956 Determinants of Brand Equity: Offering a Model to Chocolate Industry

Authors: Emari Hossien

Abstract:

This study examined the underlying dimensions of brand equity in the chocolate industry. For this purpose, researchers developed a model to identify which factors are influential in building brand equity. The second purpose was to assess brand loyalty and brand images mediating effect between brand attitude, brand personality, brand association with brand equity. The study employed structural equation modeling to investigate the causal relationships between the dimensions of brand equity and brand equity itself. It specifically measured the way in which consumers’ perceptions of the dimensions of brand equity affected the overall brand equity evaluations. Data were collected from a sample of consumers of chocolate industry in Iran. The results of this empirical study indicate that brand loyalty and brand image are important components of brand equity in this industry. Moreover, the role of brand loyalty and brand image as mediating factors in the intention of brand equity are supported. The principal contribution of the present research is that it provides empirical evidence of the multidimensionality of consumer based brand equity, supporting Aaker´s and Keller´s conceptualization of brand equity. The present research also enriched brand equity building by incorporating the brand personality and brand image, as recommended by previous researchers. Moreover, creating the brand equity index in chocolate industry of Iran particularly is novel.

Keywords: brand equity, brand personality, structural equationmodeling, Iran.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3589
5955 Development of Coronal Field and Solar Wind Components for MHD Interplanetary Simulations

Authors: Ljubomir Nikolic, Larisa Trichtchenko

Abstract:

The connection between solar activity and adverse phenomena in the Earth’s environment that can affect space and ground based technologies has spurred interest in Space Weather (SW) research. A great effort has been put on the development of suitable models that can provide advanced forecast of SW events. With the progress in computational technology, it is becoming possible to develop operational large scale physics based models which can incorporate the most important physical processes and domains of the Sun-Earth system. In order to enhance our SW prediction capabilities we are developing advanced numerical tools. With operational requirements in mind, our goal is to develop a modular simulation framework of propagation of the disturbances from the Sun through interplanetary space to the Earth. Here, we report and discuss on the development of coronal field and solar wind components for a large scale MHD code. The model for these components is based on a potential field source surface model and an empirical Wang-Sheeley-Arge solar wind relation. 

Keywords: Space weather, numerical modeling, coronal field, solar wind.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2117
5954 FEM Simulation of HE Blast-Fragmentation Warhead and the Calculation of Lethal Range

Authors: G. Tanapornraweekit, W. Kulsirikasem

Abstract:

This paper presents the simulation of fragmentation warhead using a hydrocode, Autodyn. The goal of this research is to determine the lethal range of such a warhead. This study investigates the lethal range of warheads with and without steel balls as preformed fragments. The results from the FE simulation, i.e. initial velocities and ejected spray angles of fragments, are further processed using an analytical approach so as to determine a fragment hit density and probability of kill of a modelled warhead. In order to simulate a plenty of preformed fragments inside a warhead, the model requires expensive computation resources. Therefore, this study attempts to model the problem in an alternative approach by considering an equivalent mass of preformed fragments to the mass of warhead casing. This approach yields approximately 7% and 20% difference of fragment velocities from the analytical results for one and two layers of preformed fragments, respectively. The lethal ranges of the simulated warheads are 42.6 m and 56.5 m for warheads with one and two layers of preformed fragments, respectively, compared to 13.85 m for a warhead without preformed fragment. These lethal ranges are based on the requirement of fragment hit density. The lethal ranges which are based on the probability of kill are 27.5 m, 61 m and 70 m for warheads with no preformed fragment, one and two layers of preformed fragments, respectively.

Keywords: Lethal Range, Natural Fragment, Preformed Fragment, Warhead.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4277
5953 Stereotypical Motor Movement Recognition Using Microsoft Kinect with Artificial Neural Network

Authors: M. Jazouli, S. Elhoufi, A. Majda, A. Zarghili, R. Aalouane

Abstract:

Autism spectrum disorder is a complex developmental disability. It is defined by a certain set of behaviors. Persons with Autism Spectrum Disorders (ASD) frequently engage in stereotyped and repetitive motor movements. The objective of this article is to propose a method to automatically detect this unusual behavior. Our study provides a clinical tool which facilitates for doctors the diagnosis of ASD. We focus on automatic identification of five repetitive gestures among autistic children in real time: body rocking, hand flapping, fingers flapping, hand on the face and hands behind back. In this paper, we present a gesture recognition system for children with autism, which consists of three modules: model-based movement tracking, feature extraction, and gesture recognition using artificial neural network (ANN). The first one uses the Microsoft Kinect sensor, the second one chooses points of interest from the 3D skeleton to characterize the gestures, and the last one proposes a neural connectionist model to perform the supervised classification of data. The experimental results show that our system can achieve above 93.3% recognition rate.

Keywords: ASD, stereotypical motor movements, repetitive gesture, kinect, artificial neural network.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1882
5952 Evaluation of Fitts’ Law Index of Difficulty Formulation for Screen Size Variations

Authors: Hidehiko Okada, Takayuki Akiba

Abstract:

It is well-known as Fitts’ law that the time for a user to point a target on a GUI screen can be modeled as a linear function of “index of difficulty (ID).” In this paper, the authors investigate whether the traditional ID formulation is appropriate independently of device screen sizes. Result of our experiment reveals that the ID formulation may not consistently capture actual difficulty: users’ pointing performances are not consistent among pointing target variations of which index of difficulty are consistent. The term A/W may not be appropriate because the term causes the observed inconsistency. Based on this finding, the authors then evaluate the applicability of possible models other than Fitts’ one. Multiple regression models are found to be able to appropriately represent the effects of target design variations. The authors next make an attempt to improve the definition of ID in Fitts’ model. Our idea is to raise the size or the distance values depending on the screen size. The modified model is found to fit well to the users’ pointing data, which supports the idea. 

Keywords: Fitts’ law, pointing device, small screen, touch user interface, usability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1614
5951 In silico Repopulation Model of Various Tumour Cells during Treatment Breaks in Head and Neck Cancer Radiotherapy

Authors: Loredana G. Marcu, David Marcu, Sanda M. Filip

Abstract:

Advanced head and neck cancers are aggressive tumours, which require aggressive treatment. Treatment efficiency is often hindered by cancer cell repopulation during radiotherapy, which is due to various mechanisms triggered by the loss of tumour cells and involves both stem and differentiated cells. The aim of the current paper is to present in silico simulations of radiotherapy schedules on a virtual head and neck tumour grown with biologically realistic kinetic parameters. Using the linear quadratic formalism of cell survival after radiotherapy, altered fractionation schedules employing various treatment breaks for normal tissue recovery are simulated and repopulation mechanism implemented in order to evaluate the impact of various cancer cell contribution on tumour behaviour during irradiation. The model has shown that the timing of treatment breaks is an important factor influencing tumour control in rapidly proliferating tissues such as squamous cell carcinomas of the head and neck. Furthermore, not only stem cells but also differentiated cells, via the mechanism of abortive division, can contribute to malignant cell repopulation during treatment.

Keywords: Radiation, tumour repopulation, squamous cell carcinoma, stem cell.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1952
5950 Effect of Flowrate and Coolant Temperature on the Efficiency of Progressive Freeze Concentration on Simulated Wastewater

Authors: M. Jusoh, R. Mohd Yunus, M. A. Abu Hassan

Abstract:

Freeze concentration freezes or crystallises the water molecules out as ice crystals and leaves behind a highly concentrated solution. In conventional suspension freeze concentration where ice crystals formed as a suspension in the mother liquor, separation of ice is difficult. The size of the ice crystals is still very limited which will require usage of scraped surface heat exchangers, which is very expensive and accounted for approximately 30% of the capital cost. This research is conducted using a newer method of freeze concentration, which is progressive freeze concentration. Ice crystals were formed as a layer on the designed heat exchanger surface. In this particular research, a helical structured copper crystallisation chamber was designed and fabricated. The effect of two operating conditions on the performance of the newly designed crystallisation chamber was investigated, which are circulation flowrate and coolant temperature. The performance of the design was evaluated by the effective partition constant, K, calculated from the volume and concentration of the solid and liquid phase. The system was also monitored by a data acquisition tool in order to see the temperature profile throughout the process. On completing the experimental work, it was found that higher flowrate resulted in a lower K, which translated into high efficiency. The efficiency is the highest at 1000 ml/min. It was also found that the process gives the highest efficiency at a coolant temperature of -6 °C.

Keywords: Freeze concentration, progressive freeze concentration, freeze wastewater treatment, ice crystals.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2156
5949 Stating Best Commercialization Method: An Unanswered Question from Scholars and Practitioners

Authors: Saheed A. Gbadegeshin

Abstract:

Commercialization method is a means to make inventions available at the market for final consumption. It is described as an important tool for keeping business enterprises sustainable and improving national economic growth. Thus, there are several scholarly publications on it, either presenting or testing different methods for commercialization. However, young entrepreneurs, technologists and scientists would like to know the best method to commercialize their innovations. Then, this question arises: What is the best commercialization method? To answer the question, a systematic literature review was conducted, and practitioners were interviewed. The literary results revealed that there are many methods but new methods are needed to improve commercialization especially during these times of economic crisis and political uncertainty. Similarly, the empirical results showed there are several methods, but the best method is the one that reduces costs, reduces the risks associated with uncertainty, and improves customer participation and acceptability. Therefore, it was concluded that new commercialization method is essential for today's high technologies and a method was presented.

Keywords: Commercialization method, high technology, lean start-up methodology, technology, knowledge.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1286
5948 Matching Pursuit based Removal of Cardiac Pulse-Related Artifacts in EEG/fMRI

Authors: Rainer Schneider, Stephan Lau, Levin Kuhlmann, Simon Vogrin, Maciej Gratkowski, Mark Cook, Jens Haueisen

Abstract:

Cardiac pulse-related artifacts in the EEG recorded simultaneously with fMRI are complex and highly variable. Their effective removal is an unsolved problem. Our aim is to develop an adaptive removal algorithm based on the matching pursuit (MP) technique and to compare it to established methods using a visual evoked potential (VEP). We recorded the VEP inside the static magnetic field of an MR scanner (with artifacts) as well as in an electrically shielded room (artifact free). The MP-based artifact removal outperformed average artifact subtraction (AAS) and optimal basis set removal (OBS) in terms of restoring the EEG field map topography of the VEP. Subsequently, a dipole model was fitted to the VEP under each condition using a realistic boundary element head model. The source location of the VEP recorded inside the MR scanner was closest to that of the artifact free VEP after cleaning with the MP-based algorithm as well as with AAS. While none of the tested algorithms offered complete removal, MP showed promising results due to its ability to adapt to variations of latency, frequency and amplitude of individual artifact occurrences while still utilizing a common template.

Keywords: matching pursuit, ballistocardiogram, artifactremoval, EEG/fMRI.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1671
5947 Evaluation of Settlement of Coastal Embankments Using Finite Elements Method

Authors: Sina Fadaie, Seyed Abolhassan Naeini

Abstract:

Coastal embankments play an important role in coastal structures by reducing the effect of the wave forces and controlling the movement of sediments. Many coastal areas are underlain by weak and compressible soils. Estimation of during construction settlement of coastal embankments is highly important in design and safety control of embankments and appurtenant structures. Accordingly, selecting and establishing of an appropriate model with a reasonable level of complication is one of the challenges for engineers. Although there are advanced models in the literature regarding design of embankments, there is not enough information on the prediction of their associated settlement, particularly in coastal areas having considerable soft soils. Marine engineering study in Iran is important due to the existence of two important coastal areas located in the northern and southern parts of the country. In the present study, the validity of Terzaghi’s consolidation theory has been investigated. In addition, the settlement of these coastal embankments during construction is predicted by using special methods in PLAXIS software by the help of appropriate boundary conditions and soil layers. The results indicate that, for the existing soil condition at the site, some parameters are important to be considered in analysis. Consequently, a model is introduced to estimate the settlement of the embankments in such geotechnical conditions.

Keywords: Consolidation, coastal embankments, settlement, numerical methods, finite elements method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 770