Search results for: requirement specification
159 A Study on Removal Characteristics of (Mn2+) from Aqueous Solution by CNT
Authors: Nassereldeen A. Kabashi, Suleyman A. Muyibi. Mohammed E. Saeed., Farhana I. Yahya
Abstract:
It is important to remove manganese from water because of its effects on human and the environment. Human activities are one of the biggest contributors for excessive manganese concentration in the environment. The proposed method to remove manganese in aqueous solution by using adsorption as in carbon nanotubes (CNT) at different parameters: The parameters are CNT dosage, pH, agitation speed and contact time. Different pHs are pH 6.0, pH 6.5, pH 7.0, pH 7.5 and pH 8.0, CNT dosages are 5mg, 6.25mg, 7.5mg, 8.75mg or 10mg, contact time are 10 min, 32.5 min, 55 min, 87.5 min and 120 min while the agitation speeds are 100rpm, 150rpm, 200rpm, 250rpm and 300rpm. The parameters chosen for experiments are based on experimental design done by using Central Composite Design, Design Expert 6.0 with 4 parameters, 5 levels and 2 replications. Based on the results, condition set at pH 7.0, agitation speed of 300 rpm, 7.5mg and contact time 55 minutes gives the highest removal with 75.5%. From ANOVA analysis in Design Expert 6.0, the residual concentration will be very much affected by pH and CNT dosage. Initial manganese concentration is 1.2mg/L while the lowest residual concentration achieved is 0.294mg/L, which almost satisfy DOE Malaysia Standard B requirement. Therefore, further experiments must be done to remove manganese from model water to the required standard (0.2 mg/L) with the initial concentration set to 0.294 mg/L.Keywords: Adsorption, CNT, DOE, Manganese, Parameters.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1472158 The Use of the Flat Field Panel for the On-Ground Calibration of Metis Coronagraph on Board of Solar Orbiter
Authors: C. Casini, V. Da Deppo, P. Zuppella, P. Chioetto, A. Slemer, F. Frassetto, M. Romoli, F. Landini, M. Pancrazzi, V. Andretta, E. Antonucci, A. Bemporad, M. Casti, Y. De Leo, M. Fabi, S. Fineschi, F. Frassati, C. Grimani, G. Jerse, P. Heinzel, K. Heerlein, A. Liberatore, E. Magli, G. Naletto, G. Nicolini, M.G. Pelizzo, P. Romano, C. Sasso, D. Spadaro, M. Stangalini, T. Straus, R. Susino, L. Teriaca, M. Uslenghi, A. Volpicelli
Abstract:
Solar Orbiter, launched on February 9th 2020, is an ESA/NASA mission conceived to study the Sun. The payload is composed of 10 instruments, among which there is the Metis coronagraph. A coronagraph aims at taking images of the solar corona: the occulter element simulates a total solar eclipse. This work presents some of the results obtained in the visible light band (580-640 nm) using a flat field panel source. The flat field panel gives a uniform illumination; consequently, it has been used during the on-ground calibration for several purposes: evaluating the response of each pixel of the detector (linearity); and characterizing the Field of View of the coronagraph. As a conclusion, a major result is the verification that the requirement for the Field of View (FoV) of Metis is fulfilled. Some investigations are in progress in order to verify that the performance measured on-ground did not change after launch.Keywords: Space instrumentation, Metis, solar coronagraph, flat field.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 701157 Mathematical Modelling and Numerical Simulation of Maisotsenko Cycle
Authors: Rasikh Tariq, Fatima Z. Benarab
Abstract:
Evaporative coolers has a minimum potential to reach the wet-bulb temperature of intake air which is not enough to handle a large cooling load; therefore, it is not a feasible option to overcome cooling requirement of a building. The invention of Maisotsenko (M) cycle has led evaporative cooling technology to reach the sub-wet-bulb temperature of the intake air; therefore, it brings an innovation in evaporative cooling techniques. In this work, we developed a mathematical model of the Maisotsenko based air cooler by applying energy and mass balance laws on different air channels. The governing ordinary differential equations are discretized and simulated on MATLAB. The temperature and the humidity plots are shown in the simulation results. A parametric study is conducted by varying working air inlet conditions (temperature and humidity), inlet air velocity, geometric parameters and water temperature. The influence of these aforementioned parameters on the cooling effectiveness of the HMX is reported. Results have shown that the effectiveness of the M-Cycle is increased by increasing the ambient temperature and decreasing absolute humidity. An air velocity of 0.5 m/sec and a channel height of 6-8mm is recommended.
Keywords: Renewable energy, indirect evaporative cooling, Maisotsenko cycle, HMX, mathematical model, numerical simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1282156 Artificial Neural Network based Modeling of Evaporation Losses in Reservoirs
Authors: Surinder Deswal, Mahesh Pal
Abstract:
An Artificial Neural Network based modeling technique has been used to study the influence of different combinations of meteorological parameters on evaporation from a reservoir. The data set used is taken from an earlier reported study. Several input combination were tried so as to find out the importance of different input parameters in predicting the evaporation. The prediction accuracy of Artificial Neural Network has also been compared with the accuracy of linear regression for predicting evaporation. The comparison demonstrated superior performance of Artificial Neural Network over linear regression approach. The findings of the study also revealed the requirement of all input parameters considered together, instead of individual parameters taken one at a time as reported in earlier studies, in predicting the evaporation. The highest correlation coefficient (0.960) along with lowest root mean square error (0.865) was obtained with the input combination of air temperature, wind speed, sunshine hours and mean relative humidity. A graph between the actual and predicted values of evaporation suggests that most of the values lie within a scatter of ±15% with all input parameters. The findings of this study suggest the usefulness of ANN technique in predicting the evaporation losses from reservoirs.Keywords: Artificial neural network, evaporation losses, multiple linear regression, modeling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1982155 The Use of Seashell by-Products in Pervious Concrete Pavers
Authors: Dang Hanh Nguyen, Nassim Sebaibi, Mohamed Boutouil, Lydia Leleyter, Fabienne Baraud
Abstract:
Pervious concrete is a green alternative to conventional pavements with minimal fine aggregate and a high void content. Pervious concrete allows water to infiltrate through the pavement, thereby reducing the runoff and the requirement for stormwater management systems.
Seashell By-Products (SBP) are produced in an important quantity in France and are considered as waste. This work investigated to use SBP in pervious concrete and produce an even more environmentally friendly product, Pervious Concrete Pavers.
The research methodology involved substituting the coarse aggregate in the previous concrete mix design with 20%, 40% and 60% SBP. The testing showed that pervious concrete containing less than 40% SBP had strengths, permeability and void content which are comparable to the pervious concrete containing with only natural aggregate. The samples that contained 40% SBP or higher had a significant loss in strength and an increase in permeability and a void content from the control mix pervious concrete. On the basis of the results in this research, it was found that the natural aggregate can be substituted by SBP without affecting the delicate balance of a pervious concrete mix. Additional, it is recommended that the optimum replacement percentage for SBP in pervious concrete is 40 % direct replacement of natural coarse aggregate while maintaining the structural performance and drainage capabilities of the pervious concrete.
Keywords: Seashell by-products, pervious concrete pavers, permeability and mechanical strength.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4909154 Computer Modeling and Plant-Wide Dynamic Simulation for Industrial Flare Minimization
Authors: Sujing Wang, Song Wang, Jian Zhang, Qiang Xu
Abstract:
Flaring emissions during abnormal operating conditions such as plant start-ups, shut-downs, and upsets in chemical process industries (CPI) are usually significant. Flare minimization can help to save raw material and energy for CPI plants, and to improve local environmental sustainability. In this paper, a systematic methodology based on plant-wide dynamic simulation is presented for CPI plant flare minimizations under abnormal operating conditions. Since off-specification emission sources are inevitable during abnormal operating conditions, to significantly reduce flaring emission in a CPI plant, they must be either recycled to the upstream process for online reuse, or stored somewhere temporarily for future reprocessing, when the CPI plant manufacturing returns to stable operation. Thus, the off-spec products could be reused instead of being flared. This can be achieved through the identification of viable design and operational strategies during normal and abnormal operations through plant-wide dynamic scheduling, simulation, and optimization. The proposed study includes three stages of simulation works: (i) developing and validating a steady-state model of a CPI plant; (ii) transiting the obtained steady-state plant model to the dynamic modeling environment; and refining and validating the plant dynamic model; and (iii) developing flare minimization strategies for abnormal operating conditions of a CPI plant via a validated plant-wide dynamic model. This cost-effective methodology has two main merits: (i) employing large-scale dynamic modeling and simulations for industrial flare minimization, which involves various unit models for modeling hundreds of CPI plant facilities; (ii) dealing with critical abnormal operating conditions of CPI plants such as plant start-up and shut-down. Two virtual case studies on flare minimizations for start-up operation (over 50% of emission savings) and shut-down operation (over 70% of emission savings) of an ethylene plant have been employed to demonstrate the efficacy of the proposed study.
Keywords: Flare minimization, large-scale modeling and simulation, plant shut-down, plant start-up.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1736153 Research on IBR-Driven Distributed Collaborative Visualization System
Authors: Yin Runmin, Song Changfeng
Abstract:
Image-based Rendering(IBR) techniques recently reached in broad fields which leads to a critical challenge to build up IBR-Driven visualization platform where meets requirement of high performance, large bounds of distributed visualization resource aggregation and concentration, multiple operators deploying and CSCW design employing. This paper presents an unique IBR-based visualization dataflow model refer to specific characters of IBR techniques and then discusses prominent feature of IBR-Driven distributed collaborative visualization (DCV) system before finally proposing an novel prototype. The prototype provides a well-defined three level modules especially work as Central Visualization Server, Local Proxy Server and Visualization Aid Environment, by which data and control for collaboration move through them followed the previous dataflow model. With aid of this triple hierarchy architecture of that, IBR oriented application construction turns to be easy. The employed augmented collaboration strategy not only achieve convenient multiple users synchronous control and stable processing management, but also is extendable and scalable.Keywords: Image-Based Rendering, Distributed CollaborativeVisualization, Computer Supported Cooperative Work, Model andSimulation, Modular Visualization Environment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1487152 Energy Efficient Resource Allocation and Scheduling in Cloud Computing Platform
Authors: Shuen-Tai Wang, Ying-Chuan Chen, Yu-Ching Lin
Abstract:
There has been renewal of interest in the relation between Green IT and cloud computing in recent years. Cloud computing has to be a highly elastic environment which provides stable services to users. The growing use of cloud computing facilities has caused marked energy consumption, putting negative pressure on electricity cost of computing center or data center. Each year more and more network devices, storages and computers are purchased and put to use, but it is not just the number of computers that is driving energy consumption upward. We could foresee that the power consumption of cloud computing facilities will double, triple, or even more in the next decade. This paper aims at resource allocation and scheduling technologies that are short of or have not well developed yet to reduce energy utilization in cloud computing platform. In particular, our approach relies on recalling services dynamically onto appropriate amount of the machines according to user’s requirement and temporarily shutting down the machines after finish in order to conserve energy. We present initial work on integration of resource and power management system that focuses on reducing power consumption such that they suffice for meeting the minimizing quality of service required by the cloud computing platform.Keywords: Cloud computing, energy utilization, power consumption, resource allocation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1444151 Fingerprint Compression Using Contourlet Transform and Multistage Vector Quantization
Authors: S. Esakkirajan, T. Veerakumar, V. Senthil Murugan, R. Sudhakar
Abstract:
This paper presents a new fingerprint coding technique based on contourlet transform and multistage vector quantization. Wavelets have shown their ability in representing natural images that contain smooth areas separated with edges. However, wavelets cannot efficiently take advantage of the fact that the edges usually found in fingerprints are smooth curves. This issue is addressed by directional transforms, known as contourlets, which have the property of preserving edges. The contourlet transform is a new extension to the wavelet transform in two dimensions using nonseparable and directional filter banks. The computation and storage requirements are the major difficulty in implementing a vector quantizer. In the full-search algorithm, the computation and storage complexity is an exponential function of the number of bits used in quantizing each frame of spectral information. The storage requirement in multistage vector quantization is less when compared to full search vector quantization. The coefficients of contourlet transform are quantized by multistage vector quantization. The quantized coefficients are encoded by Huffman coding. The results obtained are tabulated and compared with the existing wavelet based ones.Keywords: Contourlet Transform, Directional Filter bank, Laplacian Pyramid, Multistage Vector Quantization
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2014150 Fuzzy C-Means Clustering for Biomedical Documents Using Ontology Based Indexing and Semantic Annotation
Authors: S. Logeswari, K. Premalatha
Abstract:
Search is the most obvious application of information retrieval. The variety of widely obtainable biomedical data is enormous and is expanding fast. This expansion makes the existing techniques are not enough to extract the most interesting patterns from the collection as per the user requirement. Recent researches are concentrating more on semantic based searching than the traditional term based searches. Algorithms for semantic searches are implemented based on the relations exist between the words of the documents. Ontologies are used as domain knowledge for identifying the semantic relations as well as to structure the data for effective information retrieval. Annotation of data with concepts of ontology is one of the wide-ranging practices for clustering the documents. In this paper, indexing based on concept and annotation are proposed for clustering the biomedical documents. Fuzzy c-means (FCM) clustering algorithm is used to cluster the documents. The performances of the proposed methods are analyzed with traditional term based clustering for PubMed articles in five different diseases communities. The experimental results show that the proposed methods outperform the term based fuzzy clustering.
Keywords: MeSH Ontology, Concept Indexing, Annotation, semantic relations, Fuzzy c-means.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2304149 MIMCA: A Modelling and Simulation Approach in Support of the Design and Construction of Manufacturing Control Systems Using Modular Petri net
Authors: S. Ariffin, K. Hasnan, R.H. Weston
Abstract:
A new generation of manufacturing machines so-called MIMCA (modular and integrated machine control architecture) capable of handling much increased complexity in manufacturing control-systems is presented. Requirement for more flexible and effective control systems for manufacturing machine systems is investigated and dimensioned-which highlights a need for improved means of coordinating and monitoring production machinery and equipment used to- transport material. The MIMCA supports simulation based on machine modeling, was conceived by the authors to address the issues. Essentially MIMCA comprises an organized unification of selected architectural frameworks and modeling methods, which include: NISTRCS, UMC and Colored Timed Petri nets (CTPN). The unification has been achieved; to support the design and construction of hierarchical and distributed machine control which realized the concurrent operation of reusable and distributed machine control components; ability to handle growing complexity; and support requirements for real- time control systems. Thus MIMCA enables mapping between 'what a machine should do' and 'how the machine does it' in a well-defined but flexible way designed to facilitate reconfiguration of machine systems.Keywords: Machine control, architectures, Petri nets, modularity, modeling, simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1588148 Impact of Process Parameters on Tensile Strength of Fused Deposition Modeling Printed Crisscross Poylactic Acid
Authors: Shilpesh R. Rajpurohit, Harshit K. Dave
Abstract:
Additive manufacturing gains the popularity in recent times, due to its capability to create prototype as well functional as end use product directly from CAD data without any specific requirement of tooling. Fused deposition modeling (FDM) is one of the widely used additive manufacturing techniques that are used to create functional end use part of polymer that is comparable with the injection-molded parts. FDM printed part has an application in various fields such as automobile, aerospace, medical, electronic, etc. However, application of FDM part is greatly affected by poor mechanical properties. Proper selection of the process parameter could enhance the mechanical performance of the printed part. In the present study, experimental investigation has been carried out to study the behavior of the mechanical performance of the printed part with respect to process variables. Three process variables viz. raster angle, raster width and layer height have been varied to understand its effect on tensile strength. Further, effect of process variables on fractured surface has been also investigated.
Keywords: 3D printing, fused deposition modeling, layer height, raster angle, raster width, tensile strength.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1661147 FEM Simulation of HE Blast-Fragmentation Warhead and the Calculation of Lethal Range
Authors: G. Tanapornraweekit, W. Kulsirikasem
Abstract:
This paper presents the simulation of fragmentation warhead using a hydrocode, Autodyn. The goal of this research is to determine the lethal range of such a warhead. This study investigates the lethal range of warheads with and without steel balls as preformed fragments. The results from the FE simulation, i.e. initial velocities and ejected spray angles of fragments, are further processed using an analytical approach so as to determine a fragment hit density and probability of kill of a modelled warhead. In order to simulate a plenty of preformed fragments inside a warhead, the model requires expensive computation resources. Therefore, this study attempts to model the problem in an alternative approach by considering an equivalent mass of preformed fragments to the mass of warhead casing. This approach yields approximately 7% and 20% difference of fragment velocities from the analytical results for one and two layers of preformed fragments, respectively. The lethal ranges of the simulated warheads are 42.6 m and 56.5 m for warheads with one and two layers of preformed fragments, respectively, compared to 13.85 m for a warhead without preformed fragment. These lethal ranges are based on the requirement of fragment hit density. The lethal ranges which are based on the probability of kill are 27.5 m, 61 m and 70 m for warheads with no preformed fragment, one and two layers of preformed fragments, respectively.Keywords: Lethal Range, Natural Fragment, Preformed Fragment, Warhead.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4322146 STEP Implementation on Turn-mill Manufacturing Environment
Authors: Ahmad Majdi Bin Abdul-Rani, Mesfin Gizaw, Yusri Yusof
Abstract:
Researches related to standard product model and development of neutral manufacturing interfaces for numerical control machines becomes a significant topic since the last 25 years. In this paper, a detail description of STEP implementation on turnmill manufacturing has been discussed. It shows requirements of information contents from ISO14649 data model. It covers to describe the design of STEP-NC framework applicable to turn-mill manufacturing. In the framework, EXPRESS-G and UML modeling tools are used to depict the information contents of the system and established the bases of information model requirement. A product and manufacturing data model applicable for STEP compliant manufacturing. The next generation turn-mill operations requirements have been represented by a UML diagram. An object oriented classes of ISO1449 has been developed on Visual Basic dot NET platform for binding the static information model represented by the UML diagram. An architect of the proposed system implementation has been given on the bases of the design and manufacturing module of STEP-NC interface established. Finally, a part 21 file process plan generated for an illustration of turn-mill components.Keywords: CAPP, ISO14649, Product modeling, STEP-NC
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1663145 Tagged Grid Matching Based Object Detection in Wavelet Neural Network
Authors: R. Arulmurugan, P. Sengottuvelan
Abstract:
Object detection using Wavelet Neural Network (WNN) plays a major contribution in the analysis of image processing. Existing cluster-based algorithm for co-saliency object detection performs the work on the multiple images. The co-saliency detection results are not desirable to handle the multi scale image objects in WNN. Existing Super Resolution (SR) scheme for landmark images identifies the corresponding regions in the images and reduces the mismatching rate. But the Structure-aware matching criterion is not paying attention to detect multiple regions in SR images and fail to enhance the result percentage of object detection. To detect the objects in the high-resolution remote sensing images, Tagged Grid Matching (TGM) technique is proposed in this paper. TGM technique consists of the three main components such as object determination, object searching and object verification in WNN. Initially, object determination in TGM technique specifies the position and size of objects in the current image. The specification of the position and size using the hierarchical grid easily determines the multiple objects. Second component, object searching in TGM technique is carried out using the cross-point searching. The cross out searching point of the objects is selected to faster the searching process and reduces the detection time. Final component performs the object verification process in TGM technique for identifying (i.e.,) detecting the dissimilarity of objects in the current frame. The verification process matches the search result grid points with the stored grid points to easily detect the objects using the Gabor wavelet Transform. The implementation of TGM technique offers a significant improvement on the multi-object detection rate, processing time, precision factor and detection accuracy level.
Keywords: Object Detection, Cross-point Searching, Wavelet Neural Network, Object Determination, Gabor Wavelet Transform, Tagged Grid Matching.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1965144 Experimental Investigation to Find Transition Temperature of VG-30 Binder
Authors: D. Latha, V. Sunitha, Samson Mathew
Abstract:
In India, most of the pavement is laid by bituminous road and the consumption of binder is high for pavement construction and also modified binders are used to satisfy any specific pavement requirement. Since the binders are visco-elastic material which is having the mechanical properties of binder transition from viscoelastic solid to visco-elastic fluid. In this paper, two different protocols were used to measure the viscosity property of binder using a Brookfield Viscometer and there is a need to find the appropriate mixing and compaction temperatures of various types of binders which can result in complete aggregate coating and adequate field density of HMA mixtures. The aim of this work is to find the transition temperature from Non-Newtonian behavior to Newtonian behavior of the binder by adopting a steady shear protocol and the shear rate ramp protocol. The transition from non-Newtonian to Newtonian can occur through an increase of temperature and shear of the material. The test has been conducted for unmodified binder VG 30. The transition temperature was found in the unmodified binder VG is 120oC. Therefore, the application of both modified binder and unmodified binder in the pavement construction needs to be studied properly by considering temperature and traffic loading factors of the respective project site.Keywords: Unmodified and modified binders, Brookfield Viscometer, transition temperature, steady shear, shear rate protocol.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1876143 Laboratory Investigation of the Pavement Condition in Lebanon: Implementation of Reclaimed Asphalt Pavement in the Base Course and Asphalt Layer
Authors: Marinelle El-Khoury, Lina Bouhaya, Nivine Abbas, Hassan Sleiman
Abstract:
The road network in the north of Lebanon is a prime example of the lack of pavement design and execution in Lebanon. These roads show major distresses and hence, should be tested and evaluated. The aim of this research is to investigate and determine the deficiencies in road surface design in Lebanon, and to propose an environmentally friendly asphalt mix design. This paper consists of several parts: (i) evaluating pavement performance and structural behavior, (ii) identifying the distresses using visual examination followed by laboratory tests, (iii) deciding the optimal solution where rehabilitation or reconstruction is required and finally, (iv) identifying a sustainable method, which uses recycled material in the proposed mix. The asphalt formula contains Reclaimed Asphalt Pavement (RAP) in the base course layer and in the asphalt layer. Visual inspection of the roads in Tripoli shows that these roads face a high level of distress severity. Consequently, the pavement should be reconstructed rather than simply rehabilitated. Coring was done to determine the pavement layer thickness. The results were compared to the American Association of State Highway and Transportation Officials (AASHTO) design methodology and showed that the existing asphalt thickness is lower than the required asphalt thickness. Prior to the pavement reconstruction, the road materials were tested according to the American Society for Testing and Materials (ASTM) specification to identify whether the materials are suitable. Accordingly, the ASTM tests that were performed on the base course are Sieve analysis, Atterberg limits, modified proctor, Los Angeles, and California Bearing Ratio (CBR) tests. Results show a CBR value higher than 70%. Hence, these aggregates could be used as a base course layer. The asphalt layer was also tested and the results of the Marshall flow and stability tests meet the ASTM specifications. In the last section, an environmentally friendly mix was proposed. An optimal RAP percentage of 30%, which produced a well graded base course and asphalt mix, was determined through a series of trials.Keywords: Asphalt mix, reclaimed asphalt pavement, California bearing ratio, sustainability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 730142 A Reusability Evaluation Model for OO-Based Software Components
Authors: Parvinder S. Sandhu, Hardeep Singh
Abstract:
The requirement to improve software productivity has promoted the research on software metric technology. There are metrics for identifying the quality of reusable components but the function that makes use of these metrics to find reusability of software components is still not clear. These metrics if identified in the design phase or even in the coding phase can help us to reduce the rework by improving quality of reuse of the component and hence improve the productivity due to probabilistic increase in the reuse level. CK metric suit is most widely used metrics for the objectoriented (OO) software; we critically analyzed the CK metrics, tried to remove the inconsistencies and devised the framework of metrics to obtain the structural analysis of OO-based software components. Neural network can learn new relationships with new input data and can be used to refine fuzzy rules to create fuzzy adaptive system. Hence, Neuro-fuzzy inference engine can be used to evaluate the reusability of OO-based component using its structural attributes as inputs. In this paper, an algorithm has been proposed in which the inputs can be given to Neuro-fuzzy system in form of tuned WMC, DIT, NOC, CBO , LCOM values of the OO software component and output can be obtained in terms of reusability. The developed reusability model has produced high precision results as expected by the human experts.Keywords: CK-Metric, ID3, Neuro-fuzzy, Reusability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1819141 The Effect of Cross-Curriculum of L1 and L2 on Elementary School Students’ Linguistic Proficiency: To Sympathize with Others
Authors: Reiko Yamamoto
Abstract:
This paper reports on a project to integrate Japanese (as a first language) and English (as a second language) education. This study focuses on the mutual effects of the two languages on the linguistic proficiency of elementary school students. The research team consisted of elementary school teachers and researchers at a university. The participants of the experiment were students between 3rd and 6th grades at an elementary school. The research process consisted of seven steps: 1) specifying linguistic proficiency; 2) developing the cross-curriculum of L1 and L2; 3) forming can-do statements; 4) creating a self-evaluation questionnaire; 5) executing the self-evaluation questionnaire at the beginning of the school year; 6) instructing L1 and L2 based on the curriculum; and 7) executing the self-evaluation questionnaire at the beginning of the next school year. In Step 1, the members of the research team brainstormed ways to specify elementary school students’ linguistic proficiency that can be observed in various scenes. It was revealed that the teachers evaluate their students’ linguistic proficiency on the basis of the students’ utterances, but also informed by their non-verbal communication abilities. This led to the idea that competency for understanding others’ minds through the use of physical movement or bodily senses in communication in L1 – to sympathize with others – can be transferred to that same competency in communication in L2. Based on the specification of linguistic proficiency that L1 and L2 have in common, a cross-curriculum of L1 and L2 was developed in Step 2. In Step 3, can-do statements based on the curriculum were also formed, building off of the action-oriented approach from the Common European Framework of Reference for Languages (CEFR) used in Europe. A self-evaluation questionnaire consisting of the main can-do statements was given to the students between 3rd grade and 6th grade at the beginning of the school year (Step 4 and Step 5), and all teachers gave L1 and L2 instruction based on the curriculum to the students for one year (Step 6). The same questionnaire was given to the students at the beginning of the next school year (Step 7). The results of statistical analysis proved the enhancement of the students’ linguistic proficiency. This verified the validity of developing the cross-curriculum of L1 and L2 and adapting it in elementary school. It was concluded that elementary school students do not distinguish between L1 and L2, and that they just try to understand others’ minds through physical movement or senses in any language.Keywords: Cross-curriculum of L1 and L2, elementary school education, language proficiency, sympathy with others.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1309140 Co-tier and Co-channel Interference Avoidance Algorithm for Femtocell Networks
Authors: S. Padmapriya, M. Tamilarasi
Abstract:
Femtocells are regarded as a milestone for next generation cellular networks. As femtocells are deployed in an unplanned manner, there is a chance of assigning same resource to neighboring femtocells. This scenario may induce co-channel interference and may seriously affect the service quality of neighboring femtocells. In addition, the dominant transmit power of a femtocell will induce co-tier interference to neighboring femtocells. Thus to jointly handle co-tier and co-channel interference, we propose an interference-free power and resource block allocation (IFPRBA) algorithm for closely located, closed access femtocells. Based on neighboring list, inter-femto-base station distance and uplink noise power, the IFPRBA algorithm assigns non-interfering power and resource to femtocells. The IFPRBA algorithm also guarantees the quality of service to femtouser based on the knowledge of resource requirement, connection type, and the tolerable delay budget. Simulation result shows that the interference power experienced in IFPRBA algorithm is below the tolerable interference power and hence the overall service success ratio, PRB efficiency and network throughput are maximum when compared to conventional resource allocation framework for femtocell (RAFF) algorithm.
Keywords: Co-channel interference, co-tier interference, femtocells, guaranteed QoS, power optimization, resource assignment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2467139 Detecting Interactions between Behavioral Requirements with OWL and SWRL
Authors: Haibo Hu, Dan Yang, Chunxiao Ye, Chunlei Fu, Ren Li
Abstract:
High quality requirements analysis is one of the most crucial activities to ensure the success of a software project, so that requirements verification for software system becomes more and more important in Requirements Engineering (RE) and it is one of the most helpful strategies for improving the quality of software system. Related works show that requirement elicitation and analysis can be facilitated by ontological approaches and semantic web technologies. In this paper, we proposed a hybrid method which aims to verify requirements with structural and formal semantics to detect interactions. The proposed method is twofold: one is for modeling requirements with the semantic web language OWL, to construct a semantic context; the other is a set of interaction detection rules which are derived from scenario-based analysis and represented with semantic web rule language (SWRL). SWRL based rules are working with rule engines like Jess to reason in semantic context for requirements thus to detect interactions. The benefits of the proposed method lie in three aspects: the method (i) provides systematic steps for modeling requirements with an ontological approach, (ii) offers synergy of requirements elicitation and domain engineering for knowledge sharing, and (3)the proposed rules can systematically assist in requirements interaction detection.Keywords: Requirements Engineering, Semantic Web, OWL, Requirements Interaction Detection, SWRL.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1798138 RUPSec: An Extension on RUP for Developing Secure Systems - Requirements Discipline
Authors: Mohammad Reza Ayatollahzadeh Shirazi, Pooya Jaferian, Golnaz Elahi, Hamid Baghi, Babak Sadeghian
Abstract:
The world is moving rapidly toward the deployment of information and communication systems. Nowadays, computing systems with their fast growth are found everywhere and one of the main challenges for these systems is increasing attacks and security threats against them. Thus, capturing, analyzing and verifying security requirements becomes a very important activity in development process of computing systems, specially in developing systems such as banking, military and e-business systems. For developing every system, a process model which includes a process, methods and tools is chosen. The Rational Unified Process (RUP) is one of the most popular and complete process models which is used by developers in recent years. This process model should be extended to be used in developing secure software systems. In this paper, the Requirement Discipline of RUP is extended to improve RUP for developing secure software systems. These proposed extensions are adding and integrating a number of Activities, Roles, and Artifacts to RUP in order to capture, document and model threats and security requirements of system. These extensions introduce a group of clear and stepwise activities to developers. By following these activities, developers assure that security requirements are captured and modeled. These models are used in design, implementation and test activitie Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2812137 Study on Status and Development of Hydraulic System Protection: Pump Combined With Air Chamber
Authors: I. Abuiziah, A. Oulhaj, K. Sebari, D. Ouazar, A. A. Saber
Abstract:
Fluid transient analysis is one of the more challenging and complicated flow problems in the design and the operation of water pipeline systems (wps). When transient conditions "water hammer" exists, the life expectancy of the wps can be adversely impacted, resulting in pump and valve failures and catastrophic pipe ruptures. Transient control has become an essential requirement for ensuring safe operation of wps. An accurate analysis and suitable protection devices should be used to protect wps. This paper presents the problem of modeling and simulation of transient phenomena in wps based on the characteristics method. Also, it provides the influence of using the protection devices to control the adverse effects due to excessive and low pressure occur in the transient. The developed model applied for main wps: pump combined with closed surge tank connected to a reservoir. The results obtained provide that the model is an efficient tool for water hammer analysis. Moreover; using the closed surge tank reduces the unfavorable effects of transients.
Keywords: Flow Transient, Water hammer, Pipeline System, Closed Surge Tank, Simulation Model, Protection Devices, Characteristics Method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2214136 Optimization of Element Type for FE Model and Verification of Analyses with Physical Tests
Authors: M. Tufekci, C. Guven
Abstract:
In Automotive Industry, sliding door systems that are also used as body closures are safety members. Extreme product tests are realized to prevent failures in design process, but these tests realized experimentally result in high costs. Finite element analysis is an effective tool used for design process. These analyses are used before production of prototype for validation of design according to customer requirement. In result of this, substantial amount of time and cost is saved. Finite element model is created for geometries that are designed in 3D CAD programs. Different element types as bar, shell and solid, can be used for creating mesh model. Cheaper model can be created by selection of element type, but combination of element type that was used in model, number and geometry of element and degrees of freedom affects the analysis result. Sliding door system is a good example which used these methods for this study. Structural analysis was realized for sliding door mechanism by using FE models. As well, physical tests that have same boundary conditions with FE models were realized. Comparison study for these element types, were done regarding test and analyses results then optimum combination was achieved.Keywords: Finite Element Analysis, Sliding Door Mechanism, Element Type, Structural Analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1965135 The Effects of Sewage Sludge Usage and Manure on Some Heavy Metals Uptake in Savory (Satureja hortensis L.)
Authors: A. Hani
Abstract:
In recent decades with the development of technology and lack of food sources, sewage sludge in production of human foods is inevitable. Various sources of municipal and industrial sewage sludge that is produced can provide the requirement of plant nutrients. Soils in arid, semi-arid climate of central Iran that most affected by water drainage, iron and zinc deficiencies, using of sewage sludge is helpful. Therefore, the aim of this study is investigation of sewage sludge and manure application on Ni, Pb and Cd uptake by Savory. An experiment in a randomized complete block design with three replications was performed. Sewage sludge treatments consisted of four levels, control, 15, 30, 80 tons per hectares; the manure was used in four levels of control, 20, 40 and 80 tons per hectare. Results showed that the wet and dry weights was not affected by sewage sludge using, while, manure has significant effect on them. The effect of sewage sludge on the cadmium and lead concentrations were significant. Interactions of sewage sludge and manure on dry weight values were not significant. Compare mean analysis showed that increasing the amount of sewage sludge had no significant effect on cadmium concentration and it reduced when sewage sludge usage increased. This is probably due to increased plant growth and reduced concentrations of these elements in the plant.
Keywords: Savory, lead, cadmium, sewage sludge, manure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1726134 An Efficient Approach to Mining Frequent Itemsets on Data Streams
Authors: Sara Ansari, Mohammad Hadi Sadreddini
Abstract:
The increasing importance of data stream arising in a wide range of advanced applications has led to the extensive study of mining frequent patterns. Mining data streams poses many new challenges amongst which are the one-scan nature, the unbounded memory requirement and the high arrival rate of data streams. In this paper, we propose a new approach for mining itemsets on data stream. Our approach SFIDS has been developed based on FIDS algorithm. The main attempts were to keep some advantages of the previous approach and resolve some of its drawbacks, and consequently to improve run time and memory consumption. Our approach has the following advantages: using a data structure similar to lattice for keeping frequent itemsets, separating regions from each other with deleting common nodes that results in a decrease in search space, memory consumption and run time; and Finally, considering CPU constraint, with increasing arrival rate of data that result in overloading system, SFIDS automatically detect this situation and discard some of unprocessing data. We guarantee that error of results is bounded to user pre-specified threshold, based on a probability technique. Final results show that SFIDS algorithm could attain about 50% run time improvement than FIDS approach.Keywords: Data stream, frequent itemset, stream mining.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1421133 Selection the Optimum Cooling Scheme for Generators based on the Electro-Thermal Analysis
Authors: Diako Azizi, Ahmad Gholami, Vahid Abbasi
Abstract:
Optimal selection of electrical insulations in electrical machinery insures reliability during operation. From the insulation studies of view for electrical machines, stator is the most important part. This fact reveals the requirement for inspection of the electrical machine insulation along with the electro-thermal stresses. In the first step of the study, a part of the whole structure of machine in which covers the general characteristics of the machine is chosen, then based on the electromagnetic analysis (finite element method), the machine operation is simulated. In the simulation results, the temperature distribution of the total structure is presented simultaneously by using electro-thermal analysis. The results of electro-thermal analysis can be used for designing an optimal cooling system. In order to design, review and comparing the cooling systems, four wiring structures in the slots of Stator are presented. The structures are compared to each other in terms of electrical, thermal distribution and remaining life of insulation by using Finite Element analysis. According to the steps of the study, an optimization algorithm has been presented for selection of appropriate structure.Keywords: Electrical field, field distribution, insulation, winding, finite element method, electro thermal
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1750132 Quad Tree Decomposition Based Analysis of Compressed Image Data Communication for Lossy and Lossless Using WSN
Authors: N. Muthukumaran, R. Ravi
Abstract:
The Quad Tree Decomposition based performance analysis of compressed image data communication for lossy and lossless through wireless sensor network is presented. Images have considerably higher storage requirement than text. While transmitting a multimedia content there is chance of the packets being dropped due to noise and interference. At the receiver end the packets that carry valuable information might be damaged or lost due to noise, interference and congestion. In order to avoid the valuable information from being dropped various retransmission schemes have been proposed. In this proposed scheme QTD is used. QTD is an image segmentation method that divides the image into homogeneous areas. In this proposed scheme involves analysis of parameters such as compression ratio, peak signal to noise ratio, mean square error, bits per pixel in compressed image and analysis of difficulties during data packet communication in Wireless Sensor Networks. By considering the above, this paper is to use the QTD to improve the compression ratio as well as visual quality and the algorithm in MATLAB 7.1 and NS2 Simulator software tool.
Keywords: Image compression, Compression Ratio, Quad tree decomposition, Wireless sensor networks, NS2 simulator.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2391131 Cyber Security Situational Awareness among Students: A Case Study in Malaysia
Authors: Yunos Zahri, Ab Hamid R. Susanty, Ahmad Mustaffa
Abstract:
This paper explores the need for a national baseline study on understanding the level of cyber security situational awareness among primary and secondary school students in Malaysia. The online survey method was deployed to administer the data collection exercise. The target groups were divided into three categories: Group 1 (primary school aged 7-9 years old), Group 2 (primary school aged 10-12 years old), and Group 3 (secondary school aged 13-17 years old). A different questionnaire set was designed for each group. The survey topics/areas included Internet and digital citizenship knowledge. Respondents were randomly selected from rural and urban areas throughout all 14 states in Malaysia. A total of 9,158 respondents participated in the survey, with most states meeting the minimum sample size requirement to represent the country’s demographics. The findings and recommendations from this baseline study are fundamental to develop teaching modules required for children to understand the security risks and threats associated with the Internet throughout their years in school. Early exposure and education will help ensure healthy cyber habits among millennials in Malaysia.
Keywords: Cyber security awareness, cyber security education, cyber security, students.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2990130 Assessing the Global Water Productivity of Some Irrigation Command Areas in Iran
Authors: A. Montazar
Abstract:
The great challenge of the agricultural sector is to produce more crop from less water, which can be achieved by increasing crop water productivity. The modernization of the irrigation systems offers a number of possibilities to expand the economic productivity of water and improve the virtual water status. The objective of the present study is to assess the global water productivity (GWP) within the major irrigation command areas of I.R. Iran. For this purpose, fourteen irrigation command areas where located in different areas of Iran were selected. In order to calculate the global water productivity of irrigation command areas, all data on the delivered water to cropping pattern, cultivated area, crops water requirement, and yield production rate during 2002-2006 were gathered. In each of the command areas it seems that the cultivated crops have a higher amount of virtual water and thus can be replaced by crops with less virtual water. This is merely suggested due to crop water consumption and at the time of replacing crops, economic value as well as cultural and political factors must be considered. The results indicated that the lowest GWP belongs to Mahyar and Borkhar irrigation areas, 0.24 kg m-3, and the highest is that of the Dez irrigation area, 0.81 kg m-3. The findings demonstrated that water management in the two irrigation areas is just efficient. The difference in the GWP of irrigation areas is due to variations in the cropping pattern, amount of crop productions, in addition to the effective factors in the water use efficiency in the irrigation areas.Keywords: Iran, Irrigation command area, Water productivity, Virtual water.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1678