Search results for: Sumanpreet Kaur
41 Solution of Fuzzy Maximal Flow Problems Using Fuzzy Linear Programming
Authors: Amit Kumar, Manjot Kaur
Abstract:
In this paper, the fuzzy linear programming formulation of fuzzy maximal flow problems are proposed and on the basis of the proposed formulation a method is proposed to find the fuzzy optimal solution of fuzzy maximal flow problems. In the proposed method all the parameters are represented by triangular fuzzy numbers. By using the proposed method the fuzzy optimal solution of fuzzy maximal flow problems can be easily obtained. To illustrate the proposed method a numerical example is solved and the obtained results are discussed.Keywords: Fuzzy linear programming, Fuzzy maximal flow problem, Ranking function, Triangular fuzzy number
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 197540 Application of Artificial Neural Network for Predicting Maintainability Using Object-Oriented Metrics
Authors: K. K. Aggarwal, Yogesh Singh, Arvinder Kaur, Ruchika Malhotra
Abstract:
Importance of software quality is increasing leading to development of new sophisticated techniques, which can be used in constructing models for predicting quality attributes. One such technique is Artificial Neural Network (ANN). This paper examined the application of ANN for software quality prediction using Object- Oriented (OO) metrics. Quality estimation includes estimating maintainability of software. The dependent variable in our study was maintenance effort. The independent variables were principal components of eight OO metrics. The results showed that the Mean Absolute Relative Error (MARE) was 0.265 of ANN model. Thus we found that ANN method was useful in constructing software quality model.
Keywords: Software quality, Measurement, Metrics, Artificial neural network, Coupling, Cohesion, Inheritance, Principal component analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 257339 Comparison of Imputation Techniques for Efficient Prediction of Software Fault Proneness in Classes
Authors: Geeta Sikka, Arvinder Kaur Takkar, Moin Uddin
Abstract:
Missing data is a persistent problem in almost all areas of empirical research. The missing data must be treated very carefully, as data plays a fundamental role in every analysis. Improper treatment can distort the analysis or generate biased results. In this paper, we compare and contrast various imputation techniques on missing data sets and make an empirical evaluation of these methods so as to construct quality software models. Our empirical study is based on NASA-s two public dataset. KC4 and KC1. The actual data sets of 125 cases and 2107 cases respectively, without any missing values were considered. The data set is used to create Missing at Random (MAR) data Listwise Deletion(LD), Mean Substitution(MS), Interpolation, Regression with an error term and Expectation-Maximization (EM) approaches were used to compare the effects of the various techniques.Keywords: Missing data, Imputation, Missing Data Techniques.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 166738 Analytical Study of Component Based Software Engineering
Authors: Iqbaldeep Kaur, Parvinder S. Sandhu, Hardeep Singh, Vandana Saini
Abstract:
This paper is a survey of current component-based software technologies and the description of promotion and inhibition factors in CBSE. The features that software components inherit are also discussed. Quality Assurance issues in componentbased software are also catered to. The feat research on the quality model of component based system starts with the study of what the components are, CBSE, its development life cycle and the pro & cons of CBSE. Various attributes are studied and compared keeping in view the study of various existing models for general systems and CBS. When illustrating the quality of a software component an apt set of quality attributes for the description of the system (or components) should be selected. Finally, the research issues that can be extended are tabularized.Keywords: Component, COTS, Component based development, Component-based Software Engineering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 274037 Design and Optimization of Coplanar Waveguide-Fed Sensing Antennas for ISM Band Applications
Authors: Vivek Bharti, Inderpreet Kaur, Saurabh Verma, Renu Gangwar, Hari Kumar Singh
Abstract:
This work presents a dual-band Coplanar Waveguide (CPW) fed rectangular patch antenna designed for applications in Industrial, Scientific, and Medical (ISM) Band sensing. Fabricated on a cost-effective FR-4 substrate with specific dimensions, the antenna incorporates a rectangular slot and a copper patch, optimized through simulation to achieve desired characteristics, particularly focusing on the reflection coefficient. A unique feature is introduced through the integration of copper cells forming a Partially Reflecting Surface (PRS) to feed the antenna. Dual-band functionality is achieved, covering frequencies of 1.28 GHz-1.3 GHz and 4.9 GHz-5.2 GHz, catering to diverse communication needs within these frequency ranges. The addition of a superstrate enhances the antenna’s gain, resulting in 4.5 dBi at 1.28 GHz-1.3 GHz and 6.5 dBi at 4.9 GHz-5.2 GHz, with an efficiency of 58%.
Keywords: ISM, Coplanar waveguide fed, superstrate, circularly polarized.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13436 Comparative Analysis of the Software Effort Estimation Models
Authors: Jaswinder Kaur, Satwinder Singh, Karanjeet Singh Kahlon
Abstract:
Accurate software cost estimates are critical to both developers and customers. They can be used for generating request for proposals, contract negotiations, scheduling, monitoring and control. The exact relationship between the attributes of the effort estimation is difficult to establish. A neural network is good at discovering relationships and pattern in the data. So, in this paper a comparative analysis among existing Halstead Model, Walston-Felix Model, Bailey-Basili Model, Doty Model and Neural Network Based Model is performed. Neural Network has outperformed the other considered models. Hence, we proposed Neural Network system as a soft computing approach to model the effort estimation of the software systems.Keywords: Effort Estimation, Neural Network, Halstead Model, Walston-Felix Model, Bailey-Basili Model, Doty Model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 222135 Modeling of Reusability of Object Oriented Software System
Authors: Parvinder S. Sandhu, Harpreet Kaur, Amanpreet Singh
Abstract:
Automatic reusability appraisal is helpful in evaluating the quality of developed or developing reusable software components and in identification of reusable components from existing legacy systems; that can save cost of developing the software from scratch. But the issue of how to identify reusable components from existing systems has remained relatively unexplored. In this research work, structural attributes of software components are explored using software metrics and quality of the software is inferred by different Neural Network based approaches, taking the metric values as input. The calculated reusability value enables to identify a good quality code automatically. It is found that the reusability value determined is close to the manual analysis used to be performed by the programmers or repository managers. So, the developed system can be used to enhance the productivity and quality of software development.Keywords: Neural Network, Software Reusability, Software Metric, Accuracy, MAE, RMSE.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 208134 Analysis of One Dimensional Advection Diffusion Model Using Finite Difference Method
Authors: Vijay Kumar Kukreja, Ravneet Kaur
Abstract:
In this paper, one dimensional advection diffusion model is analyzed using finite difference method based on Crank-Nicolson scheme. A practical problem of filter cake washing of chemical engineering is analyzed. The model is converted into dimensionless form. For the grid Ω × ω = [0, 1] × [0, T], the Crank-Nicolson spatial derivative scheme is used in space domain and forward difference scheme is used in time domain. The scheme is found to be unconditionally convergent, stable, first order accurate in time and second order accurate in space domain. For a test problem, numerical results are compared with the analytical ones for different values of parameter.Keywords: Consistency, Crank-Nicolson scheme, Gerschgorin circle, Lax-Richtmyer theorem, Peclet number, stability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 76133 A K-Means Based Clustering Approach for Finding Faulty Modules in Open Source Software Systems
Authors: Parvinder S. Sandhu, Jagdeep Singh, Vikas Gupta, Mandeep Kaur, Sonia Manhas, Ramandeep Sidhu
Abstract:
Prediction of fault-prone modules provides one way to support software quality engineering. Clustering is used to determine the intrinsic grouping in a set of unlabeled data. Among various clustering techniques available in literature K-Means clustering approach is most widely being used. This paper introduces K-Means based Clustering approach for software finding the fault proneness of the Object-Oriented systems. The contribution of this paper is that it has used Metric values of JEdit open source software for generation of the rules for the categorization of software modules in the categories of Faulty and non faulty modules and thereafter empirically validation is performed. The results are measured in terms of accuracy of prediction, probability of Detection and Probability of False Alarms.Keywords: K-Means, Software Fault, Classification, ObjectOriented Metrics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 230432 Performance Evaluation of One and Two Dimensional Prime Codes for Optical Code Division Multiple Access Systems
Authors: Gurjit Kaur, Neena Gupta
Abstract:
In this paper, we have analyzed and compared the performance of various coding schemes. The basic ID prime sequence codes are unique in only dimension, i.e. time slots, whereas 2D coding techniques are not unique by their time slots but with their wavelengths also. In this research, we have evaluated and compared the performance of 1D and 2D coding techniques constructed using prime sequence coding pattern for Optical Code Division Multiple Access (OCDMA) system on a single platform. Analysis shows that 2D prime code supports lesser number of active users than 1D codes, but they are having large code family and are the most secure codes compared to other codes. The performance of all these codes is analyzed on basis of number of active users supported at a Bit Error Rate (BER) of 10-9.Keywords: CDMA, OCDMA, BER, OOC, PC, EPC, MPC, 2-D PC/PC, λc, λa.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 112331 Dielectric and Impedance Spectroscopy of Samarium and Lanthanum Doped Barium Titanate at Room Temperature
Authors: Sukhleen Bindra Narang, Dalveer Kaur, Kunal Pubby
Abstract:
Dielectric ceramic samples in the BaO-Re2O3-TiO2 ternary system were synthesized with structural formula Ba2- xRe4+2x/3Ti8O24 where Re= rare earth metal and Re= Sm and La where x varies from 0.0 to 0.6 with step size 0.1. Polycrystalline samples were prepared by the conventional solid state reaction technique. The dielectric, electrical and impedance analysis of all the samples in the frequency range 1KHz- 1MHz at room temperature (25°C) have been done to get the understanding of electrical conduction and dielectric relaxation and their correlation. Dielectric response of the samples at lower frequencies shows dielectric dispersion while at higher frequencies it shows dielectric relaxation. The ac conductivity is well fitted by the Jonscher law. The spectroscopic data in the impedance plane confirms the existence of grain contribution to the relaxation. All the properties are found out to be function of frequency as well as the amount of substitution.Keywords: Dielectric ceramics, Dielectric constant, Loss tangent, AC conductivity, Impedance spectroscopy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 252130 Cultural Effect on Using New Technologies
Authors: Nazli Ebrahimi, Sharan Kaur Garib Singh, Reza Sigari Tabrizi
Abstract:
One of the main concerns in the Information Technology field is adoption with new technologies in organizations which may result in increasing the usage paste of these technologies.This study aims to look at the issue of culture-s role in accepting and using new technologies in organizations. The study examines the effect of culture on accepting and intention to use new technology in organizations. Studies show culture is one of the most important barriers in adoption new technologies. The model used for accepting and using new technology is Technology Acceptance Model (TAM), while for culture and dimensions a well-known theory by Hofsted was used. Results of the study show significant effect of culture on intention to use new technologies. All four dimensions of culture were tested to find the strength of relationship with behavioral intention to use new technologies. Findings indicate the important role of culture in the level of intention to use new technologies and different role of each dimension to improve adaptation process. The study suggests that transferring of new technologies efforts are most likely to be successful if the parties are culturally aligned.
Keywords: Human-computer interaction, accepting new technologies, culture, behavioral intention.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 295929 A Study of Gaps in CBMIR Using Different Methods and Prospective
Authors: Pradeep Singh, Sukhwinder Singh, Gurjinder Kaur
Abstract:
In recent years, rapid advances in software and hardware in the field of information technology along with a digital imaging revolution in the medical domain facilitate the generation and storage of large collections of images by hospitals and clinics. To search these large image collections effectively and efficiently poses significant technical challenges, and it raises the necessity of constructing intelligent retrieval systems. Content-based Image Retrieval (CBIR) consists of retrieving the most visually similar images to a given query image from a database of images[5]. Medical CBIR (content-based image retrieval) applications pose unique challenges but at the same time offer many new opportunities. On one hand, while one can easily understand news or sports videos, a medical image is often completely incomprehensible to untrained eyes.
Keywords: Classification, clustering, content-based image retrieval (CBIR), relevance feedback (RF), statistical similarity matching, support vector machine (SVM).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 178728 Heavy Metals and Polycyclic Aromatic Hydrocarbons in Roadside Soil Samples: A Review
Authors: R. Kaur, J. K. Katnoria
Abstract:
Diverse contaminants released into the environment through progress of urbanization and industrialization adversely affect human health. Among various sources of contaminants, especially, in big cities, automobiles play a significant role in aggravating the pollution. Various pollutants viz., heavy metals (Pb, Mn, Ni, Zn, As, Hg, Cd) and Polyaromatic hydrocarbons (Benzo-a-pyrene, fluoranthene, pyrene, benzo-b-anthracene, benzo-b-fluoranthene, acenaphthylene, fluorine, phenantherene, anthracene, chrysene, benzo-k-fluoranthene, benzo-e-pyrene, indenol-1,2,3-cd-pyrene, dibenzo-a,h-anthracene, benzo-ghi-perylene) are released by vehicles. Further, these pollutants are expected to cause severe mutagenic, genotoxic and carcinogenic effects. Considering this, many authors monitored the levels of pollution in roadside soil, water and plants. The present review focuses upon the analysis and effects of heavy metals and polycyclic aromatic hydrocarbons from the roadside samples.
Keywords: Automobiles, Carcinogenicity, Atomic Absorption Spectrophotometer, Gas Chromatography – Mass Spectroscopy, Soil pollution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 444027 Modeling and Analysis of an SIRS Epidemic Model with Effect of Awareness Programs by Media
Authors: Navjot Kaur, Mini Ghosh, S.S. Bhatia
Abstract:
This paper proposes and analyzes an SIRS epidemic model incorporating the effects of the awareness programs driven by the media. Media and media driven awareness programs play a promising role in disseminating the information about outbreak of any disease across the globe. This motivates people to take precautionary measures and guides the infected individuals to get hospitalized. Timely hospitalization helps to reduce diagnostic delays and hence results in fast recovery of infected individuals. The aim of this study is to investigate the impact of the media on the spread and control of infectious diseases. This model is analyzed using stability theory of differential equations. The sensitivity of parameters has been discussed and it has been found that the awareness programs driven by the media have positive impact in reducing the infection prevalence of the infective population in the region under consideration.
Keywords: Infectious diseases, SIRS model, Media, Stability theory, Simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 291326 A Novel Algorithm for Parsing IFC Models
Authors: Raninder Kaur Dhillon, Mayur Jethwa, Hardeep Singh Rai
Abstract:
Information technology has made a pivotal progress across disparate disciplines, one of which is AEC (Architecture, Engineering and Construction) industry. CAD is a form of computer-aided building modulation that architects, engineers and contractors use to create and view two- and three-dimensional models. The AEC industry also uses building information modeling (BIM), a newer computerized modeling system that can create four-dimensional models; this software can greatly increase productivity in the AEC industry. BIM models generate open source IFC (Industry Foundation Classes) files which aim for interoperability for exchanging information throughout the project lifecycle among various disciplines. The methods developed in previous studies require either an IFC schema or MVD and software applications, such as an IFC model server or a Building Information Modeling (BIM) authoring tool, to extract a partial or complete IFC instance model. This paper proposes an efficient algorithm for extracting a partial and total model from an Industry Foundation Classes (IFC) instance model without an IFC schema or a complete IFC model view definition (MVD).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 270425 A Study of Agile-Based Approaches to Improve Software Quality
Authors: Gurmeet Kaur, Jyoti Pruthi
Abstract:
Agile Software development approaches and techniques are being considered as efficient, effective, and popular methods to the development of software. Agile software developments are useful for developing high-quality software that completes client requirements with zero defects, and in short delivery period. In agile software development methodology, quality is related to coding, which means quality, is managed through the use of approaches like refactoring, pair programming, test-driven development, behavior-driven development, acceptance test-driven development, and demand-driven development. The quality of software is measured using metrics like the number of defects during the development and improvement of the software. Usage of the above-mentioned methods or approaches reduces the possibilities of defects in developed software, and hence improves quality. This paper focuses on the study of agile-based quality methods or approaches for software development that ensures improved quality of software as well as reduced cost, and customer satisfaction.
Keywords: Agile software development, ASD, Acceptance test-driven development, ATDD, Behavior-driven development, BDD, Demand-driven development. DDD, Test-driven development, TDD.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 69124 Fingerprint Verification System Using Minutiae Extraction Technique
Authors: Manvjeet Kaur, Mukhwinder Singh, Akshay Girdhar, Parvinder S. Sandhu
Abstract:
Most fingerprint recognition techniques are based on minutiae matching and have been well studied. However, this technology still suffers from problems associated with the handling of poor quality impressions. One problem besetting fingerprint matching is distortion. Distortion changes both geometric position and orientation, and leads to difficulties in establishing a match among multiple impressions acquired from the same finger tip. Marking all the minutiae accurately as well as rejecting false minutiae is another issue still under research. Our work has combined many methods to build a minutia extractor and a minutia matcher. The combination of multiple methods comes from a wide investigation into research papers. Also some novel changes like segmentation using Morphological operations, improved thinning, false minutiae removal methods, minutia marking with special considering the triple branch counting, minutia unification by decomposing a branch into three terminations, and matching in the unified x-y coordinate system after a two-step transformation are used in the work.
Keywords: Biometrics, Minutiae, Crossing number, False Accept Rate (FAR), False Reject Rate (FRR).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 657423 A Subtractive Clustering Based Approach for Early Prediction of Fault Proneness in Software Modules
Authors: Ramandeep S. Sidhu, Sunil Khullar, Parvinder S. Sandhu, R. P. S. Bedi, Kiranbir Kaur
Abstract:
In this paper, subtractive clustering based fuzzy inference system approach is used for early detection of faults in the function oriented software systems. This approach has been tested with real time defect datasets of NASA software projects named as PC1 and CM1. Both the code based model and joined model (combination of the requirement and code based metrics) of the datasets are used for training and testing of the proposed approach. The performance of the models is recorded in terms of Accuracy, MAE and RMSE values. The performance of the proposed approach is better in case of Joined Model. As evidenced from the results obtained it can be concluded that Clustering and fuzzy logic together provide a simple yet powerful means to model the earlier detection of faults in the function oriented software systems.
Keywords: Subtractive clustering, fuzzy inference system, fault proneness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 258022 Design and Implementation of Reed Solomon Encoder on FPGA
Authors: Amandeep Singh, Mandeep Kaur
Abstract:
Error correcting codes are used for detection and correction of errors in digital communication system. Error correcting coding is based on appending of redundancy to the information message according to a prescribed algorithm. Reed Solomon codes are part of channel coding and withstand the effect of noise, interference and fading. Galois field arithmetic is used for encoding and decoding reed Solomon codes. Galois field multipliers and linear feedback shift registers are used for encoding the information data block. The design of Reed Solomon encoder is complex because of use of LFSR and Galois field arithmetic. The purpose of this paper is to design and implement Reed Solomon (255, 239) encoder with optimized and lesser number of Galois Field multipliers. Symmetric generator polynomial is used to reduce the number of GF multipliers. To increase the capability toward error correction, convolution interleaving will be used with RS encoder. The Design will be implemented on Xilinx FPGA Spartan II.
Keywords: Galois Field, Generator polynomial, LFSR, Reed Solomon.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 484421 A Second Look at Gesture-Based Passwords: Usability and Vulnerability to Shoulder-Surfing Attacks
Authors: Lakshmidevi Sreeramareddy, Komalpreet Kaur, Nane Pothier
Abstract:
For security purposes, it is important to detect passwords entered by unauthorized users. With traditional alphanumeric passwords, if the content of a password is acquired and correctly entered by an intruder, it is impossible to differentiate the password entered by the intruder from those entered by the authorized user because the password entries contain precisely the same character set. However, no two entries for the gesture-based passwords, even those entered by the person who created the password, will be identical. There are always variations between entries, such as the shape and length of each stroke, the location of each stroke, and the speed of drawing. It is possible that passwords entered by the unauthorized user contain higher levels of variations when compared with those entered by the authorized user (the creator). The difference in the levels of variations may provide cues to detect unauthorized entries. To test this hypothesis, we designed an empirical study, collected and analyzed the data with the help of machine-learning algorithms. The results of the study are significant.
Keywords: Authentication, gesture-based passwords, machine learning algorithms, shoulder-surfing attacks, usability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 61620 A Study on Early Prediction of Fault Proneness in Software Modules using Genetic Algorithm
Authors: Parvinder S. Sandhu, Sunil Khullar, Satpreet Singh, Simranjit K. Bains, Manpreet Kaur, Gurvinder Singh
Abstract:
Fault-proneness of a software module is the probability that the module contains faults. To predict faultproneness of modules different techniques have been proposed which includes statistical methods, machine learning techniques, neural network techniques and clustering techniques. The aim of proposed study is to explore whether metrics available in the early lifecycle (i.e. requirement metrics), metrics available in the late lifecycle (i.e. code metrics) and metrics available in the early lifecycle (i.e. requirement metrics) combined with metrics available in the late lifecycle (i.e. code metrics) can be used to identify fault prone modules using Genetic Algorithm technique. This approach has been tested with real time defect C Programming language datasets of NASA software projects. The results show that the fusion of requirement and code metric is the best prediction model for detecting the faults as compared with commonly used code based model.Keywords: Genetic Algorithm, Fault Proneness, Software Faultand Software Quality.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 198419 Image Haze Removal Using Scene Depth Based Spatially Varying Atmospheric Light in Haar Lifting Wavelet Domain
Authors: Prabh Preet Singh, Harpreet Kaur
Abstract:
This paper presents a method for single image dehazing based on dark channel prior (DCP). The property that the intensity of the dark channel gives an approximate thickness of the haze is used to estimate the transmission and atmospheric light. Instead of constant atmospheric light, the proposed method employs scene depth to estimate spatially varying atmospheric light as it truly occurs in nature. Haze imaging model together with the soft matting method has been used in this work to produce high quality haze free image. Experimental results demonstrate that the proposed approach produces better results than the classic DCP approach as color fidelity and contrast of haze free image are improved and no over-saturation in the sky region is observed. Further, lifting Haar wavelet transform is employed to reduce overall execution time by a factor of two to three as compared to the conventional approach.
Keywords: Depth based atmospheric light, dark channel prior, lifting wavelet.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 55318 Production of Eicosapentaenoic Acid and Fucoxanthin in Cold-Tolerant Diatom Strains
Authors: Nga Phuong Dang, Terje Vasskog, Ashwiny Pandey, Rajnish Kaur Calay
Abstract:
Diatoms hold great potential for nutraceutical production as they are source of high value products such as eicosapentaenoic acid (EPA) and pigment fucoxanthin. EPA has proven human health benefits, and fucoxanthin can be used as both medicinal and nutritional ingredient to prevent and treat chronic diseases. The economic perspective of commercial production of a single product from microalgae is not attractive due to the high production cost. To improve the economic viability, we explore the concept of combining the production of both EPA and fucoxanthin in a single process. In our current study, we isolated twelve new microalgae isolates from Ofotfjord. Eight of them are diatoms and 4 of them are cyanobacteria and microalgae. Screening the new diatom isolates revealed that two strains are cold-tolerant diatom which can grow at 10 °C. They accumulated significant amount of lipid, which was up to 40-60% of the dried mass. The EPA contents from the two strains ranged from 15-19% of the total fatty acid, while fucoxanthin concentrations were between 1-1.4% of the dried biomass. Comparing with other studied diatom, this is promising result. The two strains hold promise as source for EPA and fucoxanthin production.
Keywords: Microalgae, fucoxanthin, eicosapentaenoic acid, diatom, fatty acid.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 53617 Image Dehazing Using Dark Channel Prior and Fast Guided Filter in Daubechies Lifting Wavelet Transform Domain
Authors: Harpreet Kaur, Sudipta Majumdar
Abstract:
In this paper a method for image dehazing is proposed in lifting wavelet transform domain. Lifting Daubechies (D4) wavelet has been used to obtain the approximate image and detail images. As the haze is contained in low frequency part, only the approximate image is used for further processing. This region is processed by dehazing algorithm based on dark channel prior (DCP). The dehazed approximate image is then recombined with the detail images using inverse lifting wavelet transform. Implementation of lifting wavelet transform has the advantage of auxiliary memory saving, fast implementation and simplicity. Also, the proposed method deals with near white scene problem, blue horizon issue and localized light sources in a way to enhance image quality and makes the algorithm robust. Simulation results present improvement in terms of visual quality, parameters such as root mean square (RMS) contrast, structural similarity index (SSIM), entropy and execution time.
Keywords: Dark channel prior, image dehazing, lifting wavelet transform.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 112316 Numerical Solution of a Laminar Viscous Flow Boundary Layer Equation Using Uniform Haar Wavelet Quasi-linearization Method
Authors: Harpreet Kaur, Vinod Mishra, R. C. Mittal
Abstract:
In this paper, we have proposed a Haar wavelet quasilinearization method to solve the well known Blasius equation. The method is based on the uniform Haar wavelet operational matrix defined over the interval [0, 1]. In this method, we have proposed the transformation for converting the problem on a fixed computational domain. The Blasius equation arises in the various boundary layer problems of hydrodynamics and in fluid mechanics of laminar viscous flows. Quasi-linearization is iterative process but our proposed technique gives excellent numerical results with quasilinearization for solving nonlinear differential equations without any iteration on selecting collocation points by Haar wavelets. We have solved Blasius equation for 1≤α ≤ 2 and the numerical results are compared with the available results in literature. Finally, we conclude that proposed method is a promising tool for solving the well known nonlinear Blasius equation.
Keywords: Boundary layer Blasius equation, collocation points, quasi-linearization process, uniform haar wavelets.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 327215 An Evaluation of Pesticide Stress Induced Proteins in three Cyanobacterial Species-Anabaena Fertilissima, Aulosira Fertilissima and Westiellopsis Prolifica using SDS-PAGE
Authors: Nirmal Kumar, Rita N. Kumar, Anubhuti Bora, Manmeet Kaur Amb
Abstract:
The whole-cell protein-profiling technique was evaluated for studying differences in banding pattern of three different species of Cyanobacteria i.e. Anabaena fertilissima, Aulosira fertilissima and Westiellopsis prolifica under the influence of four different pesticides-2,4-D (Ethyl Ester of 2,4-Dichloro Phenoxy Acetic Acid), Pencycuron (N-[(4-chlorophenyl)methyl]-Ncyclopentyl- N'–phenylurea), Endosulfan (6,7,8,9,10,10hexachloro- 1,5,5a,6,9,9a-hexahydro-6,9-methano-2,4,3-benzodioxathiepine-3- oxide) and Tebuconazole (1-(4-Chlorophenyl)-4,4-dimethyl-3-(1,2,4- triazol-1-ylmethyl)pentan-3-ol). Whole-cell extracts were obtained by sonication treatment (Sonifier cell disruptor -Branson Digital Sonifier S-450D, USA) and were analyzed by sodium dodecyl sulfate polyacrylamide gel electrophoresis (SDS-PAGE). SDS-PAGE analyses of the total protein profile of Anabaena fertilissima, Aulosira fertilissima and Westiellopsis prolifica showed a linear decrease in the protein content with increasing pesticide stress when administered to different concentrations of 2, 4-D, Pencycuron, Endosulfan and Tebuconazole. The results indicate that different stressors exert specific effects on cyanobacterial protein synthesis.Keywords: Cyanobacteria, pesticide, SDS-PAGE
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 250914 Cluster Algorithm for Genetic Diversity
Authors: Manpreet Singh, Keerat Kaur, Bhavdeep Singh
Abstract:
With the hardware technology advancing, the cost of storing is decreasing. Thus there is an urgent need for new techniques and tools that can intelligently and automatically assist us in transferring this data into useful knowledge. Different techniques of data mining are developed which are helpful for handling these large size databases [7]. Data mining is also finding its role in the field of biotechnology. Pedigree means the associated ancestry of a crop variety. Genetic diversity is the variation in the genetic composition of individuals within or among species. Genetic diversity depends upon the pedigree information of the varieties. Parents at lower hierarchic levels have more weightage for predicting genetic diversity as compared to the upper hierarchic levels. The weightage decreases as the level increases. For crossbreeding, the two varieties should be more and more genetically diverse so as to incorporate the useful characters of the two varieties in the newly developed variety. This paper discusses the searching and analyzing of different possible pairs of varieties selected on the basis of morphological characters, Climatic conditions and Nutrients so as to obtain the most optimal pair that can produce the required crossbreed variety. An algorithm was developed to determine the genetic diversity between the selected wheat varieties. Cluster analysis technique is used for retrieving the results.Keywords: Genetic diversity, pedigree, nutrients.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 180213 Actionable Rules: Issues and New Directions
Authors: Harleen Kaur
Abstract:
Knowledge Discovery in Databases (KDD) is the process of extracting previously unknown, hidden and interesting patterns from a huge amount of data stored in databases. Data mining is a stage of the KDD process that aims at selecting and applying a particular data mining algorithm to extract an interesting and useful knowledge. It is highly expected that data mining methods will find interesting patterns according to some measures, from databases. It is of vital importance to define good measures of interestingness that would allow the system to discover only the useful patterns. Measures of interestingness are divided into objective and subjective measures. Objective measures are those that depend only on the structure of a pattern and which can be quantified by using statistical methods. While, subjective measures depend only on the subjectivity and understandability of the user who examine the patterns. These subjective measures are further divided into actionable, unexpected and novel. The key issues that faces data mining community is how to make actions on the basis of discovered knowledge. For a pattern to be actionable, the user subjectivity is captured by providing his/her background knowledge about domain. Here, we consider the actionability of the discovered knowledge as a measure of interestingness and raise important issues which need to be addressed to discover actionable knowledge.
Keywords: Data Mining Community, Knowledge Discovery inDatabases (KDD), Interestingness, Subjective Measures, Actionability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 194212 PSO Based Weight Selection and Fixed Structure Robust Loop Shaping Control for Pneumatic Servo System with 2DOF Controller
Authors: Randeep Kaur, Jyoti Ohri
Abstract:
This paper proposes a new technique to design a fixed-structure robust loop shaping controller for the pneumatic servosystem. In this paper, a new method based on a particle swarm optimization (PSO) algorithm for tuning the weighting function parameters to design an H∞ controller is presented. The PSO algorithm is used to minimize the infinity norm of the transfer function of the nominal closed loop system to obtain the optimal parameters of the weighting functions. The optimal stability margin is used as an objective in PSO for selecting the optimal weighting parameters; it is shown that the proposed method can simplify the design procedure of H∞ control to obtain optimal robust controller for pneumatic servosystem. In addition, the order of the proposed controller is much lower than that of the conventional robust loop shaping controller, making it easy to implement in practical works. Also two-degree-of-freedom (2DOF) control design procedure is proposed to improve tracking performance in the face of noise and disturbance. Result of simulations demonstrates the advantages of the proposed controller in terms of simple structure and robustness against plant perturbations and disturbances.
Keywords: Robust control, Pneumatic Servosystem, PSO, H∞ control, 2DOF.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2424