Search results for: Step pulsed active thermography (SPAT)
667 Novel Anti-leukemia Calanone Compounds by Quantitative Structure-Activity Relationship AM1 Semiempirical Method
Authors: Ponco Iswanto, Mochammad Chasani, Muhammad Hanafi, Iqmal Tahir, Eva Vaulina YD, Harjono, Lestari Solikhati, Winkanda S. Putra, Yayuk Yuliantini
Abstract:
Quantitative Structure-Activity Relationship (QSAR) approach for discovering novel more active Calanone derivative as anti-leukemia compound has been conducted. There are 6 experimental activities of Calanone compounds against leukemia cell L1210 that are used as material of the research. Calculation of theoretical predictors (independent variables) was performed by AM1 semiempirical method. The QSAR equation is determined by Principle Component Regression (PCR) analysis, with Log IC50 as dependent variable and the independent variables are atomic net charges, dipole moment (μ), and coefficient partition of noctanol/ water (Log P). Three novel Calanone derivatives that obtained by this research have higher activity against leukemia cell L1210 than pure Calanone.Keywords: AM1 semiempirical calculation, Calanone, Principle Component Regression, QSAR approach.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1476666 Reduction of Linear Time-Invariant Systems Using Routh-Approximation and PSO
Authors: S. Panda, S. K. Tomar, R. Prasad, C. Ardil
Abstract:
Order reduction of linear-time invariant systems employing two methods; one using the advantages of Routh approximation and other by an evolutionary technique is presented in this paper. In Routh approximation method the denominator of the reduced order model is obtained using Routh approximation while the numerator of the reduced order model is determined using the indirect approach of retaining the time moments and/or Markov parameters of original system. By this method the reduced order model guarantees stability if the original high order model is stable. In the second method Particle Swarm Optimization (PSO) is employed to reduce the higher order model. PSO method is based on the minimization of the Integral Squared Error (ISE) between the transient responses of original higher order model and the reduced order model pertaining to a unit step input. Both the methods are illustrated through numerical examples.
Keywords: Model Order Reduction, Markov Parameters, Routh Approximation, Particle Swarm Optimization, Integral Squared Error, Steady State Stability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3287665 Potential cIBR-Conjugated PLGA Nanoparticles for Selective Targeting to Leukemic Cells
Authors: Rungsinee Phongpradist, Sawitree Chiampanichayakul, Singkome Tima, Teruna J. Siahaan, Cory J. Berkland, Songyot Anuchapreeda, Chadarat Ampasavate
Abstract:
The expression of LFA-1 diverges from the physiological condition, thus active targeting carrier can provide the benefits from difference into LFA-1 expression in various conditions. Here, the selectivity of cIBR-conjugated nanoparticles (cIBR-NPs), in terms of uptake, was investigated using PBMCs, Mixed PBMCMolt- 3 cells and Molt-3 cells. The expressions of LFA-1 on Molt-3 cells, from flow cytometry and Western blot, possessed the highest level whereas PBMCs showed the lowest level. The kinetic uptake profiles of cIBR-NPs were obtained by flow cytometry, which the degree of cellular uptake presented a similar trend with the level of LFA-1 indicating the influence of LFA-1 expression on the cellular uptake of cIBR-NPs. The conformation of LFA-1 had a slight effect on the cellular uptake of cIBR-NPs. Overall we demonstrated that cIBR-NPs enhanced cellular uptake and improved the selectivity of drug carriers to LFA-1 on the leukemia cells, which related with the order of LFA-1 expression.Keywords: cIBR, LFA-1, Molt-3, PBMCs
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1937664 Segmentation of Lungs from CT Scan Images for Early Diagnosis of Lung Cancer
Authors: Nisar Ahmed Memon, Anwar Majid Mirza, S.A.M. Gilani
Abstract:
Segmentation is an important step in medical image analysis and classification for radiological evaluation or computer aided diagnosis. The CAD (Computer Aided Diagnosis ) of lung CT generally first segment the area of interest (lung) and then analyze the separately obtained area for nodule detection in order to diagnosis the disease. For normal lung, segmentation can be performed by making use of excellent contrast between air and surrounding tissues. However this approach fails when lung is affected by high density pathology. Dense pathologies are present in approximately a fifth of clinical scans, and for computer analysis such as detection and quantification of abnormal areas it is vital that the entire and perfectly lung part of the image is provided and no part, as present in the original image be eradicated. In this paper we have proposed a lung segmentation technique which accurately segment the lung parenchyma from lung CT Scan images. The algorithm was tested against the 25 datasets of different patients received from Ackron Univeristy, USA and AGA Khan Medical University, Karachi, Pakistan.Keywords: Computer Aided Diagnosis, Medical ImageProcessing, Region Growing, Segmentation, Thresholding,
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2597663 Specialized Reduced Models of Dynamic Flows in 2-Stroke Engines
Authors: S. Cagin, X. Fischer, E. Delacourt, N. Bourabaa, C. Morin, D. Coutellier, B. Carré, S. Loumé
Abstract:
The complexity of scavenging by ports and its impact on engine efficiency create the need to understand and to model it as realistically as possible. However, there are few empirical scavenging models and these are highly specialized. In a design optimization process, they appear very restricted and their field of use is limited. This paper presents a comparison of two methods to establish and reduce a model of the scavenging process in 2-stroke diesel engines. To solve the lack of scavenging models, a CFD model has been developed and is used as the referent case. However, its large size requires a reduction. Two techniques have been tested depending on their fields of application: The NTF method and neural networks. They both appear highly appropriate drastically reducing the model’s size (over 90% reduction) with a low relative error rate (under 10%). Furthermore, each method produces a reduced model which can be used in distinct specialized fields of application: the distribution of a quantity (mass fraction for example) in the cylinder at each time step (pseudo-dynamic model) or the qualification of scavenging at the end of the process (pseudo-static model).
Keywords: Diesel engine, Design optimization, Model reduction, Neural network, NTF algorithm, Scavenging.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1329662 Segmentation Problems and Solutions in Printed Degraded Gurmukhi Script
Authors: M. K. Jindal, G. S. Lehal, R. K. Sharma
Abstract:
Character segmentation is an important preprocessing step for text recognition. In degraded documents, existence of touching characters decreases recognition rate drastically, for any optical character recognition (OCR) system. In this paper we have proposed a complete solution for segmenting touching characters in all the three zones of printed Gurmukhi script. A study of touching Gurmukhi characters is carried out and these characters have been divided into various categories after a careful analysis. Structural properties of the Gurmukhi characters are used for defining the categories. New algorithms have been proposed to segment the touching characters in middle zone, upper zone and lower zone. These algorithms have shown a reasonable improvement in segmenting the touching characters in degraded printed Gurmukhi script. The algorithms proposed in this paper are applicable only to machine printed text. We have also discussed a new and useful technique to segment the horizontally overlapping lines.Keywords: Character Segmentation, Middle Zone, Upper Zone, Lower Zone, Touching Characters, Horizontally Overlapping Lines.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1695661 CMOS-Compatible Silicon Nanoplasmonics for On-Chip Integration
Authors: Shiyang Zhu, Guo-Qiang Lo, Dim-Lee Kwong
Abstract:
Although silicon photonic devices provide a significantly larger bandwidth and dissipate a substantially less power than the electronic devices, they suffer from a large size due to the fundamental diffraction limit and the weak optical response of Si. A potential solution is to exploit Si plasmonics, which may not only miniaturize the photonic device far beyond the diffraction limit, but also enhance the optical response in Si due to the electromagnetic field confinement. In this paper, we discuss and summarize the recently developed metal-insulator-Si-insulator-metal nanoplasmonic waveguide as well as various passive and active plasmonic components based on this waveguide, including coupler, bend, power splitter, ring resonator, MZI, modulator, detector, etc. All these plasmonic components are CMOS compatible and could be integrated with electronic and conventional dielectric photonic devices on the same SOI chip. More potential plasmonic devices as well as plasmonic nanocircuits with complex functionalities are also addressed.
Keywords: Silicon nanoplasmonics, Silicon nanophotonics, Onchip integration, CMOS
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1905660 Faster FPGA Routing Solution using DNA Computing
Authors: Manpreet Singh, Parvinder Singh Sandhu, Manjinder Singh Kahlon
Abstract:
There are many classical algorithms for finding routing in FPGA. But Using DNA computing we can solve the routes efficiently and fast. The run time complexity of DNA algorithms is much less than other classical algorithms which are used for solving routing in FPGA. The research in DNA computing is in a primary level. High information density of DNA molecules and massive parallelism involved in the DNA reactions make DNA computing a powerful tool. It has been proved by many research accomplishments that any procedure that can be programmed in a silicon computer can be realized as a DNA computing procedure. In this paper we have proposed two tier approaches for the FPGA routing solution. First, geometric FPGA detailed routing task is solved by transforming it into a Boolean satisfiability equation with the property that any assignment of input variables that satisfies the equation specifies a valid routing. Satisfying assignment for particular route will result in a valid routing and absence of a satisfying assignment implies that the layout is un-routable. In second step, DNA search algorithm is applied on this Boolean equation for solving routing alternatives utilizing the properties of DNA computation. The simulated results are satisfactory and give the indication of applicability of DNA computing for solving the FPGA Routing problem.Keywords: FPGA, Routing, DNA Computing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1591659 A Numerical Description of a Fibre Reinforced Concrete Using a Genetic Algorithm
Authors: Henrik L. Funke, Lars Ulke-Winter, Sandra Gelbrich, Lothar Kroll
Abstract:
This work reports about an approach for an automatic adaptation of concrete formulations based on genetic algorithms (GA) to optimize a wide range of different fit-functions. In order to achieve the goal, a method was developed which provides a numerical description of a fibre reinforced concrete (FRC) mixture regarding the production technology and the property spectrum of the concrete. In a first step, the FRC mixture with seven fixed components was characterized by varying amounts of the components. For that purpose, ten concrete mixtures were prepared and tested. The testing procedure comprised flow spread, compressive and bending tensile strength. The analysis and approximation of the determined data was carried out by GAs. The aim was to obtain a closed mathematical expression which best describes the given seven-point cloud of FRC by applying a Gene Expression Programming with Free Coefficients (GEP-FC) strategy. The seven-parametric FRC-mixtures model which is generated according to this method correlated well with the measured data. The developed procedure can be used for concrete mixtures finding closed mathematical expressions, which are based on the measured data.
Keywords: Concrete design, fibre reinforced concrete, genetic algorithms, GEP-FC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 988658 The Impact of Germination and In Vitro Digestion on the Formation of Angiotensin Converting Enzyme (ACE) Inhibitory Peptides from Lentil Proteins Compared to Whey Proteins
Authors: F. Bamdad, Sh. Dokhani, J. Keramat, R. Zareie
Abstract:
Biologically active peptides are of particular interest in food science and human nutrition because they have been shown to play several physiological roles. In vitro gastrointestinal digestion of lentil and whey proteins in this study produced high angiotensin-I converting enzyme inhibitory activity with 75.5±1.9 and 91.4±2.3% inhibition, respectively. High ACE inhibitory activity was observed in lentil after 5 days of germination (84.3±1.2%). Fractionation by reverse phase chromatography gave inhibitory activities as high as 86.3±2.0 for lentil, 94.8±1.8% for whey and 93.7±1.7% at 5th day of germination. Further purification by HPLC resulted in several inhibitory peptides with IC50 values ranging from 0.064 to 0.164 mg/ml. These results demonstrate that lentil proteins are a good source of peptides with ACE inhibitory activity that can be released by germination or gastrointestinal digestion. Despite the lower bioactivity in comparison with whey proteins, incorporation of lentil proteins in functional food formulations and natural drugs look promising.Keywords: ACE inhibitory peptides, digestion, germination, lentil proteins, whey proteins
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2448657 Determining of Threshold Levels of Burst by Burst AQAM/CDMA in Slow Rayleigh Fading Environments
Authors: F. Nejadebrahimi, M. ArdebiliPour
Abstract:
In this paper, we are going to determine the threshold levels of adaptive modulation in a burst by burst CDMA system by a suboptimum method so that the above method attempts to increase the average bit per symbol (BPS) rate of transceiver system by switching between the different modulation modes in variable channel condition. In this method, we choose the minimum values of average bit error rate (BER) and maximum values of average BPS on different values of average channel signal to noise ratio (SNR) and then calculate the relative threshold levels of them, so that when the instantaneous SNR increases, a higher order modulation be employed for increasing throughput and vise-versa when the instantaneous SNR decreases, a lower order modulation be employed for improvement of BER. In transmission step, by this adaptive modulation method, in according to comparison between obtained estimation of pilot symbols and a set of above suboptimum threshold levels, above system chooses one of states no transmission, BPSK, 4QAM and square 16QAM for modulation of data. The expected channel in this paper is a slow Rayleigh fading.
Keywords: AQAM, burst, BER, BPS, CDMA, threshold.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1532656 A Two Level Load Balancing Approach for Cloud Environment
Authors: Anurag Jain, Rajneesh Kumar
Abstract:
Cloud computing is the outcome of rapid growth of internet. Due to elastic nature of cloud computing and unpredictable behavior of user, load balancing is the major issue in cloud computing paradigm. An efficient load balancing technique can improve the performance in terms of efficient resource utilization and higher customer satisfaction. Load balancing can be implemented through task scheduling, resource allocation and task migration. Various parameters to analyze the performance of load balancing approach are response time, cost, data processing time and throughput. This paper demonstrates a two level load balancer approach by combining join idle queue and join shortest queue approach. Authors have used cloud analyst simulator to test proposed two level load balancer approach. The results are analyzed and compared with the existing algorithms and as observed, proposed work is one step ahead of existing techniques.
Keywords: Cloud Analyst, Cloud Computing, Join Idle Queue, Join Shortest Queue, Load balancing, Task Scheduling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1991655 Evaluation of Minimization of Moment Ratio Method by Physical Modeling
Authors: Amin Eslami, Jafar Bolouri Bazaz
Abstract:
Under active stress conditions, a rigid cantilever retaining wall tends to rotate about a pivot point located within the embedded depth of the wall. For purely granular and cohesive soils, a methodology was previously reported called minimization of moment ratio to determine the location of the pivot point of rotation. The usage of this new methodology is to estimate the rotational stability safety factor. Moreover, the degree of improvement required in a backfill to get a desired safety factor can be estimated by the concept of the shear strength demand. In this article, the accuracy of this method for another type of cantilever walls called Contiguous Bored Pile (CBP) retaining wall is evaluated by using physical modeling technique. Based on observations, the results of moment ratio minimization method are in good agreement with the results of the carried out physical modeling.Keywords: Cantilever Retaining Wall, Physical Modeling, Minimization of Moment Ratio Method, Pivot Point.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1581654 Effect of Curing Profile to Eliminate the Voids / Black Dots Formation in Underfill Epoxy for Hi-CTE Flip Chip Packaging
Authors: Zainudin Kornain, Azman Jalar, Rozaidi Rasid, Fong Chee Seng
Abstract:
Void formation in underfill is considered as failure in flip chip manufacturing process. Void formation possibly caused by several factors such as poor soldering and flux residue during die attach process, void entrapment due moisture contamination, dispense pattern process and setting up the curing process. This paper presents the comparison of single step and two steps curing profile towards the void and black dots formation in underfill for Hi-CTE Flip Chip Ceramic Ball Grid Array Package (FC-CBGA). Statistic analysis was conducted to analyze how different factors such as wafer lot, sawing technique, underfill fillet height and curing profile recipe were affected the formation of voids and black dots. A C-Mode Scanning Aqoustic Microscopy (C-SAM) was used to scan the total count of voids and black dots. It was shown that the 2 steps curing profile provided solution for void elimination and black dots in underfill after curing process.Keywords: black dots formation, curing profile, FC-CBGA, underfill, void formation,
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4071653 Inflammatory Markers in the Blood and Chronic Periodontitis
Authors: Saimir Heta, Ilma Robo, Nevila Alliu, Tea Meta
Abstract:
Background: Plasma levels of inflammatory markers are the expression of the infectious wastes of existing periodontitis, as well as of existing inflammation everywhere in the body. Materials and Methods: The study consists of the clinical part of the measurement of inflammatory markers of 23 patients diagnosed with chronic periodontitis and the recording of parental periodontal parameters of patient periodontal status: hemorrhage index and probe values, before and 7-10 days after non-surgical periodontal treatment. Results: The level of fibrinogen drops according to the categorization of disease progression, active and passive, with the biggest % (18%-30%) at the fluctuation 10-20 mg/d. Fluctuations in fibrinogen level according to the age of patients in the range 0-10 mg/dL under 40 years and over 40 years was 13%-26%, in the range 10-20 mg/dL was 26%-22%, in the 20-40 mg/dL was 9%-4%. Conclusions: Non-surgical periodontal treatment significantly reduces the level of non-inflammatory markers in the blood. Oral health significantly reduces the potential source for periodontal bacteria, with the potential of promoting thromboembolism, through interaction between thrombocytes.
Keywords: Chronic periodontitis, atherosclerosis, risk factor, inflammatory markers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 562652 The State, Local Community and Participatory Governance Practices: Prospects of Change
Authors: Gaysu R. Arvind
Abstract:
In policy discourse of 1990s, more inclusive spaces have been constructed for realizing full and meaningful participation of common people in education. These participatory spaces provide an alternative possibility for universalizing elementary education against the backdrop of a history of entrenched forms of social and economical exclusion; inequitable education provisions; and shrinking role of the state in today-s neo-liberal times. Drawing on case-studies of bottom-up approaches to school governance, the study examines an array of innovative ways through which poor people gained a sense of identity and agency by evolving indigenous solutions to issues regarding schooling of their children. In the process, state-s institutions and practices became more accountable and responsive to educational concerns of the marginalized people. The deliberative participation emerged as an active way of experiencing deeper forms of empowerment and democracy than its passive realization as mere bearers of citizen rights.Keywords: Deliberative Forum, Inclusive Spaces, Participatory Governance, People's Agency
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1656651 Project Management at University: Towards an Evaluation Process around Cooperative Learning
Authors: J. L. Andrade-Pineda, J.M. León-Blanco, M. Calle, P. L. González-R
Abstract:
The enrollment in current Master's degree programs usually pursues gaining the expertise required in real-life workplaces. The experience we present here concerns the learning process of "Project Management Methodology (PMM)", around a cooperative/collaborative mechanism aimed at affording students measurable learning goals and providing the teacher with the ability of focusing on the weaknesses detected. We have designed a mixed summative/formative evaluation, which assures curriculum engage while enriches the comprehension of PMM key concepts. In this experience we converted the students into active actors in the evaluation process itself and we endowed ourselves as teachers with a flexible process in which along with qualifications (score), other attitudinal feedback arises. Despite the high level of self-affirmation on their discussion within the interactive assessment sessions, they ultimately have exhibited a great ability to review and correct the wrong reasoning when that was the case.
Keywords: Cooperative-collaborative learning, educational management, formative-summative assessment, leadership training.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1361650 Low Power and Less Area Architecture for Integer Motion Estimation
Authors: C Hisham, K Komal, Amit K Mishra
Abstract:
Full search block matching algorithm is widely used for hardware implementation of motion estimators in video compression algorithms. In this paper we are proposing a new architecture, which consists of a 2D parallel processing unit and a 1D unit both working in parallel. The proposed architecture reduces both data access power and computational power which are the main causes of power consumption in integer motion estimation. It also completes the operations with nearly the same number of clock cycles as compared to a 2D systolic array architecture. In this work sum of absolute difference (SAD)-the most repeated operation in block matching, is calculated in two steps. The first step is to calculate the SAD for alternate rows by a 2D parallel unit. If the SAD calculated by the parallel unit is less than the stored minimum SAD, the SAD of the remaining rows is calculated by the 1D unit. Early termination, which stops avoidable computations has been achieved with the help of alternate rows method proposed in this paper and by finding a low initial SAD value based on motion vector prediction. Data reuse has been applied to the reference blocks in the same search area which significantly reduced the memory access.
Keywords: Sum of absolute difference, high speed DSP.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1489649 Inadequate Requirements Engineering Process: A Key Factor for Poor Software Development in Developing Nations: A Case Study
Authors: K. Adu Michael, K. Alese Boniface
Abstract:
Developing a reliable and sustainable software products is today a big challenge among up–coming software developers in Nigeria. The inability to develop a comprehensive problem statement needed to execute proper requirements engineering process is missing. The need to describe the ‘what’ of a system in one document, written in a natural language is a major step in the overall process of Software Engineering. Requirements Engineering is a process use to discover, analyze and validate system requirements. This process is needed in reducing software errors at the early stage of the development of software. The importance of each of the steps in Requirements Engineering is clearly explained in the context of using detailed problem statement from client/customer to get an overview of an existing system along with expectations from the new system. This paper elicits inadequate Requirements Engineering principle as the major cause of poor software development in developing nations using a case study of final year computer science students of a tertiary-education institution in Nigeria.
Keywords: Client/Customer, Problem Statement, Requirements Engineering, Software Developers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2454648 Risk Assessment in Durations and Costs for Construction of Industrial Facilities in Egypt Using Equations and Computer
Authors: M. Kamal Elbokl, Negadi Kheira
Abstract:
Risk Evaluation is an important step in protecting your workers and your business, as well as complying with the law. It helps you focus on the risks that really matter in your workplace – the ones with the potential to cause real harm. We are in this paper introduce basics of risk assessment then we mention some of ways to risk evaluation by computer especially Monte Carlo simulation and Microsoft project.
We use Program Evaluation and Review Technique (PERT) to deal with Risks in Industrial Facilities in Evaluation and Assessment for this risk. Using PERT Technique in Microsoft Project by the PERT toolbar and using PERTMASTER Program with Primavera Program we evaluate many hazards and make calculations for that by mathematical equation to make right decisions. We define and calculate risk factor and risk severity to ranking the type of the risk then dealing with it using in that many ways like probability computation, curves, and tables. By introducing variables in the equation of functions in computer programs we calculate the risk in the time and the cost in general case and then mention some examples in industrial facilities field.
Keywords: Risk, Industrial Facilities, PERT, Monte Carlo Simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1951647 Duration Analysis of New Firms in the Banking Industry
Authors: Jesus Orbe, Vicente Nunez-Anton
Abstract:
This paper studies the duration or survival time of commercial banks active in the Moscovian three month Rouble deposits market, during the 1994-1997 period. The privatization process of the Russian commercial banking industry, after the 1988 banking reform, caused a massive entry of new banks followed by a period of high rates of exit. As a consequence, many firms went bankrupt without refunding their deposits. Therefore, both for the banks and for the banks- depositors, it is of interest to analyze which are the significant characteristics that motivate the exit or the closing of the bank. We propose a different methodology based on penalized weighted least squares which represents a very general, flexible and innovative approach for this type of analysis. The more relevant results are that smaller banks exit sooner, banks that enter the market in the last part of the study have shorter durations. As expected, the more experienced banks have a longer duration in the market. In addition, the mean survival time is lower for banks which offer extreme interest rates.
Keywords: Banking, censored, duration, Kaplan-Meier.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1548646 Characteristics of Ozone Generated from Dielectric Barrier Discharge Plasma Actuators
Authors: R. Osada, S. Ogata, T. Segawa
Abstract:
Dielectric barrier discharge plasma actuators (DBD-PAs) have been developed for active flow control devices. However, it is necessary to reduce ozone produced by DBD toward practical applications using DBD-PAs. In this study, variations of ozone concentration, flow velocity, power consumption were investigated by changing exposed electrodes of DBD-PAs. Two exposed electrode prototypes were prepared: span-type with exposed electrode width of 0.1 mm, and normal-type with width of 5 mm. It was found that span-type shows lower power consumption and higher flow velocity than that of normal-type at Vp-p = 4.0-6.0 kV. Ozone concentration of span-type higher than normal-type at Vp-p = 4.0-8.0 kV. In addition, it was confirmed that catalyst located in downstream from the exposed electrode can reduce ozone concentration between 18 and 42% without affecting the induced flow.Keywords: Dielectric barrier discharge plasma actuators, ozone diffusion, PIV measurement, power consumption.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1188645 An Analysis of the Social Network Structure of Knowledge Management Students at NTU
Authors: Guo Yanru, Zhu Xiaobo, Lee Chu Keong
Abstract:
This paper maps the structure of the social network of the 2011 class ofsixty graduate students of the Masters of Science (Knowledge Management) programme at the Nanyang Technological University, based on their friending relationships on Facebook. To ensure anonymity, actual names were not used. Instead, they were replaced with codes constructed from their gender, nationality, mode of study, year of enrollment and a unique number. The relationships between friends within the class, and among the seniors and alumni of the programme wereplotted. UCINet and Pajek were used to plot the sociogram, to compute the density, inclusivity, and degree, global, betweenness, and Bonacich centralities, to partition the students into two groups, namely, active and peripheral, and to identify the cut-points. Homophily was investigated, and it was observed for nationality and study mode. The groups students formed on Facebook were also studied, and of fifteen groups, eight were classified as dead, which we defined as those that have been inactive for over two months.Keywords: Facebook, friending relationships, Social network analysis, social network sites, structural position
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1744644 Preparation and Characterization of Recycled PET/PP Blends from Automotive Textile Waste for Use in the Furniture Edge Banding Sector
Authors: Merve Ozer, Tolga Gokkurt, Yasemen Gokkurt, Ezgi Bozbey
Abstract:
In this study, research has been conducted on the recovery of automotive textile waste, which has heavy use in the automotive sector and consists of PET/PP content, through the upcycling technique of post-product and post-consumer usage. The aim is to investigate the formulation and production methods that will enable the substitution of original PP raw materials, used in the production of plastic edge bands, with PP/PET alloys. The lamination structure of the mentioned waste makes it impossible to separate the incompatible PP and PET phases, thereby hindering the production of high-quality raw materials or products through recycling. In this study, a comprehensive process was examined through a two-step production process using different types of block and maleic-grafted copolymers to achieve compatibility between these two incompatible phases. The obtained plastic raw materials, referred to as PP/PET blends, were examined in detail, with a focus on their mechanical, thermal, and morphological properties, to discuss their substitutability for the original raw materials.
Keywords: Twin screw extruders, mechanical recycling, melt blending, plastic blends, polyethylene, polypropylene, recycling of plastics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 174643 Control of Commutation of SR Motor Using Its Magnetic Characteristics and Back-of-Core Saturation Effects
Authors: Dr. N.H. Mvungi
Abstract:
The control of commutation of switched reluctance (SR) motor has nominally depended on a physical position detector. The physical rotor position sensor limits robustness and increases size and inertia of the SR drive system. The paper describes a method to overcome these limitations by using magnetization characteristics of the motor to indicate rotor and stator teeth overlap status. The method is using active current probing pulses of same magnitude that is used to simulate flux linkage in the winding being probed. A microprocessor is used for processing magnetization data to deduce rotor-stator teeth overlap status and hence rotor position. However, the back-of-core saturation and mutual coupling introduces overlap detection errors, hence that of commutation control. This paper presents the concept of the detection scheme and the effects of backof core saturation.Keywords: Microprocessor control, rotor position, sensorless, switched reluctance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1282642 Improvement of Durability of Wood by Maleic Anhydride
Authors: Yong F. Li, Yi X. Liu, Xiang M. Wang, Feng H. Wang
Abstract:
Wood as a natural renewable material is vulnerable to degradation by microorganisms and susceptible to change in dimension by water. In order to effectively improve the durability of wood, an active reagent, maleic anhydride (Man) was selected for wood modification. Man was first dissolved into a solvent, and then penetrated into wood porous structure under a vacuum/pressure condition. After a final catalyst-thermal treatment, wood modification was finished. The test results indicate that acetone is a good solvent for transporting Man into wood matrix. SEM observation proved that wood samples treated by Man kept a good cellular structure, indicating a well penetration of Man into wood cell walls. FTIR analysis suggested that Man reacted with hydroxyl groups on wood cell walls by its ring-ether group, resulting in reduction of amount of hydroxyl groups and resultant good dimensional stability as well as fine decay resistance. Consequently, Man modifying wood to improve its durability is an effective method.Keywords: Wood, porous structure, durability improvement, maleic anhydride
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2101641 Bioethanol: Indonesian Macro-Algae as a Renewable Feedstock for Liquid Fuel
Authors: T. Poespowati, E. Marsyahyo, R. Kartika-Dewi
Abstract:
This experimental study aims at studying the conversion of macro-algae into bioethanol under several steps of procedure: preparation, pre-treatment, fermentation, and distillation. The main objective of this work was to investigate the role of buffer’s type as a stabiliser of pH level and fermentation time on the yield of ethanol. For this purpose, experiments were carried out on biomass macro-algae to de-couple the pre-treatment and fermentation processes from those associated with distillation process. β- glucosidase was used as cellulose decomposer during hydrolysis step and yeast was used during fermentation process. The species of macro-algae utilised as energy feedstock was Ulva lactuca and it was harvested from southern coast of Central of Java Island – Indonesia. Experiments were conducted in a simple fermenter over a different buffer: citrate buffer and acetic buffer, and over a range of fermentation times between 5 to 20 days. The ethanol production was found to be significantly affected by both variables. The optimum time of fermentation was 10 days with citrate buffer; result in 0.88458% of ethanol, and the ethanol content after distillation process was shown 0.985015%.
Keywords: Fermentation, ulva-lactuca, buffer, β-glucosidase, bioethanol.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2436640 Decontamination of Cr(VI) Polluted Wastewater by use of Low Cost Industrial Wastes
Authors: Marius Gheju, Rodica Pode
Abstract:
The reduction of hexavalent chromium by scrap iron was investigated in continuous system, using long-term column experiments, for aqueous Cr(VI) solutions having low buffering capacities, over the Cr(VI) concentration range of 5 – 40 mg/L. The results showed that the initial Cr(VI) concentration significantly affects the reduction capacity of scrap iron. Maximum reduction capacity of scrap iron was observed at the beginning of the column experiments; the lower the Cr(VI) concentration, the greater the experiment duration with maximum scrap iron reduction capacity. However, due to passivation of active surface, scrap iron reduction capacity continuously decreased in time, especially after Cr(VI) breakthrough. The experimental results showed that highest reduction capacity recorded until Cr(VI) breakthrough was 22.8 mg Cr(VI)/g scrap iron, at CI = 5 mg/L, and decreased with increasing Cr(VI) concentration. In order to assure total reduction of greater Cr(VI) concentrations for a longer period of time, either the mass of scrap iron filling, or the hydraulic retention time should be increased.Keywords: hexavalent chromium, heavy metals, scrap iron, reduction capacity, wastewater treatment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1840639 Scheduling Multiple Workflow Using De-De Dodging Algorithm and PBD Algorithm in Cloud: Detailed Study
Authors: B. Arun Kumar, T. Ravichandran
Abstract:
Workflow scheduling is an important part of cloud computing and based on different criteria it decides cost, execution time, and performances. A cloud workflow system is a platform service facilitating automation of distributed applications based on new cloud infrastructure. An aspect which differentiates cloud workflow system from others is market-oriented business model, an innovation which challenges conventional workflow scheduling strategies. Time and Cost optimization algorithm for scheduling Hybrid Clouds (TCHC) algorithm decides which resource should be chartered from public providers is combined with a new De-De algorithm considering that every instance of single and multiple workflows work without deadlocks. To offset this, two new concepts - De-De Dodging Algorithm and Priority Based Decisive Algorithm - combine with conventional deadlock avoidance issues by proposing one algorithm that maximizes active (not just allocated) resource use and reduces Makespan.Keywords: Workflow Scheduling, cloud workflow, TCHC algorithm, De-De Dodging Algorithm, Priority Based Decisive Algorithm (PBD), Makespan.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2795638 Performance Evaluation of a Minimum Mean Square Error-Based Physical Sidelink Share Channel Receiver under Fading Channel
Authors: Yang Fu, Jaime Rodrigo Navarro, Jose F. Monserrat, Faiza Bouchmal, Oscar Carrasco Quilis
Abstract:
Cellular Vehicle to Everything (C-V2X) is considered a promising solution for future autonomous driving. From Release 16 to Release 17, the Third Generation Partnership Project (3GPP) has introduced the definitions and services for 5G New Radio (NR) V2X. Since establishing a simulator for C-V2X communications is an essential preliminary step to achieve reliable and stable communication links, this paper proposes a complete framework of a link-level simulator based on the 3GPP specifications for the Physical Sidelink Share Channel (PSSCH) of the 5G NR Physical Layer (PHY). In this framework, several algorithms in the receiver part, i.e., sliding window in channel estimation and Minimum Mean Square Error (MMSE)-based equalization, are developed. Finally, the performance of the developed PSSCH receiver is validated through extensive simulations under different assumptions.
Keywords: Yang Fu, Jaime Rodrigo Navarro, Jose F. Monserrat, Faiza Bouchmal, Oscar Carrasco Quilis
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 147