Search results for: Delivery performance
151 Hyperspectral Imaging and Nonlinear Fukunaga-Koontz Transform Based Food Inspection
Authors: Hamidullah Binol, Abdullah Bal
Abstract:
Nowadays, food safety is a great public concern; therefore, robust and effective techniques are required for detecting the safety situation of goods. Hyperspectral Imaging (HSI) is an attractive material for researchers to inspect food quality and safety estimation such as meat quality assessment, automated poultry carcass inspection, quality evaluation of fish, bruise detection of apples, quality analysis and grading of citrus fruits, bruise detection of strawberry, visualization of sugar distribution of melons, measuring ripening of tomatoes, defect detection of pickling cucumber, and classification of wheat kernels. HSI can be used to concurrently collect large amounts of spatial and spectral data on the objects being observed. This technique yields with exceptional detection skills, which otherwise cannot be achieved with either imaging or spectroscopy alone. This paper presents a nonlinear technique based on kernel Fukunaga-Koontz transform (KFKT) for detection of fat content in ground meat using HSI. The KFKT which is the nonlinear version of FKT is one of the most effective techniques for solving problems involving two-pattern nature. The conventional FKT method has been improved with kernel machines for increasing the nonlinear discrimination ability and capturing higher order of statistics of data. The proposed approach in this paper aims to segment the fat content of the ground meat by regarding the fat as target class which is tried to be separated from the remaining classes (as clutter). We have applied the KFKT on visible and nearinfrared (VNIR) hyperspectral images of ground meat to determine fat percentage. The experimental studies indicate that the proposed technique produces high detection performance for fat ratio in ground meat.Keywords: Food (Ground meat) inspection, Fukunaga-Koontz transform, hyperspectral imaging, kernel methods.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1500150 Impact of Quality Assurance Mechanisms on the Work Efficiency of Staff in the Educational Space of Georgia
Authors: B. Gechbaia, K. Goletiani, G. Gabedava, N. Mikeltadze
Abstract:
At this stage, Georgia is a country which is actively involved in the European integration process, for which the primary priority is effective integration in the European education system. The modern Georgian higher education system is the process of establishing a new sociocultural reality, whose main priorities are determined by the Quality System as a continuous cycle of planning, implementation, checking and acting. Obviously, in this situation, the issue of management of education institutions comes out in the foreground, since the proper planning and implementation of personnel management processes is one of the main determinants of the company's performance. At the same time, one of the most important factors is the psychological comfort of the personnel, ensuring their protection and efficiency of stress management policy.
The purpose of this research is to determine how intensely the relationship is between the psychological comfort of the personnel and the efficiency of the quality system in the institution as the quality assurance mechanisms of educational institutions affect the stability of personnel, prevention and management of the stressful situation. The research was carried out within the framework of the Internal Grant Project «The Role of Organizational Culture in the Process of Settlement of Management of Stress and Conflict, Georgian Reality and European Experience » of the Batumi Navigation Teaching University, based on the analysis of the survey results of target groups. The small-scale research conducted by us has revealed that the introduction of quality assurance system and its active implementation increased the quality of management of Georgian educational institutions, increased the level of universal engagement in internal and external processes and as a result, it has improved the quality of education as well as social and psychological comfort indicators of the society.
Keywords: Quality assurance, effective management, stability of personnel, psychological comfort, stress management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1144149 Evaluating Generative Neural Attention Weights-Based Chatbot on Customer Support Twitter Dataset
Authors: Sinarwati Mohamad Suhaili, Naomie Salim, Mohamad Nazim Jambli
Abstract:
Sequence-to-sequence (seq2seq) models augmented with attention mechanisms are increasingly important in automated customer service. These models, adept at recognizing complex relationships between input and output sequences, are essential for optimizing chatbot responses. Central to these mechanisms are neural attention weights that determine the model’s focus during sequence generation. Despite their widespread use, there remains a gap in the comparative analysis of different attention weighting functions within seq2seq models, particularly in the context of chatbots utilizing the Customer Support Twitter (CST) dataset. This study addresses this gap by evaluating four distinct attention-scoring functions—dot, multiplicative/general, additive, and an extended multiplicative function with a tanh activation parameter — in neural generative seq2seq models. Using the CST dataset, these models were trained and evaluated over 10 epochs with the AdamW optimizer. Evaluation criteria included validation loss and BLEU scores implemented under both greedy and beam search strategies with a beam size of k = 3. Results indicate that the model with the tanh-augmented multiplicative function significantly outperforms its counterparts, achieving the lowest validation loss (1.136484) and the highest BLEU scores (0.438926 under greedy search, 0.443000 under beam search, k = 3). These findings emphasize the crucial influence of selecting an appropriate attention-scoring function to enhance the performance of seq2seq models for chatbots, particularly highlighting the model integrating tanh activation as a promising approach to improving chatbot quality in customer support contexts.
Keywords: Attention weight, chatbot, encoder-decoder, neural generative attention, score function, sequence-to-sequence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 91148 Ramp Rate and Constriction Factor Based Dual Objective Economic Load Dispatch Using Particle Swarm Optimization
Authors: Himanshu Shekhar Maharana, S. K .Dash
Abstract:
Economic Load Dispatch (ELD) proves to be a vital optimization process in electric power system for allocating generation amongst various units to compute the cost of generation, the cost of emission involving global warming gases like sulphur dioxide, nitrous oxide and carbon monoxide etc. In this dissertation, we emphasize ramp rate constriction factor based particle swarm optimization (RRCPSO) for analyzing various performance objectives, namely cost of generation, cost of emission, and a dual objective function involving both these objectives through the experimental simulated results. A 6-unit 30 bus IEEE test case system has been utilized for simulating the results involving improved weight factor advanced ramp rate limit constraints for optimizing total cost of generation and emission. This method increases the tendency of particles to venture into the solution space to ameliorate their convergence rates. Earlier works through dispersed PSO (DPSO) and constriction factor based PSO (CPSO) give rise to comparatively higher computational time and less good optimal solution at par with current dissertation. This paper deals with ramp rate and constriction factor based well defined ramp rate PSO to compute various objectives namely cost, emission and total objective etc. and compares the result with DPSO and weight improved PSO (WIPSO) techniques illustrating lesser computational time and better optimal solution.
Keywords: Economic load dispatch, constriction factor based particle swarm optimization, dispersed particle swarm optimization, weight improved particle swarm optimization, ramp rate and constriction factor based particle swarm optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1260147 A Hybrid Image Fusion Model for Generating High Spatial-Temporal-Spectral Resolution Data Using OLI-MODIS-Hyperion Satellite Imagery
Authors: Yongquan Zhao, Bo Huang
Abstract:
Spatial, Temporal, and Spectral Resolution (STSR) are three key characteristics of Earth observation satellite sensors; however, any single satellite sensor cannot provide Earth observations with high STSR simultaneously because of the hardware technology limitations of satellite sensors. On the other hand, a conflicting circumstance is that the demand for high STSR has been growing with the remote sensing application development. Although image fusion technology provides a feasible means to overcome the limitations of the current Earth observation data, the current fusion technologies cannot enhance all STSR simultaneously and provide high enough resolution improvement level. This study proposes a Hybrid Spatial-Temporal-Spectral image Fusion Model (HSTSFM) to generate synthetic satellite data with high STSR simultaneously, which blends the high spatial resolution from the panchromatic image of Landsat-8 Operational Land Imager (OLI), the high temporal resolution from the multi-spectral image of Moderate Resolution Imaging Spectroradiometer (MODIS), and the high spectral resolution from the hyper-spectral image of Hyperion to produce high STSR images. The proposed HSTSFM contains three fusion modules: (1) spatial-spectral image fusion; (2) spatial-temporal image fusion; (3) temporal-spectral image fusion. A set of test data with both phenological and land cover type changes in Beijing suburb area, China is adopted to demonstrate the performance of the proposed method. The experimental results indicate that HSTSFM can produce fused image that has good spatial and spectral fidelity to the reference image, which means it has the potential to generate synthetic data to support the studies that require high STSR satellite imagery.Keywords: Hybrid spatial-temporal-spectral fusion, high resolution synthetic imagery, least square regression, sparse representation, spectral transformation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1235146 The Current Practices of Analysis of Reinforced Concrete Panels Subjected to Blast Loading
Authors: Palak J. Shukla, Atul K. Desai, Chentankumar D. Modhera
Abstract:
For any country in the world, it has become a priority to protect the critical infrastructure from looming risks of terrorism. In any infrastructure system, the structural elements like lower floors, exterior columns, walls etc. are key elements which are the most susceptible to damage due to blast load. The present study revisits the state of art review of the design and analysis of reinforced concrete panels subjected to blast loading. Various aspects in association with blast loading on structure, i.e. estimation of blast load, experimental works carried out previously, the numerical simulation tools, various material models, etc. are considered for exploring the current practices adopted worldwide. Discussion on various parametric studies to investigate the effect of reinforcement ratios, thickness of slab, different charge weight and standoff distance is also made. It was observed that for the simulation of blast load, CONWEP blast function or equivalent numerical equations were successfully employed by many researchers. The study of literature indicates that the researches were carried out using experimental works and numerical simulation using well known generalized finite element methods, i.e. LS-DYNA, ABAQUS, AUTODYN. Many researchers recommended to use concrete damage model to represent concrete and plastic kinematic material model to represent steel under action of blast loads for most of the numerical simulations. Most of the studies reveal that the increase reinforcement ratio, thickness of slab, standoff distance was resulted in better blast resistance performance of reinforced concrete panel. The study summarizes the various research results and appends the present state of knowledge for the structures exposed to blast loading.
Keywords: Blast phenomenon, experimental methods, material models, numerical methods.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1136145 Experimental Investigation of Visual Comfort Requirement in Garment Factories and Identify the Cost Saving Opportunities
Authors: M. A. Wijewardane, S. A. N. C. Sudasinghe, H. K. G. Punchihewa, W. K. D. L. Wickramasinghe, S. A. Philip, M. R. S. U. Kumara
Abstract:
Visual comfort is one of the major parameters that can be taken to measure the human comfort in any environment. If the provided illuminance level in a working environment does not meet the workers visual comfort, it will lead to eye-strain, fatigue, headache, stress, accidents and finally, poor productivity. However, improvements in lighting do not necessarily mean that the workplace requires more light. Unnecessarily higher illuminance levels will also cause poor visual comfort and health risks. In addition, more power consumption on lighting will also result in higher energy costs. So, during this study, visual comfort and the illuminance requirement for the workers in textile/apparel industry were studied to perform different tasks (i.e. cutting, sewing and knitting) at their workplace. Experimental studies were designed to identify the optimum illuminance requirement depending upon the varied fabric colour and type and finally, energy saving potentials due to controlled illuminance level depending on the workforce requirement were analysed. Visual performance of workers during the sewing operation was studied using the ‘landolt ring experiment’. It was revealed that around 36.3% of the workers would like to work if the illuminance level varies from 601 lux to 850 lux illuminance level and 45.9% of the workers are not happy to work if the illuminance level reduces less than 600 lux and greater than 850 lux. Moreover, more than 65% of the workers who do not satisfy with the existing illuminance levels of the production floors suggested that they have headache, eye diseases, or both diseases due to poor visual comfort. In addition, findings of the energy analysis revealed that the energy-saving potential of 5%, 10%, 24%, 8% and 16% can be anticipated for fabric colours, red, blue, yellow, black and white respectively, when the 800 lux is the prevailing illuminance level for sewing operation.
Keywords: Landolt ring experiment, lighting energy consumption, illuminance, textile and apparel industry, visual comfort.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 830144 Taguchi Robust Design for Optimal Setting of Process Wastes Parameters in an Automotive Parts Manufacturing Company
Authors: Charles Chikwendu Okpala, Christopher Chukwutoo Ihueze
Abstract:
As a technique that reduces variation in a product by lessening the sensitivity of the design to sources of variation, rather than by controlling their sources, Taguchi Robust Design entails the designing of ideal goods, by developing a product that has minimal variance in its characteristics and also meets the desired exact performance. This paper examined the concept of the manufacturing approach and its application to brake pad product of an automotive parts manufacturing company. Although the firm claimed that only defects, excess inventory, and over-production were the few wastes that grossly affect their productivity and profitability, a careful study and analysis of their manufacturing processes with the application of Single Minute Exchange of Dies (SMED) tool showed that the waste of waiting is the fourth waste that bedevils the firm. The selection of the Taguchi L9 orthogonal array which is based on the four parameters and the three levels of variation for each parameter revealed that with a range of 2.17, that waiting is the major waste that the company must reduce in order to continue to be viable. Also, to enhance the company’s throughput and profitability, the wastes of over-production, excess inventory, and defects with ranges of 2.01, 1.46, and 0.82, ranking second, third, and fourth respectively must also be reduced to the barest minimum. After proposing -33.84 as the highest optimum Signal-to-Noise ratio to be maintained for the waste of waiting, the paper advocated for the adoption of all the tools and techniques of Lean Production System (LPS), and Continuous Improvement (CI), and concluded by recommending SMED in order to drastically reduce set up time which leads to unnecessary waiting.Keywords: Taguchi Robust Design, signal to noise ratio, Single Minute Exchange of Dies, lean production system, waste.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 976143 Classifying Turbomachinery Blade Mode Shapes Using Artificial Neural Networks
Authors: Ismail Abubakar, Hamid Mehrabi, Reg Morton
Abstract:
Currently, extensive signal analysis is performed in order to evaluate structural health of turbomachinery blades. This approach is affected by constraints of time and the availability of qualified personnel. Thus, new approaches to blade dynamics identification that provide faster and more accurate results are sought after. Generally, modal analysis is employed in acquiring dynamic properties of a vibrating turbomachinery blade and is widely adopted in condition monitoring of blades. The analysis provides useful information on the different modes of vibration and natural frequencies by exploring different shapes that can be taken up during vibration since all mode shapes have their corresponding natural frequencies. Experimental modal testing and finite element analysis are the traditional methods used to evaluate mode shapes with limited application to real live scenario to facilitate a robust condition monitoring scheme. For a real time mode shape evaluation, rapid evaluation and low computational cost is required and traditional techniques are unsuitable. In this study, artificial neural network is developed to evaluate the mode shape of a lab scale rotating blade assembly by using result from finite element modal analysis as training data. The network performance evaluation shows that artificial neural network (ANN) is capable of mapping the correlation between natural frequencies and mode shapes. This is achieved without the need of extensive signal analysis. The approach offers advantage from the perspective that the network is able to classify mode shapes and can be employed in real time including simplicity in implementation and accuracy of the prediction. The work paves the way for further development of robust condition monitoring system that incorporates real time mode shape evaluation.
Keywords: Modal analysis, artificial neural network, mode shape, natural frequencies, pattern recognition.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 908142 Numerical Investigation on Anchored Sheet Pile Quay Wall with Separated Relieving Platform
Authors: Mahmoud Roushdy, Mohamed El Naggar, Ahmed Yehia Abdelaziz
Abstract:
Anchored sheet pile has been used worldwide as front quay walls for decades. With the increase in vessel drafts and weights, those sheet pile walls need to be upgraded by increasing the depth of the dredging line in front of the wall. One of the upgrades for the sheet pile wall is to add a separated platform to the system, where the platform is structurally separated from the front wall. The platform is structurally separated from the front wall. This paper presents a numerical investigation utilizing finite element analysis on the behavior of separated relieve platforms installed within existing anchored sheet pile quay walls. The investigation was done in two steps: a verification step followed by a parametric study. In the verification step, the numerical model was verified based on field measurements performed by others. The validated model was extended within the parametric study to a series of models with different backfill soils, separation gap width, and number of pile rows supporting the platform. The results of the numerical investigation show that using stiff clay as backfill soil (neglecting consolidation) gives better performance for the front wall and the first pile row adjacent to sandy backfills. The degree of compaction of the sandy backfill slightly increases lateral deformations but reduces bending moment acting on pile rows, while the effect is minor on the front wall. In addition, the increase in the separation gap width gradually increases bending moments on the front wall regardless of the backfill soil type, while this effect is reversed on pile rows (gradually decrease). Finally, the paper studies the possibility of deepening the basin along with the separation to take advantage of the positive separation effect on piles, and front wall.
Keywords: Anchored sheet pile, relieving platform, separation gap, upgrade quay wall.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 277141 Text Mining Technique for Data Mining Application
Authors: M. Govindarajan
Abstract:
Text Mining is around applying knowledge discovery techniques to unstructured text is termed knowledge discovery in text (KDT), or Text data mining or Text Mining. In decision tree approach is most useful in classification problem. With this technique, tree is constructed to model the classification process. There are two basic steps in the technique: building the tree and applying the tree to the database. This paper describes a proposed C5.0 classifier that performs rulesets, cross validation and boosting for original C5.0 in order to reduce the optimization of error ratio. The feasibility and the benefits of the proposed approach are demonstrated by means of medial data set like hypothyroid. It is shown that, the performance of a classifier on the training cases from which it was constructed gives a poor estimate by sampling or using a separate test file, either way, the classifier is evaluated on cases that were not used to build and evaluate the classifier are both are large. If the cases in hypothyroid.data and hypothyroid.test were to be shuffled and divided into a new 2772 case training set and a 1000 case test set, C5.0 might construct a different classifier with a lower or higher error rate on the test cases. An important feature of see5 is its ability to classifiers called rulesets. The ruleset has an error rate 0.5 % on the test cases. The standard errors of the means provide an estimate of the variability of results. One way to get a more reliable estimate of predictive is by f-fold –cross- validation. The error rate of a classifier produced from all the cases is estimated as the ratio of the total number of errors on the hold-out cases to the total number of cases. The Boost option with x trials instructs See5 to construct up to x classifiers in this manner. Trials over numerous datasets, large and small, show that on average 10-classifier boosting reduces the error rate for test cases by about 25%.Keywords: C5.0, Error Ratio, text mining, training data, test data.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2489140 Spatial Variation of WRF Model Rainfall Prediction over Uganda
Authors: Isaac Mugume, Charles Basalirwa, Daniel Waiswa, Triphonia Ngailo
Abstract:
Rainfall is a major climatic parameter affecting many sectors such as health, agriculture and water resources. Its quantitative prediction remains a challenge to weather forecasters although numerical weather prediction models are increasingly being used for rainfall prediction. The performance of six convective parameterization schemes, namely the Kain-Fritsch scheme, the Betts-Miller-Janjic scheme, the Grell-Deveny scheme, the Grell-3D scheme, the Grell-Fretas scheme, the New Tiedke scheme of the weather research and forecast (WRF) model regarding quantitative rainfall prediction over Uganda is investigated using the root mean square error for the March-May (MAM) 2013 season. The MAM 2013 seasonal rainfall amount ranged from 200 mm to 900 mm over Uganda with northern region receiving comparatively lower rainfall amount (200–500 mm); western Uganda (270–550 mm); eastern Uganda (400–900 mm) and the lake Victoria basin (400–650 mm). A spatial variation in simulated rainfall amount by different convective parameterization schemes was noted with the Kain-Fritsch scheme over estimating the rainfall amount over northern Uganda (300–750 mm) but also presented comparable rainfall amounts over the eastern Uganda (400–900 mm). The Betts-Miller-Janjic, the Grell-Deveny, and the Grell-3D underestimated the rainfall amount over most parts of the country especially the eastern region (300–600 mm). The Grell-Fretas captured rainfall amount over the northern region (250–450 mm) but also underestimated rainfall over the lake Victoria Basin (150–300 mm) while the New Tiedke generally underestimated rainfall amount over many areas of Uganda. For deterministic rainfall prediction, the Grell-Fretas is recommended for rainfall prediction over northern Uganda while the Kain-Fritsch scheme is recommended over eastern region.Keywords: Convective parameterization schemes, March-May 2013 rainfall season, spatial variation of parameterization schemes over Uganda, WRF model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1229139 Systems Engineering and Project Management Process Modeling in the Aeronautics Context: Case Study of SMEs
Authors: S. Lemoussu, J. C. Chaudemar, R. A. Vingerhoeds
Abstract:
The aeronautics sector is currently living an unprecedented growth largely due to innovative projects. In several cases, such innovative developments are being carried out by Small and Medium sized-Enterprises (SMEs). For instance, in Europe, a handful of SMEs are leading projects like airships, large civil drones, or flying cars. These SMEs have all limited resources, must make strategic decisions, take considerable financial risks and in the same time must take into account the constraints of safety, cost, time and performance as any commercial organization in this industry. Moreover, today, no international regulations fully exist for the development and certification of this kind of projects. The absence of such a precise and sufficiently detailed regulatory framework requires a very close contact with regulatory instances. But, SMEs do not always have sufficient resources and internal knowledge to handle this complexity and to discuss these issues. This poses additional challenges for those SMEs that have system integration responsibilities and that must provide all the necessary means of compliance to demonstrate their ability to design, produce, and operate airships with the expected level of safety and reliability. The final objective of our research is thus to provide a methodological framework supporting SMEs in their development taking into account recent innovation and institutional rules of the sector. We aim to provide a contribution to the problematic by developing a specific Model-Based Systems Engineering (MBSE) approach. Airspace regulation, aeronautics standards and international norms on systems engineering are taken on board to be formalized in a set of models. This paper presents the on-going research project combining Systems Engineering and Project Management process modeling and taking into account the metamodeling problematic.
Keywords: Aeronautics, certification, process modeling, project management, SME, systems engineering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1434138 A Novel GNSS Integrity Augmentation System for Civil and Military Aircraft
Authors: Roberto Sabatini, Terry Moore, Chris Hill
Abstract:
This paper presents a novel Global Navigation Satellite System (GNSS) Avionics Based Integrity Augmentation (ABIA) system architecture suitable for civil and military air platforms, including Unmanned Aircraft Systems (UAS). Taking the move from previous research on high-accuracy Differential GNSS (DGNSS) systems design, integration and experimental flight test activities conducted at the Italian Air Force Flight Test Centre (CSV-RSV), our research focused on the development of a novel approach to the problem of GNSS ABIA for mission- and safety-critical air vehicle applications and for multi-sensor avionics architectures based on GNSS. Detailed mathematical models were developed to describe the main causes of GNSS signal outages and degradation in flight, namely: antenna obscuration, multipath, fading due to adverse geometry and Doppler shift. Adopting these models in association with suitable integrity thresholds and guidance algorithms, the ABIA system is able to generate integrity cautions (predictive flags) and warnings (reactive flags), as well as providing steering information to the pilot and electronic commands to the aircraft/UAS flight control systems. These features allow real-time avoidance of safety-critical flight conditions and fast recovery of the required navigation performance in case of GNSS data losses. In other words, this novel ABIA system addresses all three cornerstones of GNSS integrity augmentation in mission- and safety-critical applications: prediction (caution flags), reaction (warning flags) and correction (alternate flight path computation).
Keywords: Global Navigation Satellite Systems (GNSS), Integrity Augmentation, Unmanned Aircraft Systems, Aircraft Based Augmentation, Avionics Based Integrity Augmentation, Safety-Critical Applications.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3244137 Controller Design for Euler-Bernoulli Smart Structures Using Robust Decentralized POF via Reduced Order Modeling
Authors: T.C. Manjunath, B. Bandyopadhyay
Abstract:
This paper features the proposed modeling and design of a Robust Decentralized Periodic Output Feedback (RDPOF) control technique for the active vibration control of smart flexible multimodel Euler-Bernoulli cantilever beams for a multivariable (MIMO) case by retaining the first 6 vibratory modes. The beam structure is modeled in state space form using the concept of piezoelectric theory, the Euler-Bernoulli beam theory and the Finite Element Method (FEM) technique by dividing the beam into 4 finite elements and placing the piezoelectric sensor / actuator at two finite element locations (positions 2 and 4) as collocated pairs, i.e., as surface mounted sensor / actuator, thus giving rise to a multivariable model of the smart structure plant with two inputs and two outputs. Five such multivariable models are obtained by varying the dimensions (aspect ratios) of the aluminum beam, thus giving rise to a multimodel of the smart structure system. Using model order reduction technique, the reduced order model of the higher order system is obtained based on dominant eigen value retention and the method of Davison. RDPOF controllers are designed for the above 5 multivariable-multimodel plant. The closed loop responses with the RDPOF feedback gain and the magnitudes of the control input are observed and the performance of the proposed multimodel smart structure system with the controller is evaluated for vibration control.Keywords: Smart structure, Euler-Bernoulli beam theory, Periodic output feedback control, Finite Element Method, State space model, SISO, Embedded sensors and actuators, Vibration control, Reduced order model
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2028136 Molecular and Serological Diagnosis of Newcastle and Ornithobacterium rhinotracheale Broiler in Chicken in Fars Province, Iran
Authors: Mohammadjavad Mehrabanpour, Maryam Ranjbar Bushehri, Dorsa Mehrabanpour
Abstract:
Respiratory diseases are the most important problems in the country’s poultry industry, particularly when it comes to broiler flocks. Ornithobacterium rhinotracheale (ORT) is a species that causes poor performance in growth rate, egg production, and mortality. This pathogen causes a respiratory infection including pulmonary alveolar inflammation, and pneumonia of birds throughout the world. Newcastle disease (ND) is a highly contagious disease in poultry, and also, it causes considerable losses to the poultry industry. The aim of this study was to evaluate the simultaneous occurrence of ORT and ND and NDV isolation by inoculation in embryonated eggs and confirmed by RT-PCR in broiler chicken flocks in Fars province. In this study, 318 blood and 85 tissue samples (brain, trachea, liver, and cecal tonsils) were collected from 15 broiler chicken farms. Survey serum antibody titers against ORT by using a commercial enzyme-linked immunosorbent assay (ELISA) kit performed. Evaluation of antibody titer against ND virus is performed by hemagglutination inhibition test. Virus isolation with chick embryo eggs 9-11 and RT-PCR method were carried out. A total of 318 serum samples, 135 samples (42.5%) were positive for antibodies to ORT and titer of HI antibodies against NDV in 122 serum samples (38/4%) were 7-10 (log2) and 61 serum samples (19/2%) had occurrence antibody titer against Newcastle virus and ORT. Results of the present study indicated that 20 tissue samples were positive in embryonated egg and in rapid hemagglutination (HA) test. HI test with specific ND positive serum confirmed that 6 of 20 samples. PCR confirmed that all six samples were positive and PCR products of samples indicated 535-base pair fragments in electrophrosis. Due to the great economic importance of these two diseases in the poultry industry, it is necessary to design and implement a comprehensive plan for prevention and control of these diseases.
Keywords: ELISA, Newcastle disease, Ornithobacterium rhinotracheale, seroprevalence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1410135 The Design of Multiple Detection Parallel Combined Spread Spectrum Communication System
Authors: Lixin Tian, Wei Xue
Abstract:
Many jobs in society go underground, such as mine mining, tunnel construction and subways, which are vital to the development of society. Once accidents occur in these places, the interruption of traditional wired communication is not conducive to the development of rescue work. In order to realize the positioning, early warning and command functions of underground personnel and improve rescue efficiency, it is necessary to develop and design an emergency ground communication system. It is easy to be subjected to narrowband interference when performing conventional underground communication. Spreading communication can be used for this problem. However, general spread spectrum methods such as direct spread communication are inefficient, so it is proposed to use parallel combined spread spectrum (PCSS) communication to improve efficiency. The PCSS communication not only has the anti-interference ability and the good concealment of the traditional spread spectrum system, but also has a relatively high frequency band utilization rate and a strong information transmission capability. So, this technology has been widely used in practice. This paper presents a PCSS communication model-multiple detection parallel combined spread spectrum (MDPCSS) communication system. In this paper, the principle of MDPCSS communication system is described, that is, the sequence at the transmitting end is processed in blocks and cyclically shifted to facilitate multiple detection at the receiving end. The block diagrams of the transmitter and receiver of the MDPCSS communication system are introduced. At the same time, the calculation formula of the system bit error rate (BER) is introduced, and the simulation and analysis of the BER of the system are completed. By comparing with the common parallel PCSS communication, we can draw a conclusion that it is indeed possible to reduce the BER and improve the system performance. Furthermore, the influence of different pseudo-code lengths selected on the system BER is simulated and analyzed, and the conclusion is that the larger the pseudo-code length is, the smaller the system error rate is.
Keywords: Cyclic shift, multiple detection, parallel combined spread spectrum, PN code.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 554134 Impact of Ownership Structure on Provision of Staff and Infrastructure for Implementing Computer Aided Design Curriculum in Universities in South-East Nigeria
Authors: Kelechi E. Ezeji
Abstract:
Instruction towards acquiring skills in the use of Computer Aided Design technologies has become a vital part of architectural education curriculum in the digital era. Its implementation, however, requires deployment of extra resources to build new infrastructure, acquisition and maintenance of new equipment, retraining of staff and recruitment of new ones who are knowledgeable in this area. This study sought to examine the impact that ownership structure of Nigerian universities had on provision of staff and infrastructure for implementing computer aided design curriculum with a view to developing a framework for the evaluation for appropriate implementation by the institutions. Survey research design was employed. The focus was on departments of architecture in universities in south-east Nigeria accredited by the National Universities Commission. Data were obtained in the areas of infrastructure and personnel for CAD implementation. A multi-stage stratified random sampling method was adopted. The first stage of stratification involved the accredited departments. Random sampling by balloting was then carried out. At the second stage, sampling size formulae was applied to obtain respondents’ number. For data analysis, analysis of variance tool for testing differences of means was used. With ρ < 0.5, the study found that there was significant difference between private-funded, state-funded and federal-funded departments of architecture in the provision of personnel and infrastructure. The implications of these findings were that for successful implementation leading to attainment of CAD proficiency to occur in every institution regardless of ownership structure, minimum evaluation guidelines needed to be set. A regular comparison of implementation in institutions was recommended as a means of rating performance. This will inform better interaction with those who consistently show weakness to challenge them towards improvement.
Keywords: Computer-aided design, curriculum, funding, infrastructure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 782133 Depth Camera Aided Dead-Reckoning Localization of Autonomous Mobile Robots in Unstructured Global Navigation Satellite System Denied Environments
Authors: David L. Olson, Stephen B. H. Bruder, Adam S. Watkins, Cleon E. Davis
Abstract:
In global navigation satellite system (GNSS) denied settings, such as indoor environments, autonomous mobile robots are often limited to dead-reckoning navigation techniques to determine their position, velocity, and attitude (PVA). Localization is typically accomplished by employing an inertial measurement unit (IMU), which, while precise in nature, accumulates errors rapidly and severely degrades the localization solution. Standard sensor fusion methods, such as Kalman filtering, aim to fuse precise IMU measurements with accurate aiding sensors to establish a precise and accurate solution. In indoor environments, where GNSS and no other a priori information is known about the environment, effective sensor fusion is difficult to achieve, as accurate aiding sensor choices are sparse. However, an opportunity arises by employing a depth camera in the indoor environment. A depth camera can capture point clouds of the surrounding floors and walls. Extracting attitude from these surfaces can serve as an accurate aiding source, which directly combats errors that arise due to gyroscope imperfections. This configuration for sensor fusion leads to a dramatic reduction of PVA error compared to traditional aiding sensor configurations. This paper provides the theoretical basis for the depth camera aiding sensor method, initial expectations of performance benefit via simulation, and hardware implementation thus verifying its veracity. Hardware implementation is performed on the Quanser Qbot 2™ mobile robot, with a Vector-Nav VN-200™ IMU and Kinect™ camera from Microsoft.
Keywords: Autonomous mobile robotics, dead reckoning, depth camera, inertial navigation, Kalman filtering, localization, sensor fusion.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 720132 Cluster Based Energy Efficient and Fault Tolerant n-Coverage in Wireless Sensor Network
Authors: D. Satish Kumar, N. Nagarajan
Abstract:
Coverage conservation and extend the network lifetime are the primary issues in wireless sensor networks. Due to the large variety of applications, coverage is focus to a wide range of interpretations. The applications necessitate that each point in the area is observed by only one sensor while other applications may require that each point is enclosed by at least sensors (n>1) to achieve fault tolerance. Sensor scheduling activities in existing Transparent and non- Transparent relay modes (T-NT) Mobile Multi-Hop relay networks fails to guarantee area coverage with minimal energy consumption and fault tolerance. To overcome these issues, Cluster based Energy Competent n- coverage scheme called (CEC n-coverage scheme) to ensure the full coverage of a monitored area while saving energy. CEC n-coverage scheme uses a novel sensor scheduling scheme based on the n-density and the remaining energy of each sensor to determine the state of all the deployed sensors to be either active or sleep as well as the state durations. Hence, it is attractive to trigger a minimum number of sensors that are able to ensure coverage area and turn off some redundant sensors to save energy and therefore extend network lifetime. In addition, decisive a smallest amount of active sensors based on the degree coverage required and its level. A variety of numerical parameters are computed using ns2 simulator on existing (T-NT) Mobile Multi-Hop relay networks and CEC n-coverage scheme. Simulation results showed that CEC n-coverage scheme in wireless sensor network provides better performance in terms of the energy efficiency, 6.61% reduced fault tolerant in terms of seconds and the percentage of active sensors to guarantee the area coverage compared to exiting algorithm.
Keywords: Wireless Sensor network, Mobile Multi-Hop relay networks, n-coverage, Cluster based Energy Competent, Transparent and non- Transparent relay modes, Fault Tolerant, sensor scheduling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2148131 Performance Analysis of Organic Rankine Cycle Technology to Exploit Low-Grade Waste Heat to Power Generation in Indian Industry
Authors: Bipul Krishna Saha, Basab Chakraborty, Ashish Alex Sam, Parthasarathi Ghosh
Abstract:
The demand for energy is cumulatively increasing with time. Since the availability of conventional energy resources is dying out gradually, significant interest is being laid on searching for alternate energy resources and minimizing the wastage of energy in various fields. In such perspective, low-grade waste heat from several industrial sources can be reused to generate electricity. The present work is to further the adoption of the Organic Rankine Cycle (ORC) technology in Indian industrial sector. The present paper focuses on extending the previously reported idea to the next level through a comparative review with three different working fluids using practical data from an Indian industrial plant. For comprehensive study in the simulation platform of Aspen Hysys®, v8.6, the waste heat data has been collected from a current coke oven gas plant in India. A parametric analysis of non-regenerative ORC and regenerative ORC is executed using the working fluids R-123, R-11 and R-21 for subcritical ORC system. The primary goal is to determine the optimal working fluid considering various system parameters like turbine work output, obtained system efficiency, irreversibility rate and second law efficiency under applied multiple heat source temperature (160 °C- 180 °C). Selection of the turbo-expanders is one of the most crucial tasks for low-temperature applications in ORC system. The present work is an attempt to make suitable recommendation for the appropriate configuration of the turbine. In a nutshell, this study justifies the proficiency of integrating the ORC technology in Indian perspective and also finds the appropriate parameter of all components integrated in ORC system for building up an ORC prototype.
Keywords: Organic rankine cycle, regenerative organic rankine cycle, waste heat recovery, Indian industry.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1269130 The Determination of Cellulose Spiral Angle by Small-Angle X-Ray Scattering from Structurally Characterized Acacia mangium Cell Wall
Authors: Tamer A. Tabet, Fauziah Abdul Aziz, Shahidan Radiman
Abstract:
The spiral angle of the elementary cellulose fibril in the wood cell wall, often called microfibril angle, (MFA). Microfibril angle in hardwood is one of the key determinants of solid timber performance due to its strong influence on the stiffness, strength, shrinkage, swelling, thermal-dynamics mechanical properties and dimensional stability of wood. Variation of MFA (degree) in the S2 layer of the cell walls among Acacia mangium trees was determined using small-angle X-ray scattering (SAXS). The length and orientation of the microfibrils of the cell walls in the irradiated volume of the thin samples are measured using SAXS and optical microscope for 3D surface measurement. The undetermined parameters in the analysis are the MFA, (M) and the standard deviation (σФ) of the intensity distribution arising from the wandering of the fibril orientation about the mean value. Nine separate pairs of values are determined for nine different values of the angle of the incidence of the X-ray beam relative to the normal to the radial direction in the sample. The results show good agreement. The curve distribution of scattered intensity for the real cell wall structure is compared with that calculated with that assembly of rectangular cells with the same ratio of transverse to radial cell wall length. It is demonstrated that for β = 45°, the peaks in the curve intensity distribution for the real and the rectangular cells coincide. If this peak position is Ф45, then the MFA can be determined from the relation M = tan-1 (tan Ф45 / cos 45°), which is precise for rectangular cells. It was found that 92.93% of the variation of MFA can be attributed to the distance from pith to bark. Here we shall present our results of the MFA in the cell wall with respect to its shape, structure and the distance from pith to park as an important fast check and yet accurate towards the quality of wood, its uses and application.Keywords: Small-Angle X-Ray Scattering, Microfibril Angle, MFA, rectangular cell wall and real cell wall, Acacia mangium.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1808129 The Guideline of Overall Competitive Advantage Promotion with Key Success Paths
Authors: M. F. Wu, F. T. Cheng, C. S. Wu, M. C. Tan
Abstract:
It is a critical time to upgrade technology and increase value added with manufacturing skills developing and management strategies that will highly satisfy the customers need in the precision machinery global market. In recent years, the supply side, each precision machinery manufacturers in each country are facing the pressures of price reducing from the demand side voices that pushes the high-end precision machinery manufacturers adopts low-cost and high-quality strategy to retrieve the market. Because of the trend of the global market, the manufacturers must take price reducing strategies and upgrade technology of low-end machinery for differentiations to consolidate the market.By using six key success factors (KSFs), customer perceived value, customer satisfaction, customer service, product design, product effectiveness and machine structure quality are causal conditions to explore the impact of competitive advantage of the enterprise, such as overall profitability and product pricing power. This research uses key success paths (KSPs) approach and f/s QCA software to explore various combinations of causal relationships, so as to fully understand the performance level of KSFs and business objectives in order to achieve competitive advantage. In this study, the combination of a causal relationships, are called Key Success Paths (KSPs). The key success paths guide the enterprise to achieve the specific outcomes of business. The findings of this study indicate that there are thirteen KSPs to achieve the overall profitability, sixteen KSPs to achieve the product pricing power and seventeen KSPs to achieve both overall profitability and pricing power of the enterprise. The KSPs provide the directions of resources integration and allocation, improve utilization efficiency of limited resources to realize the continuous vision of the enterprise.
Keywords: Precision Machinery Industry, Key Success Factors (KSPs), Key Success Paths (KSPs), Overall Profitability, Product Pricing Power, Competitive Advantages.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1856128 Optimizing Organizational Performance: The Critical Role of Headcount Budgeting in Strategic Alignment and Financial Stability
Authors: Shobhit Mittal
Abstract:
Headcount budgeting stands as a pivotal element in organizational financial management, extending beyond traditional budgeting to encompass strategic resource allocation for workforce-related expenses. This process is integral to maintaining financial stability and fostering a productive workforce, requiring a comprehensive analysis of factors such as market trends, business growth projections, and evolving workforce skill requirements. It demands a collaborative approach, primarily involving Human Resources (HR) and finance departments, to align workforce planning with an organization's financial capabilities and strategic objectives. The dynamic nature of headcount budgeting necessitates continuous monitoring and adjustment in response to economic fluctuations, business strategy shifts, technological advancements, and market dynamics. Its significance in talent management is also highlighted, aligning financial planning with talent acquisition and retention strategies to ensure a competitive edge in the market. The consequences of incorrect headcount budgeting are explored, showing how it can lead to financial strain, operational inefficiencies, and hindered strategic objectives. Examining case studies like IBM's strategic workforce rebalancing and Microsoft's shift for long-term success, the importance of aligning headcount budgeting with organizational goals is underscored. These examples illustrate that effective headcount budgeting transcends its role as a financial tool, emerging as a strategic element crucial for an organization's success. This necessitates continuous refinement and adaptation to align with evolving business goals and market conditions, highlighting its role as a key driver in organizational success and sustainability.
Keywords: Strategic planning, fiscal budget, headcount planning, resource allocation, financial management, decision-making, operational efficiency, risk management, headcount budget.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 159127 Performance Analysis of HSDPA Systems using Low-Density Parity-Check (LDPC)Coding as Compared to Turbo Coding
Authors: K. Anitha Sheela, J. Tarun Kumar
Abstract:
HSDPA is a new feature which is introduced in Release-5 specifications of the 3GPP WCDMA/UTRA standard to realize higher speed data rate together with lower round-trip times. Moreover, the HSDPA concept offers outstanding improvement of packet throughput and also significantly reduces the packet call transfer delay as compared to Release -99 DSCH. Till now the HSDPA system uses turbo coding which is the best coding technique to achieve the Shannon limit. However, the main drawbacks of turbo coding are high decoding complexity and high latency which makes it unsuitable for some applications like satellite communications, since the transmission distance itself introduces latency due to limited speed of light. Hence in this paper it is proposed to use LDPC coding in place of Turbo coding for HSDPA system which decreases the latency and decoding complexity. But LDPC coding increases the Encoding complexity. Though the complexity of transmitter increases at NodeB, the End user is at an advantage in terms of receiver complexity and Bit- error rate. In this paper LDPC Encoder is implemented using “sparse parity check matrix" H to generate a codeword at Encoder and “Belief Propagation algorithm "for LDPC decoding .Simulation results shows that in LDPC coding the BER suddenly drops as the number of iterations increase with a small increase in Eb/No. Which is not possible in Turbo coding. Also same BER was achieved using less number of iterations and hence the latency and receiver complexity has decreased for LDPC coding. HSDPA increases the downlink data rate within a cell to a theoretical maximum of 14Mbps, with 2Mbps on the uplink. The changes that HSDPA enables includes better quality, more reliable and more robust data services. In other words, while realistic data rates are only a few Mbps, the actual quality and number of users achieved will improve significantly.Keywords: AMC, HSDPA, LDPC, WCDMA, 3GPP.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2048126 Optimal Image Compression Based on Sign and Magnitude Coding of Wavelet Coefficients
Authors: Mbainaibeye Jérôme, Noureddine Ellouze
Abstract:
Wavelet transforms is a very powerful tools for image compression. One of its advantage is the provision of both spatial and frequency localization of image energy. However, wavelet transform coefficients are defined by both a magnitude and sign. While algorithms exist for efficiently coding the magnitude of the transform coefficients, they are not efficient for the coding of their sign. It is generally assumed that there is no compression gain to be obtained from the coding of the sign. Only recently have some authors begun to investigate the sign of wavelet coefficients in image coding. Some authors have assumed that the sign information bit of wavelet coefficients may be encoded with the estimated probability of 0.5; the same assumption concerns the refinement information bit. In this paper, we propose a new method for Separate Sign Coding (SSC) of wavelet image coefficients. The sign and the magnitude of wavelet image coefficients are examined to obtain their online probabilities. We use the scalar quantization in which the information of the wavelet coefficient to belong to the lower or to the upper sub-interval in the uncertainly interval is also examined. We show that the sign information and the refinement information may be encoded by the probability of approximately 0.5 only after about five bit planes. Two maps are separately entropy encoded: the sign map and the magnitude map. The refinement information of the wavelet coefficient to belong to the lower or to the upper sub-interval in the uncertainly interval is also entropy encoded. An algorithm is developed and simulations are performed on three standard images in grey scale: Lena, Barbara and Cameraman. Five scales are performed using the biorthogonal wavelet transform 9/7 filter bank. The obtained results are compared to JPEG2000 standard in terms of peak signal to noise ration (PSNR) for the three images and in terms of subjective quality (visual quality). It is shown that the proposed method outperforms the JPEG2000. The proposed method is also compared to other codec in the literature. It is shown that the proposed method is very successful and shows its performance in term of PSNR.
Keywords: Image compression, wavelet transform, sign coding, magnitude coding.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1672125 The Effect of Cyclic Speed on the Wear Properties of Molybdenum Disulfide Greases under Extreme Pressure Loading Using 4 Balls Wear Tests
Authors: Gabi Nehme
Abstract:
The relationship between different types of Molybdenum disulfide greases under extreme pressure loading and different speed situations have been studied using Design of Experiment (DOE) under 1200rpm steady state rotational speed and cyclic frequencies between 2400 and 1200rpm using a Plint machine software to set up the different rotational speed situations. Research described here is aimed at providing good friction and wear performance while optimizing cyclic frequencies and MoS2 concentration due to the recent concern about grease behavior in extreme pressure applications. Extreme load of 785 Newton was used in conjunction with different cyclic frequencies (2400rpm -3.75min, 1200rpm -7.5min, 2400rpm -3.75min, 1200rpm -7.5min), to examine lithium based grease with and without MoS2 for equal number of revolutions, and a total run of 36000 revolutions; then compared to 1200rpm steady speed for the same total number of revolutions. 4 Ball wear tester was utilized to run large number of experiments randomly selected by the DOE software. The grease was combined with fine grade MoS2 or technical grade then heated to 750C and the wear scar width was collected at the end of each test. DOE model validation results verify that the data were very significant and can be applied to a wide range of extreme pressure applications. Based on simulation results and Scanning Electron images (SEM), it has been found that wear was largely dependent on the cyclic frequency condition. It is believed that technical grade MoS2 greases under faster cyclic speeds perform better and provides antiwear film that can resist extreme pressure loadings. Figures showed reduced wear scars width and improved frictional values.
Keywords: MoS2 grease, wear, friction, extreme load, cyclic frequencies, aircraft grade bearing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1898124 Advantages of Neural Network Based Air Data Estimation for Unmanned Aerial Vehicles
Authors: Angelo Lerro, Manuela Battipede, Piero Gili, Alberto Brandl
Abstract:
Redundancy requirements for UAV (Unmanned Aerial Vehicle) are hardly faced due to the generally restricted amount of available space and allowable weight for the aircraft systems, limiting their exploitation. Essential equipment as the Air Data, Attitude and Heading Reference Systems (ADAHRS) require several external probes to measure significant data as the Angle of Attack or the Sideslip Angle. Previous research focused on the analysis of a patented technology named Smart-ADAHRS (Smart Air Data, Attitude and Heading Reference System) as an alternative method to obtain reliable and accurate estimates of the aerodynamic angles. This solution is based on an innovative sensor fusion algorithm implementing soft computing techniques and it allows to obtain a simplified inertial and air data system reducing external devices. In fact, only one external source of dynamic and static pressures is needed. This paper focuses on the benefits which would be gained by the implementation of this system in UAV applications. A simplification of the entire ADAHRS architecture will bring to reduce the overall cost together with improved safety performance. Smart-ADAHRS has currently reached Technology Readiness Level (TRL) 6. Real flight tests took place on ultralight aircraft equipped with a suitable Flight Test Instrumentation (FTI). The output of the algorithm using the flight test measurements demonstrates the capability for this fusion algorithm to embed in a single device multiple physical and virtual sensors. Any source of dynamic and static pressure can be integrated with this system gaining a significant improvement in terms of versatility.Keywords: Neural network, aerodynamic angles, virtual sensor, unmanned aerial vehicle, air data system, flight test.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1023123 Collaborative Stylistic Group Project: A Drama Practical Analysis Application
Authors: Omnia F. Elkommos
Abstract:
In the course of teaching stylistics to undergraduate students of the Department of English Language and Literature, Faculty of Arts and Humanities, the linguistic tool kit of theories comes in handy and useful for the better understanding of the different literary genres: Poetry, drama, and short stories. In the present paper, a model of teaching of stylistics is compiled and suggested. It is a collaborative group project technique for use in the undergraduate diverse specialisms (Literature, Linguistics and Translation tracks) class. Students initially are introduced to the different linguistic tools and theories suitable for each literary genre. The second step is to apply these linguistic tools to texts. Students are required to watch videos performing the poems or play, for example, and search the net for interpretations of the texts by other authorities. They should be using a template (prepared by the researcher) that has guided questions leading students along in their analysis. Finally, a practical analysis would be written up using the practical analysis essay template (also prepared by the researcher). As per collaborative learning, all the steps include activities that are student-centered addressing differentiation and considering their three different specialisms. In the process of selecting the proper tools, the actual application and analysis discussion, students are given tasks that request their collaboration. They also work in small groups and the groups collaborate in seminars and group discussions. At the end of the course/module, students present their work also collaboratively and reflect and comment on their learning experience. The module/course uses a drama play that lends itself to the task: ‘The Bond’ by Amy Lowell and Robert Frost. The project results in an interpretation of its theme, characterization and plot. The linguistic tools are drawn from pragmatics, and discourse analysis among others.
Keywords: Applied linguistic theories, collaborative learning, cooperative principle, discourse analysis, drama analysis, group project, online acting performance, pragmatics, speech act theory, stylistics, technology enhanced learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1079122 Performance Evaluation and Plugging Characteristics of Controllable Self-Aggregating Colloidal Particle Profile Control Agent
Authors: Zhiguo Yang, Xiangan Yue, Minglu Shao, Yang Yue, Tianqi Yue
Abstract:
In low permeability reservoirs, the reservoir pore throat is small and the micro heterogeneity is prominent. Conventional microsphere profile control agents generally have good injectability but poor plugging effect; however, profile control agents with good plugging effect generally have poor injectability, which makes it difficult for agent to realize deep profile control of reservoir. To solve this problem, styrene and acrylamide were used as monomers in the laboratory. Emulsion polymerization was used to prepare the Controllable Self-Aggregating Colloidal Particle (CSA), which was rich in amide group. The CSA microsphere dispersion solution with a particle diameter smaller than the pore throat diameter was injected into the reservoir to ensure that the profile control agent had good inject ability. After dispersing the CSA microsphere to the deep part of the reservoir, the CSA microspheres dispersed in static for a certain period of time will self-aggregate into large-sized particle clusters to achieve plugging of hypertonic channels. The CSA microsphere has the characteristics of low expansion and avoids shear fracture in the process of migration. It can be observed by transmission electron microscope that CSA microspheres still maintain regular and uniform spherical and core-shell heterogeneous structure after aging at 100 ºC for 35 days, and CSA microspheres have good thermal stability. The results of bottle test showed that with the increase of cation concentration, the aggregation time of CSA microspheres gradually shortened, and the influence of divalent cations was greater than that of monovalent ions. Physical simulation experiments show that CSA microspheres have good injectability, and the aggregated CSA particle clusters can produce effective plugging and migrate to the deep part of the reservoir for profile control.
Keywords: Heterogeneous reservoir, deep profile control, emulsion polymerization, colloidal particles, plugging characteristic.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 486