Search results for: real estate prediction
1514 Emotion Detection in Twitter Messages Using Combination of Long Short-Term Memory and Convolutional Deep Neural Networks
Authors: Bahareh Golchin, Nooshin Riahi
Abstract:
One of the most significant issues as attended a lot in recent years is that of recognizing the sentiments and emotions in social media texts. The analysis of sentiments and emotions is intended to recognize the conceptual information such as the opinions, feelings, attitudes and emotions of people towards the products, services, organizations, people, topics, events and features in the written text. These indicate the greatness of the problem space. In the real world, businesses and organizations are always looking for tools to gather ideas, emotions, and directions of people about their products, services, or events related to their own. This article uses the Twitter social network, one of the most popular social networks with about 420 million active users, to extract data. Using this social network, users can share their information and opinions about personal issues, policies, products, events, etc. It can be used with appropriate classification of emotional states due to the availability of its data. In this study, supervised learning and deep neural network algorithms are used to classify the emotional states of Twitter users. The use of deep learning methods to increase the learning capacity of the model is an advantage due to the large amount of available data. Tweets collected on various topics are classified into four classes using a combination of two Bidirectional Long Short Term Memory network and a Convolutional network. The results obtained from this study with an average accuracy of 93%, show good results extracted from the proposed framework and improved accuracy compared to previous work.Keywords: emotion classification, sentiment analysis, social networks, deep neural networks
Procedia PDF Downloads 1371513 Energy Performance Gaps in Residences: An Analysis of the Variables That Cause Energy Gaps and Their Impact
Authors: Amrutha Kishor
Abstract:
Today, with the rising global warming and depletion of resources every industry is moving toward sustainability and energy efficiency. As part of this movement, it is nowadays obligatory for architects to play their part by creating energy predictions for their designs. But in a lot of cases, these predictions do not reflect the real quantities of energy in newly built buildings when operating. These can be described as ‘Energy Performance Gaps’. This study aims to determine the underlying reasons for these gaps. Seven houses designed by Allan Joyce Architects, UK from 1998 until 2019 were considered for this study. The data from the residents’ energy bills were cross-referenced with the predictions made with the software SefairaPro and from energy reports. Results indicated that the predictions did not match the actual energy usage. An account of how energy was used in these seven houses was made by means of personal interviews. The main factors considered in the study were occupancy patterns, heating systems and usage, lighting profile and usage, and appliances’ profile and usage. The study found that the main reasons for the creation of energy gaps were the discrepancies in occupant usage and patterns of energy consumption that are predicted as opposed to the actual ones. This study is particularly useful for energy-conscious architectural firms to fine-tune the approach to designing houses and analysing their energy performance. As the findings reveal that energy usage in homes varies based on the way residents use the space, it helps deduce the most efficient technological combinations. This information can be used to set guidelines for future policies and regulations related to energy consumption in homes. This study can also be used by the developers of simulation software to understand how architects use their product and drive improvements in its future versions.Keywords: architectural simulation, energy efficient design, energy performance gaps, environmental design
Procedia PDF Downloads 1181512 DeepNIC a Method to Transform Each Tabular Variable into an Independant Image Analyzable by Basic CNNs
Authors: Nguyen J. M., Lucas G., Ruan S., Digonnet H., Antonioli D.
Abstract:
Introduction: Deep Learning (DL) is a very powerful tool for analyzing image data. But for tabular data, it cannot compete with machine learning methods like XGBoost. The research question becomes: can tabular data be transformed into images that can be analyzed by simple CNNs (Convolutional Neuron Networks)? Will DL be the absolute tool for data classification? All current solutions consist in repositioning the variables in a 2x2 matrix using their correlation proximity. In doing so, it obtains an image whose pixels are the variables. We implement a technology, DeepNIC, that offers the possibility of obtaining an image for each variable, which can be analyzed by simple CNNs. Material and method: The 'ROP' (Regression OPtimized) model is a binary and atypical decision tree whose nodes are managed by a new artificial neuron, the Neurop. By positioning an artificial neuron in each node of the decision trees, it is possible to make an adjustment on a theoretically infinite number of variables at each node. From this new decision tree whose nodes are artificial neurons, we created the concept of a 'Random Forest of Perfect Trees' (RFPT), which disobeys Breiman's concepts by assembling very large numbers of small trees with no classification errors. From the results of the RFPT, we developed a family of 10 statistical information criteria, Nguyen Information Criterion (NICs), which evaluates in 3 dimensions the predictive quality of a variable: Performance, Complexity and Multiplicity of solution. A NIC is a probability that can be transformed into a grey level. The value of a NIC depends essentially on 2 super parameters used in Neurops. By varying these 2 super parameters, we obtain a 2x2 matrix of probabilities for each NIC. We can combine these 10 NICs with the functions AND, OR, and XOR. The total number of combinations is greater than 100,000. In total, we obtain for each variable an image of at least 1166x1167 pixels. The intensity of the pixels is proportional to the probability of the associated NIC. The color depends on the associated NIC. This image actually contains considerable information about the ability of the variable to make the prediction of Y, depending on the presence or absence of other variables. A basic CNNs model was trained for supervised classification. Results: The first results are impressive. Using the GSE22513 public data (Omic data set of markers of Taxane Sensitivity in Breast Cancer), DEEPNic outperformed other statistical methods, including XGBoost. We still need to generalize the comparison on several databases. Conclusion: The ability to transform any tabular variable into an image offers the possibility of merging image and tabular information in the same format. This opens up great perspectives in the analysis of metadata.Keywords: tabular data, CNNs, NICs, DeepNICs, random forest of perfect trees, classification
Procedia PDF Downloads 1251511 Enhance Construction Visual As-Built Schedule Management Using BIM Technology
Authors: Shu-Hui Jan, Hui-Ping Tserng, Shih-Ping Ho
Abstract:
Construction project control attempts to obtain real-time as-built schedule information and to eliminate project delays by effectively enhancing dynamic schedule control and management. Suitable platforms for enhancing an as-built schedule visually during the construction phase are necessary and important for general contractors. As the application of building information modeling (BIM) becomes more common, schedule management integrated with the BIM approach becomes essential to enhance visual construction management implementation for the general contractor during the construction phase. To enhance visualization of the updated as-built schedule for the general contractor, this study presents a novel system called the Construction BIM-assisted Schedule Management (ConBIM-SM) system for general contractors in
Keywords: building information modeling (BIM), construction schedule management, as-built schedule management, BIM schedule updating mechanism
Procedia PDF Downloads 3751510 Finite Element Analysis of the Anaconda Device: Efficiently Predicting the Location and Shape of a Deployed Stent
Authors: Faidon Kyriakou, William Dempster, David Nash
Abstract:
Abdominal Aortic Aneurysm (AAA) is a major life-threatening pathology for which modern approaches reduce the need for open surgery through the use of stenting. The success of stenting though is sometimes jeopardized by the final position of the stent graft inside the human artery which may result in migration, endoleaks or blood flow occlusion. Herein, a finite element (FE) model of the commercial medical device AnacondaTM (Vascutek, Terumo) has been developed and validated in order to create a numerical tool able to provide useful clinical insight before the surgical procedure takes place. The AnacondaTM device consists of a series of NiTi rings sewn onto woven polyester fabric, a structure that despite its column stiffness is flexible enough to be used in very tortuous geometries. For the purposes of this study, a FE model of the device was built in Abaqus® (version 6.13-2) with the combination of beam, shell and surface elements; the choice of these building blocks was made to keep the computational cost to a minimum. The validation of the numerical model was performed by comparing the deployed position of a full stent graft device inside a constructed AAA with a duplicate set-up in Abaqus®. Specifically, an AAA geometry was built in CAD software and included regions of both high and low tortuosity. Subsequently, the CAD model was 3D printed into a transparent aneurysm, and a stent was deployed in the lab following the steps of the clinical procedure. Images on the frontal and sagittal planes of the experiment allowed the comparison with the results of the numerical model. By overlapping the experimental and computational images, the mean and maximum distances between the rings of the two models were measured in the longitudinal, and the transverse direction and, a 5mm upper bound was set as a limit commonly used by clinicians when working with simulations. The two models showed very good agreement of their spatial positioning, especially in the less tortuous regions. As a result, and despite the inherent uncertainties of a surgical procedure, the FE model allows confidence that the final position of the stent graft, when deployed in vivo, can also be predicted with significant accuracy. Moreover, the numerical model run in just a few hours, an encouraging result for applications in the clinical routine. In conclusion, the efficient modelling of a complicated structure which combines thin scaffolding and fabric has been demonstrated to be feasible. Furthermore, the prediction capabilities of the location of each stent ring, as well as the global shape of the graft, has been shown. This can allow surgeons to better plan their procedures and medical device manufacturers to optimize their designs. The current model can further be used as a starting point for patient specific CFD analysis.Keywords: AAA, efficiency, finite element analysis, stent deployment
Procedia PDF Downloads 1911509 Enablers and Inhibitors of Effective Waste Management Measures in Informal Settlements in South Africa: A Case of Alaska
Authors: Lynda C. Mbadugha, Bankole Awuzie, Kwanda Khumalo, Lindokuhle Matsebula, Masenoke Kgaditsi
Abstract:
Inadequate waste management remains a fundamental issue in the majority of cities around the globe, but it becomes a threat when it concerns informal settlements. Although studies have evaluated the performance of waste management measures, only a few have addressed that with a focus on South African informal settlements and the reasons for their apparent ineffectiveness in such locations. However, there may be evidence of variations in the extant problems due to the uniqueness of each location and the factors influencing the performance. Thus, there is a knowledge deficit regarding implementing waste management measures in South African informal settlements. This study seeks to evaluate the efficacy of waste management measures in the Alaska informal settlement in South Africa to assess the previously collected data of other areas using the degree of correlation. The research investigated a real-world scenario in the specified location using a case study approach and multiple data sources. The findings described various waste management practices used in Alaska's informal settlements; however, a correlation was found between the performance of these measures and those already used. The observed differences are primarily attributable to the physical characteristics of the locations, the lack of understanding of the environmental and health consequences of careless waste disposal, and the negative attitudes of the residents toward waste management practices. This study elucidates waste management implementation in informal settlements. It contributes to the relevant bodies of knowledge by describing these practices in South Africa. This paper's practical value emphasizes the general waste management characteristics of South Africa's informal settlements to facilitate the planning and provision of necessary interventions. The study concludes that the enablers and inhibitors are mainly political, behavioral, and environmental concerns.Keywords: factors, informal settlement, performance, waste management
Procedia PDF Downloads 931508 Budgetary Performance Model for Managing Pavement Maintenance
Authors: Vivek Hokam, Vishrut Landge
Abstract:
An ideal maintenance program for an industrial road network is one that would maintain all sections at a sufficiently high level of functional and structural conditions. However, due to various constraints such as budget, manpower and equipment, it is not possible to carry out maintenance on all the needy industrial road sections within a given planning period. A rational and systematic priority scheme needs to be employed to select and schedule industrial road sections for maintenance. Priority analysis is a multi-criteria process that determines the best ranking list of sections for maintenance based on several factors. In priority setting, difficult decisions are required to be made for selection of sections for maintenance. It is more important to repair a section with poor functional conditions which includes uncomfortable ride etc. or poor structural conditions i.e. sections those are in danger of becoming structurally unsound. It would seem therefore that any rational priority setting approach must consider the relative importance of functional and structural condition of the section. The maintenance priority index and pavement performance models tend to focus mainly on the pavement condition, traffic criteria etc. There is a need to develop the model which is suitably used with respect to limited budget provisions for maintenance of pavement. Linear programming is one of the most popular and widely used quantitative techniques. A linear programming model provides an efficient method for determining an optimal decision chosen from a large number of possible decisions. The optimum decision is one that meets a specified objective of management, subject to various constraints and restrictions. The objective is mainly minimization of maintenance cost of roads in industrial area. In order to determine the objective function for analysis of distress model it is necessary to fix the realistic data into a formulation. Each type of repair is to be quantified in a number of stretches by considering 1000 m as one stretch. A stretch considered under study is having 3750 m length. The quantity has to be put into an objective function for maximizing the number of repairs in a stretch related to quantity. The distress observed in this stretch are potholes, surface cracks, rutting and ravelling. The distress data is measured manually by observing each distress level on a stretch of 1000 m. The maintenance and rehabilitation measured that are followed currently are based on subjective judgments. Hence, there is a need to adopt a scientific approach in order to effectively use the limited resources. It is also necessary to determine the pavement performance and deterioration prediction relationship with more accurate and economic benefits of road networks with respect to vehicle operating cost. The infrastructure of road network should have best results expected from available funds. In this paper objective function for distress model is determined by linear programming and deterioration model considering overloading is discussed.Keywords: budget, maintenance, deterioration, priority
Procedia PDF Downloads 2071507 Analysis of the Learning Effectiveness of the Steam-6e Course: A Case Study on the Development of Virtual Idol Product Design as an Example
Authors: Mei-Chun. Chang
Abstract:
STEAM (Science, Technology, Engineering, Art, and Mathematics) represents a cross-disciplinary and learner-centered teaching model that cultivates students to link theory with the presentation of real situations, thereby improving their various abilities. This study explores students' learning performance after using the 6E model in STEAM teaching for a professional course in the digital media design department of technical colleges, as well as the difficulties and countermeasures faced by STEAM curriculum design and its implementation. In this study, through industry experts’ work experience, activity exchanges, course teaching, and experience, learners can think about the design and development value of virtual idol products that meet the needs of users and to employ AR/VR technology to innovate their product applications. Applying action research, the investigation has 35 junior students from the department of digital media design of the school where the researcher teaches as the research subjects. The teaching research was conducted over two stages spanning ten weeks and 30 sessions. This research collected the data and conducted quantitative and qualitative data sorting analyses through ‘design draft sheet’, ‘student interview record’, ‘STEAM Product Semantic Scale’, and ‘Creative Product Semantic Scale (CPSS)’. Research conclusions are presented, and relevant suggestions are proposed as a reference for teachers or follow-up researchers. The contribution of this study is to teach college students to develop original virtual idols and product designs, improve learning effectiveness through STEAM teaching activities, and effectively cultivate innovative and practical cross-disciplinary design talents.Keywords: STEAM, 6E model, virtual idol, learning effectiveness, practical courses
Procedia PDF Downloads 1261506 Complex Network Approach to International Trade of Fossil Fuel
Authors: Semanur Soyyigit Kaya, Ercan Eren
Abstract:
Energy has a prominent role for development of nations. Countries which have energy resources also have strategic power in the international trade of energy since it is essential for all stages of production in the economy. Thus, it is important for countries to analyze the weakness and strength of the system. On the other side, it is commonly believed that international trade has complex network properties. Complex network is a tool for the analysis of complex systems with heterogeneous agents and interaction between them. A complex network consists of nodes and the interactions between these nodes. Total properties which emerge as a result of these interactions are distinct from the sum of small parts (more or less) in complex systems. Thus, standard approaches to international trade are superficial to analyze these systems. Network analysis provides a new approach to analyze international trade as a network. In this network countries constitute nodes and trade relations (export or import) constitute edges. It becomes possible to analyze international trade network in terms of high degree indicators which are specific to complex systems such as connectivity, clustering, assortativity/disassortativity, centrality, etc. In this analysis, international trade of crude oil and coal which are types of fossil fuel has been analyzed from 2005 to 2014 via network analysis. First, it has been analyzed in terms of some topological parameters such as density, transitivity, clustering etc. Afterwards, fitness to Pareto distribution has been analyzed. Finally, weighted HITS algorithm has been applied to the data as a centrality measure to determine the real prominence of countries in these trade networks. Weighted HITS algorithm is a strong tool to analyze the network by ranking countries with regards to prominence of their trade partners. We have calculated both an export centrality and an import centrality by applying w-HITS algorithm to data.Keywords: complex network approach, fossil fuel, international trade, network theory
Procedia PDF Downloads 3361505 In Stemming Out Societal Depravity: Existentialism, Realism, and Contrapuntal Criticism in Nigerian Arabic Poetry: Ibn Yusuf’s Anthology as Paradigm
Authors: Izzudeen Adetunji
Abstract:
The intrinsic nexus between man and society is apparently unknown to many people despite understanding the real responsibility and immense roles in society. Amongst the in-depth roles of a man as an agent of the societal reformer is to be a driven force towards installing normalcy and socio-cultural change in society. The paradoxical attitudes of man in engaging in social vices, illicit characters, and unwanted attitudes have given birth to decay and ill-society. However, the need for social change or socio-cultural evolution might be necessary to install normalcy and social order. Nigerian Arabic poets since the 19th century have tremendously engaged in utilizing their poetry for social change through socio-cultural, religious, economic, scientific, or technological forces. This engagement has hitherto yielded a positive outcome for societal reform. The anthology of Ibn Yusuf is one of the compendiums of poetries revealing societal depravity, man’s social vices, and atrocities; which later called to flawlessness. The theoretical framework would be examined through the Heraclitan model, focusing on a parallel to that of a living organism, which, in order to remain alive, must constantly change. Therefore, the thrust of this paper is to examine the societal maladies as elucidated in Ibn Yusuf’s anthology and proffer a contrapuntal criticism of it. Before delving into the main discussion, the paper will examine the concepts of existentialism and realism as a philosophical interface. Likewise, the issues of man and social change, an overview of Nigerian Arabic poetry, will be discussed. Ibn Yusuf’s biography and scholarship and the review of his anthology will be studied. The paper will conclude by critically examining the contrapuntal criticism of societal maladies through Ibn Yusuf’s anthology.Keywords: societal depravity, existentialism, realism, Nigeria Arabic poetry, Ibn Yusuf’s anthology, contrapuntal criticism
Procedia PDF Downloads 261504 Direct Electrical Communication of Redox Enzyme Based on 3-Dimensional Cross-Linked Redox Enzyme/Nanomaterials
Authors: A. K. M. Kafi, S. N. Nina, Mashitah M. Yusoff
Abstract:
In this work, we have described a new 3-dimensional (3D) network of cross-linked Horseradish Peroxidase/Carbon Nanotube (HRP/CNT) on a thiol-modified Au surface in order to build up the effective electrical wiring of the enzyme units with the electrode. This was achieved by the electropolymerization of aniline-functionalized carbon nanotubes (CNTs) and 4-aminothiophenol -modified-HRP on a 4-aminothiophenol monolayer-modified Au electrode. The synthesized 3D HRP/CNT networks were characterized with cyclic voltammetry and amperometry, resulting the establishment direct electron transfer between the redox active unit of HRP and the Au surface. Electrochemical measurements reveal that the immobilized HRP exhibits high biological activity and stability and a quasi-reversible redox peak of the redox center of HRP was observed at about −0.355 and −0.275 V vs. Ag/AgCl. The electron transfer rate constant, KS and electron transfer co-efficient were found to be 0.57 s-1 and 0.42, respectively. Based on the electrocatalytic process by direct electrochemistry of HRP, a biosensor for detecting H2O2 was developed. The developed biosensor exhibits excellent electrocatalytic activity for the reduction of H2O2. The proposed biosensor modified with HRP/CNT 3D network displays a broader linear range and a lower detection limit for H2O2 determination. The linear range is from 1.0×10−7 to 1.2×10−4M with a detection limit of 2.2.0×10−8M at 3σ. Moreover, this biosensor exhibits very high sensitivity, good reproducibility and long-time stability. In summary, ease of fabrication, a low cost, fast response and high sensitivity are the main advantages of the new biosensor proposed in this study. These obvious advantages would really help for the real analytical applicability of the proposed biosensor.Keywords: redox enzyme, nanomaterials, biosensors, electrical communication
Procedia PDF Downloads 4541503 Continuous and Discontinuos Modeling of Wellbore Instability in Anisotropic Rocks
Authors: C. Deangeli, P. Obentaku Obenebot, O. Omwanghe
Abstract:
The study focuses on the analysis of wellbore instability in rock masses affected by weakness planes. The occurrence of failure in such a type of rocks can occur in the rock matrix and/ or along the weakness planes, in relation to the mud weight gradient. In this case the simple Kirsch solution coupled with a failure criterion cannot supply a suitable scenario for borehole instabilities. Two different numerical approaches have been used in order to investigate the onset of local failure at the wall of a borehole. For each type of approach the influence of the inclination of weakness planes has been investigates, by considering joint sets at 0°, 35° and 90° to the horizontal. The first set of models have been carried out with FLAC 2D (Fast Lagrangian Analysis of Continua) by considering the rock material as a continuous medium, with a Mohr Coulomb criterion for the rock matrix and using the ubiquitous joint model for accounting for the presence of the weakness planes. In this model yield may occur in either the solid or along the weak plane, or both, depending on the stress state, the orientation of the weak plane and the material properties of the solid and weak plane. The second set of models have been performed with PFC2D (Particle Flow code). This code is based on the Discrete Element Method and considers the rock material as an assembly of grains bonded by cement-like materials, and pore spaces. The presence of weakness planes is simulated by the degradation of the bonds between grains along given directions. In general the results of the two approaches are in agreement. However the discrete approach seems to capture more complex phenomena related to local failure in the form of grain detachment at wall of the borehole. In fact the presence of weakness planes in the discontinuous medium leads to local instability along the weak planes also in conditions not predicted from the continuous solution. In general slip failure locations and directions do not follow the conventional wellbore breakout direction but depend upon the internal friction angle and the orientation of the bedding planes. When weakness plane is at 0° and 90° the behaviour are similar to that of a continuous rock material, but borehole instability is more severe when weakness planes are inclined at an angle between 0° and 90° to the horizontal. In conclusion, the results of the numerical simulations show that the prediction of local failure at the wall of the wellbore cannot disregard the presence of weakness planes and consequently the higher mud weight required for stability for any specific inclination of the joints. Despite the discrete approach can simulate smaller areas because of the large number of particles required for the generation of the rock material, however it seems to investigate more correctly the occurrence of failure at the miscroscale and eventually the propagation of the failed zone to a large portion of rock around the wellbore.Keywords: continuous- discontinuous, numerical modelling, weakness planes wellbore, FLAC 2D
Procedia PDF Downloads 4991502 Numerical and Experimental Investigation of Distance Between Fan and Coil Block in a Fin and Tube Air Cooler Heat Exchanger
Authors: Feyza Şahi̇n, Harun Deni̇zli̇, Mustafa Zabun, Hüseyi̇n OnbaşIoğli
Abstract:
Heat exchangers are devices that are widely used to transfer heat between fluids due to their temperature differences. As a type of heat exchanger, air coolers are heat exchangers that cool the air as it passes through the fins of the heat exchanger by transferring heat to the refrigerant in the coil tubes of the heat exchanger. An assembled fin and tube heat exchanger consists of a coil block and a casing with a fan mounted on it. The term “Fan hood” is used to define the distance between the fan and the coil block. Air coolers play a crucial role in cooling systems, and their heat transfer performance can vary depending on design parameters. These parameters can be related to the air side or the internal fluid side. For airside efficiency, the distance between the fan and the coil block affects the performance by creating dead zones at the corners of the casing and maldistribution of airflow. Therefore, a detailed study of the effect of the fan hood on the evaporator and the optimum fan hood distance is necessary for an efficient air cooler design. This study aims to investigate the value of the fan hood in a fin and tube-type air cooler heat exchanger through computational fluid dynamics (CFD) simulations and experimental investigations. CFD simulations will be used to study the airflow within the fan hood. These simulations will provide valuable insights to optimize the design of the fan hood. In addition, experimental tests will be carried out to validate the CFD results and to measure the performance of the fan hood under real conditions. The results will help us to understand the effect of fan hood design on evaporator efficiency and contribute to the development of more efficient cooling systems. This study will provide essential information for evaporator design and improving the energy efficiency of cooling systems.Keywords: heat exchanger, fan hood, heat exchanger performance, air flow performance
Procedia PDF Downloads 771501 Close-Range Remote Sensing Techniques for Analyzing Rock Discontinuity Properties
Authors: Sina Fatolahzadeh, Sergio A. Sepúlveda
Abstract:
This paper presents advanced developments in close-range, terrestrial remote sensing techniques to enhance the characterization of rock masses. The study integrates two state-of-the-art laser-scanning technologies, the HandySCAN and GeoSLAM laser scanners, to extract high-resolution geospatial data for rock mass analysis. These instruments offer high accuracy, precision, low acquisition time, and high efficiency in capturing intricate geological features in small to medium size outcrops and slope cuts. Using the HandySCAN and GeoSLAM laser scanners facilitates real-time, three-dimensional mapping of rock surfaces, enabling comprehensive assessments of rock mass characteristics. The collected data provide valuable insights into structural complexities, surface roughness, and discontinuity patterns, which are essential for geological and geotechnical analyses. The synergy of these advanced remote sensing technologies contributes to a more precise and straightforward understanding of rock mass behavior. In this case, the main parameters of RQD, joint spacing, persistence, aperture, roughness, infill, weathering, water condition, and joint orientation in a slope cut along the Sea-to-Sky Highway, BC, were remotely analyzed to calculate and evaluate the Rock Mass Rating (RMR) and Geological Strength Index (GSI) classification systems. Automatic and manual analyses of the acquired data are then compared with field measurements. The results show the usefulness of the proposed remote sensing methods and their appropriate conformity with the actual field data.Keywords: remote sensing, rock mechanics, rock engineering, slope stability, discontinuity properties
Procedia PDF Downloads 661500 Performance of an Automotive Engine Running on Gasoline-Condensate Blends
Authors: Md. Ehsan, Cyrus Ashok Arupratan Atis
Abstract:
Significantly lower cost, bulk availability, absence of identification color additives and relative ease of mixing with fuels have made gas-field condensates a lucrative option as adulterant for gasoline in Bangladesh. Widespread adulteration of fuels with gas-field condensates being a problem existing mainly in developing countries like Bangladesh, Nigeria etc., research works regarding the effect of such fuel adulteration are very limited. Since the properties of the gas-field condensate vary widely depending on geographical location, studies need to be based on local condensate feeds. This study quantitatively evaluates the effects of blending of gas-field condensates with gasoline(octane) in terms of - fuel properties, engine performance and exhaust emission. Condensate samples collected from Kailashtila gas field were blended with octane, ranging from 30% to 75% by volume. However for blends with above 60% condensate, cold starting of engine became difficult. Investigation revealed that the condensate samples had significantly higher distillation temperatures compared to octane, but were not far different in terms of heating value and carbon residues. Engine tests showed Kailashtila blends performing quite similar to octane in terms of power and thermal efficiency. No noticeable knocking was observed from in-cylinder pressure traces. For all the gasoline-condensate blends the test engine ran with relatively leaner air-fuel mixture delivering slightly lower CO emissions but HC and NOx emissions were similar to octane. Road trials of a test vehicle in real traffic condition and on a standard gradient using 50%(v/v) gasoline-condensate blend were also carried out. The test vehicle did not exhibit any noticeable difference in drivability compared to octane.Keywords: condensates, engine performance, fuel adulteration, gasoline-condensate blends
Procedia PDF Downloads 2511499 A Wearable Fluorescence Imaging Device for Intraoperative Identification of Human Brain Tumors
Authors: Guoqiang Yu, Mehrana Mohtasebi, Jinghong Sun, Thomas Pittman
Abstract:
Malignant glioma (MG) is the most common type of primary malignant brain tumor. Surgical resection of MG remains the cornerstone of therapy, and the extent of resection correlates with patient survival. A limiting factor for resection, however, is the difficulty in differentiating the tumor from normal tissue during surgery. Fluorescence imaging is an emerging technique for real-time intraoperative visualization of MGs and their boundaries. However, most clinical-grade neurosurgical operative microscopes with fluorescence imaging ability are hampered by low adoption rates due to high cost, limited portability, limited operation flexibility, and lack of skilled professionals with technical knowledge. To overcome the limitations, we innovatively integrated miniaturized light sources, flippable filters, and a recording camera to the surgical eye loupes to generate a wearable fluorescence eye loupe (FLoupe) device for intraoperative imaging of fluorescent MGs. Two FLoupe prototypes were constructed for imaging of Fluorescein and 5-aminolevulinic acid (5-ALA), respectively. The wearable FLoupe devices were tested on tumor-simulating phantoms and patients with MGs. Comparable results were observed against the standard neurosurgical operative microscope (PENTERO® 900) with fluorescence kits. The affordable and wearable FLoupe devices enable visualization of both color and fluorescence images with the same quality as the large and expensive stationary operative microscopes. The wearable FLoupe device allows for a greater range of movement, less obstruction, and faster/easier operation. Thus, it reduces surgery time and is more easily adapted to the surgical environment than unwieldy neurosurgical operative microscopes.Keywords: fluorescence guided surgery, malignant glioma, neurosurgical operative microscope, wearable fluorescence imaging device
Procedia PDF Downloads 661498 Direct Electrical Communication of Redox Enzyme Based on 3-Dimensional Crosslinked Redox Enzyme/Carbon Nanotube on a Thiol-Modified Au Surface
Authors: A. K. M. Kafi, S. N. Nina, Mashitah M. Yusoff
Abstract:
In this work, we have described a new 3-dimensional (3D) network of crosslinked Horseradish Peroxidase/Carbon Nanotube (HRP/CNT) on a thiol-modified Au surface in order to build up the effective electrical wiring of the enzyme units with the electrode. This was achieved by the electropolymerization of aniline-functionalized carbon nanotubes (CNTs) and 4-aminothiophenol -modified-HRP on a 4-aminothiophenol monolayer-modified Au electrode. The synthesized 3D HRP/CNT networks were characterized with cyclic voltammetry and amperometry, resulting the establishment direct electron transfer between the redox active unit of HRP and the Au surface. Electrochemical measurements reveal that the immobilized HRP exhibits high biological activity and stability and a quasi-reversible redox peak of the redox center of HRP was observed at about −0.355 and −0.275 V vs. Ag/AgCl. The electron transfer rate constant, KS and electron transfer co-efficient were found to be 0.57 s-1 and 0.42, respectively. Based on the electrocatalytic process by direct electrochemistry of HRP, a biosensor for detecting H2O2 was developed. The developed biosensor exhibits excellent electrocatalytic activity for the reduction of H2O2. The proposed biosensor modified with HRP/CNT 3D network displays a broader linear range and a lower detection limit for H2O2 determination. The linear range is from 1.0×10−7 to 1.2×10−4M with a detection limit of 2.2.0×10−8M at 3σ. Moreover, this biosensor exhibits very high sensitivity, good reproducibility and long-time stability. In summary, ease of fabrication, a low cost, fast response and high sensitivity are the main advantages of the new biosensor proposed in this study. These obvious advantages would really help for the real analytical applicability of the proposed biosensor.Keywords: biosensor, nanomaterials, redox enzyme, thiol-modified Au surface
Procedia PDF Downloads 3291497 Stem Cell Differentiation Toward Secretory Progenitors after Intestinal Ischemia-Reperfusion in a Rat is Accompanied by Inhibited Notch Signaling Cascade
Authors: Igor Sukhotnik
Abstract:
Objectives: Notch signaling is thought to act to drive cell versification in the lining of the small intestine. When Notch signaling is blocked, proliferation ceases, and epithelial cells become secretory. The purpose of the present study was to evaluate the role of Notch signaling pathway in stem cell differentiation in a rat model of intestinal ischemia-reperfusion (IR). Methods: Male Sprague-Dawley rats were randomly divided into four experimental groups: Sham-24 and Sham-48 rats underwent laparotomy and were killed 24 or 48 h later, respectively; IR-24 and IR-48 rats underwent occlusion of SMA and portal vein for 30 min followed by 24 or 48 h of reperfusion, respectively. Notch-related gene and protein expression were determined using Real Time PCR, Western blotting and immunohistochemistry. Wax histology and immunohistochemistry was used to determine cell differentiation toward absorptive (enterocytes) or secretory progenitors (goblet cells, enteroendocrine cells or Paneth cells). Results: IR-48 rats exhibited a significant decrease in Notch-1 protein expression (Western blot) that was coincided with a significant decrease in the number of Notch-1 positive cells (immunohistochemistry) in jejunum and ileum as well as Hes-1 positive cells in jejunum and ileum compared to Sham-48 rats. A significant down-regulation of Notch signaling related genes and proteins in IR animals was accompanied by a significant increase in the number of goblet and Paneth cells and decreased number of absorptive cells compared to control rats. Conclusions: Forty-eight hours following intestinal IR in rats, inhibited Notch signaling pathway was accompanied by intestinal stem cells differentiation toward secretory progenitors.Keywords: Intestine, notch, ischemia-reperfusion, cell differentiation, secretory
Procedia PDF Downloads 581496 Analysis of the Behavior of the Structure Under Internal Anfo Explosion
Authors: Seung-Min Ko, Seung-Jai Choi, Gun Jung, Jang-Ho Jay Kim
Abstract:
Although extensive explosion-related research has been performed in the past several decades, almost no research has focused on internal blasts. However, internal blast research is needed to understand about the behavior of a containment structure or building under internal blast loading, as in the case of the Chornobyl and Fukushima nuclear accidents. Therefore, the internal blast study concentrated on RC and PSC structures is performed. The test data obtained from reinforced concrete (RC) and prestressed concrete (PSC) tubular structures applied with an internal explosion using ammonium nitrate/fuel oil (ANFO) charge are used to assess their deformation resistance and ultimate failure load based on the structural stiffness change under various charge weight. For the internal blast charge weight, ANFO explosive charge weights of 15.88, 20.41, 22.68 and 24.95 kg were selected for the RC tubular structures, and 22.68, 24.95, 27.22, 29.48, and 31.75 kg were selected for PSC tubular structures, which were detonated at the center of cross section at the mid-span with a standoff distance of 1,000mm to the inner wall surface. Then, the test data were used to predict the internal charge weight required to fail a real scale reinforced concrete containment vessels (RCCV) and prestressed concrete containment vessel (PCCV). Then, the analytical results based on the experimental data were derived using the simple assumptions of the models, and another approach using the stiffness, deformation and explosion weight relationship was used to formulate a general method for analyzing internal blasted tubular structures. A model of the internal explosion of a steel tube was used as an example for validation. The proposed method can be used generically, using factors according to the material characteristics of the target structures. The results of the study are discussed in detail in the paper.Keywords: internal blast, reinforced concrete, RCCV, PCCV, stiffness, blast safety
Procedia PDF Downloads 791495 Exploring the Use of Drones for Corn Borer Management: A Case Study in Central Italy
Authors: Luana Centorame, Alessio Ilari, Marco Giustozzi, Ester Foppa Pedretti
Abstract:
Maize is one of the most important agricultural cash crops in the world, involving three different chains: food, feed, and bioenergy production. Nowadays, the European corn borer (ECB), Ostrinia nubilalis, to the best of the author's knowledge, is the most important pest to control for maize growers. The ECB is harmful to maize; young larvae are responsible for minor damage to the leaves, while the most serious damage is tunneling by older larvae that burrow into the stock. Soon after, larvae can affect cobs, and it was found that ECB can foster mycotoxin contamination; this is why it is crucial to control it. There are multiple control methods available: agronomic, biological, and microbiological means, agrochemicals, and genetically modified plants. Meanwhile, the European Union’s policy focuses on the transition to sustainable supply chains and translates into the goal of reducing the use of agrochemicals by 50%. The current work aims to compare the agrochemical treatment of ECB and biological control through beneficial insects released by drones. The methodology used includes field trials of both chemical and biological control, considering a farm in central Italy as a case study. To assess the mechanical and technical efficacy of drones with respect to standard machinery, the available literature was consulted. The findings are positive because drones allow them to get in the field promptly, in difficult conditions and with lower costs if compared to traditional techniques. At the same time, it is important to consider the limits of drones regarding pilot certification, no-fly zones, etc. In the future, it will be necessary to deepen the topic with the real application in the field of both systems, expanding the scenarios in which drones can be used and the type of material distributed.Keywords: beneficial insects, corn borer management, drones, precision agriculture
Procedia PDF Downloads 1031494 Evaluation of Rheological Properties, Anisotropic Shrinkage, and Heterogeneous Densification of Ceramic Materials during Liquid Phase Sintering by Numerical-Experimental Procedure
Authors: Hamed Yaghoubi, Esmaeil Salahi, Fateme Taati
Abstract:
The effective shear and bulk viscosity, as well as dynamic viscosity, describe the rheological properties of the ceramic body during the liquid phase sintering process. The rheological parameters depend on the physical and thermomechanical characteristics of the material such as relative density, temperature, grain size, and diffusion coefficient and activation energy. The main goal of this research is to acquire a comprehensive understanding of the response of an incompressible viscose ceramic material during liquid phase sintering process such as stress-strain relations, sintering and hydrostatic stress, the prediction of anisotropic shrinkage and heterogeneous densification as a function of sintering time by including the simultaneous influence of gravity field, and frictional force. After raw materials analysis, the standard hard porcelain mixture as a ceramic body was designed and prepared. Three different experimental configurations were designed including midpoint deflection, sinter bending, and free sintering samples. The numerical method for the ceramic specimens during the liquid phase sintering process are implemented in the CREEP user subroutine code in ABAQUS. The numerical-experimental procedure shows the anisotropic behavior, the complete difference in spatial displacement through three directions, the incompressibility for ceramic samples during the sintering process. The anisotropic shrinkage factor has been proposed to investigate the shrinkage anisotropy. It has been shown that the shrinkage along the normal axis of casting sample is about 1.5 times larger than that of casting direction, the gravitational force in pyroplastic deformation intensifies the shrinkage anisotropy more than the free sintering sample. The lowest and greatest equivalent creep strain occurs at the intermediate zone and around the central line of the midpoint distorted sample, respectively. In the sinter bending test sample, the equivalent creep strain approaches to the maximum near the contact area with refractory support. The inhomogeneity in Von-Misses, pressure, and principal stress intensifies the relative density non-uniformity in all samples, except in free sintering one. The symmetrical distribution of stress around the center of free sintering sample, cause to hinder the pyroplastic deformations. Densification results confirmed that the effective bulk viscosity was well-defined with relative density values. The stress analysis confirmed that the sintering stress is more than the hydrostatic stress from start to end of sintering time so, from both theoretically and experimentally point of view, the sintering process occurs completely.Keywords: anisotropic shrinkage, ceramic material, liquid phase sintering process, rheological properties, numerical-experimental procedure
Procedia PDF Downloads 3411493 Synthesis of Pd@ Cu Core−Shell Nanowires by Galvanic Displacement of Cu by Pd²⁺ Ions as a Modified Glassy Carbon Electrode for the Simultaneous Determination of Dihydroxybenzene Isomers Speciation
Authors: Majid Farsadrouh Rashti, Parisa Jahani, Amir Shafiee, Mehrdad Mofidi
Abstract:
The dihydroxybenzene isomers, hydroquinone (HQ), catechol (CC) and resorcinol (RS) have been widely recognized as important environmental pollutants due to their toxicity and low degradability in the ecological environment. Speciation of HQ, CC and RS is very important for environmental analysis because they co-exist of these isomers in environmental samples and are too difficult to degrade as an environmental contaminant with high toxicity. There are many analytical methods have been reported for detecting these isomers, such as spectrophotometry, fluorescence, High-performance liquid chromatography (HPLC) and electrochemical methods. These methods have attractive advantages such as simple and fast response, low maintenance costs, wide linear analysis range, high efficiency, excellent selectivity and high sensitivity. A novel modified glassy carbon electrode (GCE) with Pd@ Cu/CNTs core−shell nanowires for the simultaneous determination of hydroquinone (HQ), catechol (CC) and resorcinol (RS) is described. A detailed investigation by field emission scanning electron microscopy and electrochemistry was performed in order to elucidate the preparation process and properties of the GCE/ Pd/CuNWs-CNTs. The electrochemical response characteristic of the modified GPE/LFOR toward HQ, CC and RS were investigated by cyclic voltammetry, differential pulse voltammetry (DPV) and Chronoamperometry. Under optimum conditions, the calibrations curves were linear up to 228 µM for each with detection limits of 0.4, 0.6 and 0.8 µM for HQ, CC and RS, respectively. The diffusion coefficient for the oxidation of HQ, CC and RS at the modified electrode was calculated as 6.5×10⁻⁵, 1.6 ×10⁻⁵ and 8.5 ×10⁻⁵ cm² s⁻¹, respectively. DPV was used for the simultaneous determination of HQ, CC and RS at the modified electrode and the relative standard deviations were 2.1%, 1.9% and 1.7% for HQ, CC and RS, respectively. Moreover, GCE/Pd/CuNWs-CNTs was successfully used for determination of HQ, CC and RS in real samples.Keywords: dihydroxybenzene isomers, galvanized copper nanowires, electrochemical sensor, Palladium, speciation
Procedia PDF Downloads 1281492 Designing an Exhaust Gas Energy Recovery Module Following Measurements Performed under Real Operating Conditions
Authors: Jerzy Merkisz, Pawel Fuc, Piotr Lijewski, Andrzej Ziolkowski, Pawel Czarkowski
Abstract:
The paper presents preliminary results of the development of an automotive exhaust gas energy recovery module. The aim of the performed analyses was to select the geometry of the heat exchanger that would ensure the highest possible transfer of heat at minimum heat flow losses. The starting point for the analyses was a straight portion of a pipe, from which the exhaust system of the tested vehicle was made. The design of the heat exchanger had a cylindrical cross-section, was 300 mm long and was fitted with a diffuser and a confusor. The model works were performed for the mentioned geometry utilizing the finite volume method based on the Ansys CFX v12.1 and v14 software. This method consisted in dividing of the system into small control volumes for which the exhaust gas velocity and pressure calculations were performed using the Navier-Stockes equations. The heat exchange in the system was modeled based on the enthalpy balance. The temperature growth resulting from the acting viscosity was not taken into account. The heat transfer on the fluid/solid boundary in the wall layer with the turbulent flow was done based on an arbitrarily adopted dimensionless temperature. The boundary conditions adopted in the analyses included the convective condition of heat transfer on the outer surface of the heat exchanger and the mass flow and temperature of the exhaust gas at the inlet. The mass flow and temperature of the exhaust gas were assumed based on the measurements performed in actual traffic using portable PEMS analyzers. The research object was a passenger vehicle fitted with a 1.9 dm3 85 kW diesel engine. The tests were performed in city traffic conditions.Keywords: waste heat recovery, heat exchanger, CFD simulation, pems
Procedia PDF Downloads 5741491 A Step Magnitude Haptic Feedback Device and Platform for Better Way to Review Kinesthetic Vibrotactile 3D Design in Professional Training
Authors: Biki Sarmah, Priyanko Raj Mudiar
Abstract:
In the modern world of remotely interactive virtual reality-based learning and teaching, including professional skill-building training and acquisition practices, as well as data acquisition and robotic systems, the revolutionary application or implementation of field-programmable neurostimulator aids and first-hand interactive sensitisation techniques into 3D holographic audio-visual platforms have been a coveted dream of many scholars, professionals, scientists, and students. Integration of 'kinaesthetic vibrotactile haptic perception' along with an actuated step magnitude contact profiloscopy in augmented reality-based learning platforms and professional training can be implemented by using an extremely calculated and well-coordinated image telemetry including remote data mining and control technique. A real-time, computer-aided (PLC-SCADA) field calibration based algorithm must be designed for the purpose. But most importantly, in order to actually realise, as well as to 'interact' with some 3D holographic models displayed over a remote screen using remote laser image telemetry and control, all spatio-physical parameters like cardinal alignment, gyroscopic compensation, as well as surface profile and thermal compositions, must be implemented using zero-order type 1 actuators (or transducers) because they provide zero hystereses, zero backlashes, low deadtime as well as providing a linear, absolutely controllable, intrinsically observable and smooth performance with the least amount of error compensation while ensuring the best ergonomic comfort ever possible for the users.Keywords: haptic feedback, kinaesthetic vibrotactile 3D design, medical simulation training, piezo diaphragm based actuator
Procedia PDF Downloads 1661490 'Performance-Based' Seismic Methodology and Its Application in Seismic Design of Reinforced Concrete Structures
Authors: Jelena R. Pejović, Nina N. Serdar
Abstract:
This paper presents an analysis of the “Performance-Based” seismic design method, in order to overcome the perceived disadvantages and limitations of the existing seismic design approach based on force, in engineering practice. Bearing in mind, the specificity of the earthquake as a load and the fact that the seismic resistance of the structures solely depends on its behaviour in the nonlinear field, traditional seismic design approach based on force and linear analysis is not adequate. “Performance-Based” seismic design method is based on nonlinear analysis and can be used in everyday engineering practice. This paper presents the application of this method to eight-story high reinforced concrete building with combined structural system (reinforced concrete frame structural system in one direction and reinforced concrete ductile wall system in other direction). The nonlinear time-history analysis is performed on the spatial model of the structure using program Perform 3D, where the structure is exposed to forty real earthquake records. For considered building, large number of results were obtained. It was concluded that using this method we could, with a high degree of reliability, evaluate structural behavior under earthquake. It is obtained significant differences in the response of structures to various earthquake records. Also analysis showed that frame structural system had not performed well at the effect of earthquake records on soil like sand and gravel, while a ductile wall system had a satisfactory behavior on different types of soils.Keywords: ductile wall, frame system, nonlinear time-history analysis, performance-based methodology, RC building
Procedia PDF Downloads 3661489 Enhancing Power System Resilience: An Adaptive Under-Frequency Load Shedding Scheme Incorporating PV Generation and Fast Charging Stations
Authors: Sami M. Alshareef
Abstract:
In the rapidly evolving energy landscape, the integration of renewable energy sources and the electrification of transportation are essential steps toward achieving sustainability goals. However, these advancements introduce new challenges, particularly in maintaining frequency stability due to variable photovoltaic (PV) generation and the growing demand for fast charging stations. The variability of photovoltaic (PV) generation due to weather conditions can disrupt the balance between generation and load, resulting in frequency deviations. To ensure the stability of power systems, it is imperative to develop effective under frequency load-shedding schemes. This research proposal presents an adaptive under-frequency load shedding scheme based on the power swing equation, designed explicitly for the IEEE-9 Bus Test System, that includes PV generation and fast charging stations. This research aims to address these challenges by developing an advanced scheme that dynamically disconnects fast charging stations based on power imbalances. The scheme prioritizes the disconnection of stations near affected areas to expedite system frequency stabilization. To achieve these goals, the research project will leverage the power swing equation, a widely recognized model for analyzing system dynamics during under-frequency events. By utilizing this equation, the proposed scheme will adaptively adjust the load-shedding process in real-time to maintain frequency stability and prevent power blackouts. The research findings will support the transition towards sustainable energy systems by ensuring a reliable and uninterrupted electricity supply while enhancing the resilience and stability of power systems during under-frequency events.Keywords: load shedding, fast charging stations, pv generation, power system resilience
Procedia PDF Downloads 811488 The Development, Validation, and Evaluation of the Code Blue Simulation Module in Improving the Code Blue Response Time among Nurses
Authors: Siti Rajaah Binti Sayed Sultan
Abstract:
Managing the code blue event is stressful for nurses, the patient, and the patient's families. The rapid response from the first and second responders in the code blue event will improve patient outcomes and prevent tissue hypoxia that leads to brain injury and other organ failures. Providing 1 minute for the cardiac massage and 2 minutes for defibrillation will significantly improve patient outcomes. As we know, the American Heart Association came out with guidelines for managing cardiac arrest patients. The hospital must provide competent staff to manage this situation. It can be achieved when the staff is well equipped with the skill, attitude, and knowledge to manage this situation with well-planned strategies, i.e., clear guidelines for managing the code blue event, competent staff, and functional equipment. The code blue simulation (CBS) was chosen in the training program for code blue management because it can mimic real scenarios. Having the code blue simulation module will allow the staff to appreciate what they will face during the code blue event, especially since it rarely happens in that area. This CBS module training will help the staff familiarize themselves with the activities that happened during actual events and be able to operate the equipment accordingly. Being challenged and independent in managing the code blue in the early phase gives the patient a better outcome. The CBS module will help the assessor and the hospital management team with the proper tools and guidelines for managing the code blue drill accordingly. As we know, prompt action will benefit the patient and their family. It also indirectly increases the confidence and job satisfaction among the nurses, increasing the standard of care, reducing the complication and hospital burden, and enhancing cost-effective care.Keywords: code blue simulation module, development of code blue simulation module, code blue response time, code blue drill, cardiorespiratory arrest, managing code blue
Procedia PDF Downloads 671487 A Finite Element Analysis of Hexagonal Double-Arrowhead Auxetic Structure with Enhanced Energy Absorption Characteristics and Stiffness
Abstract:
Auxetic materials, as an emerging artificial designed metamaterial has attracted growing attention due to their promising negative Poisson’s ratio behaviors and tunable properties. The conventional auxetic lattice structures for which the deformation process is governed by a bending-dominated mechanism have faced the limitation of poor mechanical performance for many potential engineering applications. Recently, both load-bearing and energy absorption capabilities have become a crucial consideration in auxetic structure design. This study reports the finite element analysis of a class of hexagonal double-arrowhead auxetic structures with enhanced stiffness and energy absorption performance. The structure design was developed by extending the traditional double-arrowhead honeycomb to a hexagon frame, the stretching-dominated deformation mechanism was determined according to Maxwell’s stability criterion. The finite element (FE) models of 2D lattice structures established with stainless steel material were analyzed in ABAQUS/Standard for predicting in-plane structural deformation mechanism, failure process, and compressive elastic properties. Based on the computational simulation, the parametric analysis was studied to investigate the effect of the structural parameters on Poisson’s ratio and mechanical properties. The geometrical optimization was then implemented to achieve the optimal Poisson’s ratio for the maximum specific energy absorption. In addition, the optimized 2D lattice structure was correspondingly converted into a 3D geometry configuration by using the orthogonally splicing method. The numerical results of 2D and 3D structures under compressive quasi-static loading conditions were compared separately with the traditional double-arrowhead re-entrant honeycomb in terms of specific Young's moduli, Poisson's ratios, and specified energy absorption. As a result, the energy absorption capability and stiffness are significantly reinforced with a wide range of Poisson’s ratio compared to traditional double-arrowhead re-entrant honeycomb. The auxetic behaviors, energy absorption capability, and yield strength of the proposed structure are adjustable with different combinations of joint angle, struts thickness, and the length-width ratio of the representative unit cell. The numerical prediction in this study suggests the proposed concept of hexagonal double-arrowhead structure could be a suitable candidate for the energy absorption applications with a constant request of load-bearing capacity. For future research, experimental analysis is required for the validation of the numerical simulation.Keywords: auxetic, energy absorption capacity, finite element analysis, negative Poisson's ratio, re-entrant hexagonal honeycomb
Procedia PDF Downloads 871486 Numerical Studies on Thrust Vectoring Using Shock-Induced Self Impinging Secondary Jets
Authors: S. Vignesh, N. Vishnu, S. Vigneshwaran, M. Vishnu Anand, Dinesh Kumar Babu, V. R. Sanal Kumar
Abstract:
The study of the primary flow velocity and the self impinging secondary jet flow mixing is important from both the fundamental research and the application point of view. Real industrial configurations are more complex than simple shear layers present in idealized numerical thrust-vectoring models due to the presence of combustion, swirl and confinement. Predicting the flow features of self impinging secondary jets in a supersonic primary flow is complex owing to the fact that there are a large number of parameters involved. Earlier studies have been highlighted several key features of self impinging jets, but an extensive characterization in terms of jet interaction between supersonic flow and self impinging secondary sonic jets is still an active research topic. In this paper numerical studies have been carried out using a validated two-dimensional k-omega standard turbulence model for the design optimization of a thrust vector control system using shock induced self impinging secondary flow sonic jets using non-reacting flows. Efforts have been taken for examining the flow features of TVC system with various secondary jets at different divergent locations and jet impinging angles with the same inlet jet pressure and mass flow ratio. The results from the parametric studies reveal that in addition to the primary to the secondary mass flow ratio the characteristics of the self impinging secondary jets having bearing on an efficient thrust vectoring. We concluded that the self impinging secondary jet nozzles are better than single jet nozzle with the same secondary mass flow rate owing to the fact fixing of the self impinging secondary jet nozzles with proper jet angle could facilitate better thrust vectoring for any supersonic aerospace vehicle.Keywords: fluidic thrust vectoring, rocket steering, supersonic to sonic jet interaction, TVC in aerospace vehicles
Procedia PDF Downloads 5891485 Maximum Likelihood Estimation Methods on a Two-Parameter Rayleigh Distribution under Progressive Type-Ii Censoring
Authors: Daniel Fundi Murithi
Abstract:
Data from economic, social, clinical, and industrial studies are in some way incomplete or incorrect due to censoring. Such data may have adverse effects if used in the estimation problem. We propose the use of Maximum Likelihood Estimation (MLE) under a progressive type-II censoring scheme to remedy this problem. In particular, maximum likelihood estimates (MLEs) for the location (µ) and scale (λ) parameters of two Parameter Rayleigh distribution are realized under a progressive type-II censoring scheme using the Expectation-Maximization (EM) and the Newton-Raphson (NR) algorithms. These algorithms are used comparatively because they iteratively produce satisfactory results in the estimation problem. The progressively type-II censoring scheme is used because it allows the removal of test units before the termination of the experiment. Approximate asymptotic variances and confidence intervals for the location and scale parameters are derived/constructed. The efficiency of EM and the NR algorithms is compared given root mean squared error (RMSE), bias, and the coverage rate. The simulation study showed that in most sets of simulation cases, the estimates obtained using the Expectation-maximization algorithm had small biases, small variances, narrower/small confidence intervals width, and small root of mean squared error compared to those generated via the Newton-Raphson (NR) algorithm. Further, the analysis of a real-life data set (data from simple experimental trials) showed that the Expectation-Maximization (EM) algorithm performs better compared to Newton-Raphson (NR) algorithm in all simulation cases under the progressive type-II censoring scheme.Keywords: expectation-maximization algorithm, maximum likelihood estimation, Newton-Raphson method, two-parameter Rayleigh distribution, progressive type-II censoring
Procedia PDF Downloads 163