Search results for: multi-agent double deep Q-network (MADDQN)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3244

Search results for: multi-agent double deep Q-network (MADDQN)

2314 Real-Time Pedestrian Detection Method Based on Improved YOLOv3

Authors: Jingting Luo, Yong Wang, Ying Wang

Abstract:

Pedestrian detection in image or video data is a very important and challenging task in security surveillance. The difficulty of this task is to locate and detect pedestrians of different scales in complex scenes accurately. To solve these problems, a deep neural network (RT-YOLOv3) is proposed to realize real-time pedestrian detection at different scales in security monitoring. RT-YOLOv3 improves the traditional YOLOv3 algorithm. Firstly, the deep residual network is added to extract vehicle features. Then six convolutional neural networks with different scales are designed and fused with the corresponding scale feature maps in the residual network to form the final feature pyramid to perform pedestrian detection tasks. This method can better characterize pedestrians. In order to further improve the accuracy and generalization ability of the model, a hybrid pedestrian data set training method is used to extract pedestrian data from the VOC data set and train with the INRIA pedestrian data set. Experiments show that the proposed RT-YOLOv3 method achieves 93.57% accuracy of mAP (mean average precision) and 46.52f/s (number of frames per second). In terms of accuracy, RT-YOLOv3 performs better than Fast R-CNN, Faster R-CNN, YOLO, SSD, YOLOv2, and YOLOv3. This method reduces the missed detection rate and false detection rate, improves the positioning accuracy, and meets the requirements of real-time detection of pedestrian objects.

Keywords: pedestrian detection, feature detection, convolutional neural network, real-time detection, YOLOv3

Procedia PDF Downloads 130
2313 Deep Learning-Based Object Detection on Low Quality Images: A Case Study of Real-Time Traffic Monitoring

Authors: Jean-Francois Rajotte, Martin Sotir, Frank Gouineau

Abstract:

The installation and management of traffic monitoring devices can be costly from both a financial and resource point of view. It is therefore important to take advantage of in-place infrastructures to extract the most information. Here we show how low-quality urban road traffic images from cameras already available in many cities (such as Montreal, Vancouver, and Toronto) can be used to estimate traffic flow. To this end, we use a pre-trained neural network, developed for object detection, to count vehicles within images. We then compare the results with human annotations gathered through crowdsourcing campaigns. We use this comparison to assess performance and calibrate the neural network annotations. As a use case, we consider six months of continuous monitoring over hundreds of cameras installed in the city of Montreal. We compare the results with city-provided manual traffic counting performed in similar conditions at the same location. The good performance of our system allows us to consider applications which can monitor the traffic conditions in near real-time, making the counting usable for traffic-related services. Furthermore, the resulting annotations pave the way for building a historical vehicle counting dataset to be used for analysing the impact of road traffic on many city-related issues, such as urban planning, security, and pollution.

Keywords: traffic monitoring, deep learning, image annotation, vehicles, roads, artificial intelligence, real-time systems

Procedia PDF Downloads 184
2312 Improving Security by Using Secure Servers Communicating via Internet with Standalone Secure Software

Authors: Carlos Gonzalez

Abstract:

This paper describes the use of the Internet as a feature to enhance the security of our software that is going to be distributed/sold to users potentially all over the world. By placing in a secure server some of the features of the secure software, we increase the security of such software. The communication between the protected software and the secure server is done by a double lock algorithm. This paper also includes an analysis of intruders and describes possible responses to detect threats.

Keywords: internet, secure software, threats, cryptography process

Procedia PDF Downloads 314
2311 Performance of Rapid Impact Compaction as a Middle-Deep Ground Improvement Technique

Authors: Bashar Tarawneh, Yasser Hakam

Abstract:

Rapid Impact Compaction (RIC) is a modern dynamic compaction device mainly used to compact sandy soils, where silt and clay contents are low. The device uses the piling hammer technology to increase the bearing capacity of soils through controlled impacts. The RIC device uses "controlled impact compaction" of the ground using a 9-ton hammer dropped from the height between 0.3 m to 1.2 m onto a 1.5 m diameter steel patent foot. The delivered energy is about 26,487 to 105,948 Joules per drop. To evaluate the performance of this technique, three project sites in the United Arab Emirates were improved using RIC. In those sites, a loose to very loose fine to medium sand was encountered at a depth ranging from 1.0m to 4.0m below the ground level. To evaluate the performance of the RIC, Cone Penetration Tests (CPT) were carried out before and after improvement. Also, load tests were carried out post-RIC work to assess the settlements and bearing capacity. The soil was improved to a depth of about 5.0m below the ground level depending on the CPT friction ratio (the ratio between sleeve friction and tip resistance). CPT tip resistance was significantly increased post ground improvement work. Load tests showed enhancement in the soil bearing capacity and reduction in the potential settlements. This study demonstrates the successful application of the RIC for middle-deep improvement and compaction of the ground. Foundation design criteria were achieved in all site post-RIC work.

Keywords: compaction, RIC, ground improvement, CPT

Procedia PDF Downloads 357
2310 Concept of the Active Flipped Learning in Engineering Mechanics

Authors: Lin Li, Farshad Amini

Abstract:

The flipped classroom has been introduced to promote collaborative learning and higher-order learning objectives. In contrast to the traditional classroom, the flipped classroom has students watch prerecorded lecture videos before coming to class and then “class becomes the place to work through problems, advance concepts, and engage in collaborative learning”. In this paper, the active flipped learning combines flipped classroom with active learning that is to establish an active flipped learning (AFL) model, aiming to promote active learning, stress deep learning, encourage student engagement and highlight data-driven personalized learning. Because students have watched the lecture prior to class, contact hours can be devoted to problem-solving and gain a deeper understanding of the subject matter. The instructor is able to provide students with a wide range of learner-centered opportunities in class for greater mentoring and collaboration, increasing the possibility to engage students. Currently, little is known about the extent to which AFL improves engineering students’ performance. This paper presents the preliminary study on the core course of sophomore students in Engineering Mechanics. A series of survey and interviews have been conducted to compare students’ learning engagement, empowerment, self-efficacy, and satisfaction with the AFL. It was found that the AFL model taking advantage of advanced technology is a convenient and professional avenue for engineering students to strengthen their academic confidence and self-efficacy in the Engineering Mechanics by actively participating in learning and fostering their deep understanding of engineering statics and dynamics

Keywords: active learning, engineering mechanics, flipped classroom, performance

Procedia PDF Downloads 280
2309 The Use of Bleomycin and Analogues to Probe the Chromatin Structure of Human Genes

Authors: Vincent Murray

Abstract:

The chromatin structure at the transcription start sites (TSSs) of genes is very important in the control of gene expression. In order for gene expression to occur, the chromatin structure at the TSS has to be altered so that the transcriptional machinery can be assembled and RNA transcripts can be produced. In particular, the nucleosome structure and positioning around the TSS has to be changed. Bleomycin is utilized as an anti-tumor agent to treat Hodgkin's lymphoma, squamous cell carcinoma, and testicular cancer. Bleomycin produces DNA damage in human cells and DNA strand breaks, especially double-strand breaks, are thought to be responsible for the cancer chemotherapeutic activity of bleomycin. Bleomycin is a large glycopeptide with molecular weight of approximately 1500 Daltons and hence its DNA strand cleavage activity can be utilized as a probe of chromatin structure. In this project, Illumina next-generation DNA sequencing technology was used to determine the position of DNA double-strand breaks at the TSSs of genes in intact cells. In this genome-wide study, it was found that bleomycin cleavage preferentially occurred at the TSSs of actively transcribed human genes in comparison with non-transcribed genes. There was a correlation between the level of enhanced bleomycin cleavage at TSSs and the degree of transcriptional activity. In addition, bleomycin was able to determine the position of nucleosomes at the TSSs of human genes. Bleomycin analogues were also utilized as probes of chromatin structure at the TSSs of human genes. In a similar manner to bleomycin, the bleomycin analogues 6′-deoxy-BLM Z and zorbamycin preferentially cleaved at the TSSs of human genes. Interestingly this degree of enhanced TSS cleavage inversely correlated with the cytotoxicity (IC50 values) of BLM analogues. This indicated that the degree of cleavage by bleomycin analogues at the TSSs of human genes was very important in the cytotoxicity of bleomycin and analogues. It also provided a deeper insight into the mechanism of action of this cancer chemotherapeutic agent since actively transcribed genes were preferentially targeted.

Keywords: anti-cancer activity, chromatin structure, cytotoxicity, gene expression, next-generation DNA sequencing

Procedia PDF Downloads 110
2308 Deep Vision: A Robust Dominant Colour Extraction Framework for T-Shirts Based on Semantic Segmentation

Authors: Kishore Kumar R., Kaustav Sengupta, Shalini Sood Sehgal, Poornima Santhanam

Abstract:

Fashion is a human expression that is constantly changing. One of the prime factors that consistently influences fashion is the change in colour preferences. The role of colour in our everyday lives is very significant. It subconsciously explains a lot about one’s mindset and mood. Analyzing the colours by extracting them from the outfit images is a critical study to examine the individual’s/consumer behaviour. Several research works have been carried out on extracting colours from images, but to the best of our knowledge, there were no studies that extract colours to specific apparel and identify colour patterns geographically. This paper proposes a framework for accurately extracting colours from T-shirt images and predicting dominant colours geographically. The proposed method consists of two stages: first, a U-Net deep learning model is adopted to segment the T-shirts from the images. Second, the colours are extracted only from the T-shirt segments. The proposed method employs the iMaterialist (Fashion) 2019 dataset for the semantic segmentation task. The proposed framework also includes a mechanism for gathering data and analyzing India’s general colour preferences. From this research, it was observed that black and grey are the dominant colour in different regions of India. The proposed method can be adapted to study fashion’s evolving colour preferences.

Keywords: colour analysis in t-shirts, convolutional neural network, encoder-decoder, k-means clustering, semantic segmentation, U-Net model

Procedia PDF Downloads 97
2307 Anatomical Investigation of Superficial Fascia Relationships with the Skin and Underlying Tissue in the Greyhound Rump, Thigh, and Crus

Authors: Oday A. Al-Juhaishi, Sa`ad M. Ismail, Hung-Hsun Yen, Christina M. Murray, Helen M. S. Davies

Abstract:

The functional anatomy of the fascia in the greyhound is still poorly understood, and incompletely described. The basic knowledge of fascia stems mainly from anatomical, histological and ultrastructural analyses. In this study, twelve specimens of hindlimbs from six fresh greyhound cadavers (3 male, 3 female) were used to examine the topographical relationships of the superficial fascia with the skin and underlying tissue. The first incision was made along the dorsal midline from the level of the thoracolumbar junction caudally to the level of the mid sacrum. The second incision was begun at the level of the first incision and extended along the midline of the lateral aspect of the hindlimb distally, to just proximal to the tarsus, and, the skin margins carefully separated to observe connective tissue links between the skin and superficial fascia, attachment points of the fascia and the relationships of the fascia with blood vessels that supply the skin. A digital camera was used to record the anatomical features as they were revealed. The dissections identified fibrous septa connecting the skin with the superficial fascia and deep fascia in specific areas. The presence of the adipose tissue was found to be very rare within the superficial fascia in these specimens. On the extensor aspects of some joints, a fusion between the superficial fascia and deep fascia was observed. This fusion created a subcutaneous bursa in the following areas: a prepatellar bursa of the stifle, a tarsal bursa caudal to the calcaneus bone, and an ischiatic bursa caudal to the ischiatic tuberosity. The evaluation of blood vessels showed that the perforating vessels passed through fibrous septa in a perpendicular direction to supply the skin, with the largest branch noted in the gluteal area. The attachment points between the superficial fascia and skin were mainly found in the region of the flexor aspect of the joints, such as caudal to the stifle joint. The numerous fibrous septa between the superficial fascia and skin that have been identified in some areas, may create support for the blood vessels that penetrate fascia and into the skin, while allowing for movement between the tissue planes. The subcutaneous bursae between the skin and the superficial fascia where it is fused with the deep fascia may be useful to decrease friction between moving areas. The adhesion points may be related to the integrity and loading of the skin. The attachment points fix the skin and appear to divide the hindlimb into anatomical compartments.

Keywords: attachment points, fibrous septa, greyhound, subcutaneous bursa, superficial fascia

Procedia PDF Downloads 352
2306 Use of a Novel Intermittent Compression Shoe in Reducing Lower Limb Venous Stasis

Authors: Hansraj Riteesh Bookun, Cassandra Monique Hidajat

Abstract:

This pilot study investigated the efficacy of a newly designed shoe which will act as an intermittent pneumatic compression device to augment venous flow in the lower limb. The aim was to assess the degree with which a wearable intermittent compression device can increase the venous flow in the popliteal vein. Background: Deep venous thrombosis and chronic venous insufficiency are relatively common problems with significant morbidity and mortality. While mechanical and chemical thromboprophylaxis measures are in place in hospital environments (in the form of TED stockings, intermittent pneumatic compression devices, analgesia, antiplatelet and anticoagulant agents), there are limited options in a community setting. Additionally, many individuals are poorly tolerant of graduated compression stockings due to the difficulty in putting them on, their constant tightness and increased associated discomfort in warm weather. These factors may hinder the management of their chronic venous insufficiency. Method: The device is lightweight, easy to wear and comfortable, with a self-contained power source. It features a Bluetooth transmitter and can be controlled with a smartphone. It is externally almost indistinguishable from a normal shoe. During activation, two bladders are inflated -one overlying the metatarsal heads and the second at the pedal arch. The resulting cyclical increase in pressure squeezes blood into the deep venous system. This will decrease periods of stasis and potentially reduce the risk of deep venous thrombosis. The shoe was fitted to 2 healthy participants and the peak systolic velocity of flow in the popliteal vein was measured during and prior to intermittent compression phases. Assessments of total flow volume were also performed. All haemodynamic assessments were performed with ultrasound by a licensed sonographer. Results: Mean peak systolic velocity of 3.5 cm/s with standard deviation of 1.3 cm/s were obtained. There was a three fold increase in mean peak systolic velocity and five fold increase in total flow volume. Conclusion: The device augments venous flow in the leg significantly. This may contribute to lowered thromboembolic risk during periods of prolonged travel or immobility. This device may also serve as an adjunct in the treatment of chronic venous insufficiency. The study will be replicated on a larger scale in a multi—centre trial.

Keywords: venous, intermittent compression, shoe, wearable device

Procedia PDF Downloads 180
2305 Multimodal Biometric Cryptography Based Authentication in Cloud Environment to Enhance Information Security

Authors: D. Pugazhenthi, B. Sree Vidya

Abstract:

Cloud computing is one of the emerging technologies that enables end users to use the services of cloud on ‘pay per usage’ strategy. This technology grows in a fast pace and so is its security threat. One among the various services provided by cloud is storage. In this service, security plays a vital factor for both authenticating legitimate users and protection of information. This paper brings in efficient ways of authenticating users as well as securing information on the cloud. Initial phase proposed in this paper deals with an authentication technique using multi-factor and multi-dimensional authentication system with multi-level security. Unique identification and slow intrusive formulates an advanced reliability on user-behaviour based biometrics than conventional means of password authentication. By biometric systems, the accounts are accessed only by a legitimate user and not by a nonentity. The biometric templates employed here do not include single trait but multiple, viz., iris and finger prints. The coordinating stage of the authentication system functions on Ensemble Support Vector Machine (SVM) and optimization by assembling weights of base SVMs for SVM ensemble after individual SVM of ensemble is trained by the Artificial Fish Swarm Algorithm (AFSA). Thus it helps in generating a user-specific secure cryptographic key of the multimodal biometric template by fusion process. Data security problem is averted and enhanced security architecture is proposed using encryption and decryption system with double key cryptography based on Fuzzy Neural Network (FNN) for data storing and retrieval in cloud computing . The proposing scheme aims to protect the records from hackers by arresting the breaking of cipher text to original text. This improves the authentication performance that the proposed double cryptographic key scheme is capable of providing better user authentication and better security which distinguish between the genuine and fake users. Thus, there are three important modules in this proposed work such as 1) Feature extraction, 2) Multimodal biometric template generation and 3) Cryptographic key generation. The extraction of the feature and texture properties from the respective fingerprint and iris images has been done initially. Finally, with the help of fuzzy neural network and symmetric cryptography algorithm, the technique of double key encryption technique has been developed. As the proposed approach is based on neural networks, it has the advantage of not being decrypted by the hacker even though the data were hacked already. The results prove that authentication process is optimal and stored information is secured.

Keywords: artificial fish swarm algorithm (AFSA), biometric authentication, decryption, encryption, fingerprint, fusion, fuzzy neural network (FNN), iris, multi-modal, support vector machine classification

Procedia PDF Downloads 246
2304 A Student Centered Learning Environment in Engineering Education: Design and a Longitudinal Study of Impact

Authors: Tom O'Mahony

Abstract:

This article considers the design of a student-centered learning environment in engineering education. The learning environment integrates a number of components, including project-based learning, collaborative learning, two-stage assignments, active learning lectures, and a flipped-classroom. Together these elements place the individual learner and their learning at the center of the environment by focusing on understanding, enhancing relevance, applying learning, obtaining rich feedback, making choices, and taking responsibility. The evolution of this environment from 2014 to the present day is outlined. The impact of this environment on learners and their learning is evaluated via student questionnaires that consist of both open and closed-ended questions. The closed questions indicate that students found the learning environment to be really interesting and enjoyable (rated as 4.7 on a 5 point scale) and encouraged students to adopt a deep approach towards studying the course materials (rated as 4.0 on a 5 point scale). A content analysis of the open-ended questions provides evidence that the project, active learning lectures, and flipped classroom all contribute to the success of this environment. Furthermore, this analysis indicates that the two-stage assessment process, in which feedback is provided between a draft and final assignment, is the key component and the dominant theme. A limitation of the study is the small class size (less than 20 learners per year), but, to some degree, this is compensated for by the longitudinal nature of the study.

Keywords: deep approaches, formative assessment, project-based learning, student-centered learning

Procedia PDF Downloads 104
2303 Refined Edge Detection Network

Authors: Omar Elharrouss, Youssef Hmamouche, Assia Kamal Idrissi, Btissam El Khamlichi, Amal El Fallah-Seghrouchni

Abstract:

Edge detection is represented as one of the most challenging tasks in computer vision, due to the complexity of detecting the edges or boundaries in real-world images that contains objects of different types and scales like trees, building as well as various backgrounds. Edge detection is represented also as a key task for many computer vision applications. Using a set of backbones as well as attention modules, deep-learning-based methods improved the detection of edges compared with the traditional methods like Sobel and Canny. However, images of complex scenes still represent a challenge for these methods. Also, the detected edges using the existing approaches suffer from non-refined results while the image output contains many erroneous edges. To overcome this, n this paper, by using the mechanism of residual learning, a refined edge detection network is proposed (RED-Net). By maintaining the high resolution of edges during the training process, and conserving the resolution of the edge image during the network stage, we make the pooling outputs at each stage connected with the output of the previous layer. Also, after each layer, we use an affined batch normalization layer as an erosion operation for the homogeneous region in the image. The proposed methods are evaluated using the most challenging datasets including BSDS500, NYUD, and Multicue. The obtained results outperform the designed edge detection networks in terms of performance metrics and quality of output images.

Keywords: edge detection, convolutional neural networks, deep learning, scale-representation, backbone

Procedia PDF Downloads 90
2302 Axial Load Capacity of Drilled Shafts from In-Situ Test Data at Semani Site, in Albania

Authors: Neritan Shkodrani, Klearta Rrushi, Anxhela Shaha

Abstract:

Generally, the design of axial load capacity of deep foundations is based on the data provided from field tests, such as SPT (Standard Penetration Test) and CPT (Cone Penetration Test) tests. This paper reports the results of axial load capacity analysis of drilled shafts at a construction site at Semani, in Fier county, Fier prefecture in Albania. In this case, the axial load capacity analyses are based on the data of 416 SPT tests and 12 CPTU tests, which are carried out in this site construction using 12 boreholes (10 borings of a depth 30.0 m and 2 borings of a depth of 80.0m). The considered foundation widths range from 0.5m to 2.5 m and foundation embedment lengths is fixed at a value of 25m. SPT – based analytical methods from the Japanese practice of design (Building Standard Law of Japan) and CPT – based analytical Eslami and Fellenius methods are used for obtaining axial ultimate load capacity of drilled shafts. The considered drilled shaft (25m long and 0.5m - 2.5m in diameter) is analyzed for the soil conditions of each borehole. The values obtained from sets of calculations are shown in different charts. Then the reported axial load capacity values acquired from SPT and CPTU data are compared and some conclusions are found related to the mentioned methods of calculations.

Keywords: deep foundations, drilled shafts, axial load capacity, ultimate load capacity, allowable load capacity, SPT test, CPTU test

Procedia PDF Downloads 96
2301 Pulmonary Disease Identification Using Machine Learning and Deep Learning Techniques

Authors: Chandu Rathnayake, Isuri Anuradha

Abstract:

Early detection and accurate diagnosis of lung diseases play a crucial role in improving patient prognosis. However, conventional diagnostic methods heavily rely on subjective symptom assessments and medical imaging, often causing delays in diagnosis and treatment. To overcome this challenge, we propose a novel lung disease prediction system that integrates patient symptoms and X-ray images to provide a comprehensive and reliable diagnosis.In this project, develop a mobile application specifically designed for detecting lung diseases. Our application leverages both patient symptoms and X-ray images to facilitate diagnosis. By combining these two sources of information, our application delivers a more accurate and comprehensive assessment of the patient's condition, minimizing the risk of misdiagnosis. Our primary aim is to create a user-friendly and accessible tool, particularly important given the current circumstances where many patients face limitations in visiting healthcare facilities. To achieve this, we employ several state-of-the-art algorithms. Firstly, the Decision Tree algorithm is utilized for efficient symptom-based classification. It analyzes patient symptoms and creates a tree-like model to predict the presence of specific lung diseases. Secondly, we employ the Random Forest algorithm, which enhances predictive power by aggregating multiple decision trees. This ensemble technique improves the accuracy and robustness of the diagnosis. Furthermore, we incorporate a deep learning model using Convolutional Neural Network (CNN) with the RestNet50 pre-trained model. CNNs are well-suited for image analysis and feature extraction. By training CNN on a large dataset of X-ray images, it learns to identify patterns and features indicative of lung diseases. The RestNet50 architecture, known for its excellent performance in image recognition tasks, enhances the efficiency and accuracy of our deep learning model. By combining the outputs of the decision tree-based algorithms and the deep learning model, our mobile application generates a comprehensive lung disease prediction. The application provides users with an intuitive interface to input their symptoms and upload X-ray images for analysis. The prediction generated by the system offers valuable insights into the likelihood of various lung diseases, enabling individuals to take appropriate actions and seek timely medical attention. Our proposed mobile application has significant potential to address the rising prevalence of lung diseases, particularly among young individuals with smoking addictions. By providing a quick and user-friendly approach to assessing lung health, our application empowers individuals to monitor their well-being conveniently. This solution also offers immense value in the context of limited access to healthcare facilities, enabling timely detection and intervention. In conclusion, our research presents a comprehensive lung disease prediction system that combines patient symptoms and X-ray images using advanced algorithms. By developing a mobile application, we provide an accessible tool for individuals to assess their lung health conveniently. This solution has the potential to make a significant impact on the early detection and management of lung diseases, benefiting both patients and healthcare providers.

Keywords: CNN, random forest, decision tree, machine learning, deep learning

Procedia PDF Downloads 66
2300 Pavement Management for a Metropolitan Area: A Case Study of Montreal

Authors: Luis Amador Jimenez, Md. Shohel Amin

Abstract:

Pavement performance models are based on projections of observed traffic loads, which makes uncertain to study funding strategies in the long run if history does not repeat. Neural networks can be used to estimate deterioration rates but the learning rate and momentum have not been properly investigated, in addition, economic evolvement could change traffic flows. This study addresses both issues through a case study for roads of Montreal that simulates traffic for a period of 50 years and deals with the measurement error of the pavement deterioration model. Travel demand models are applied to simulate annual average daily traffic (AADT) every 5 years. Accumulated equivalent single axle loads (ESALs) are calculated from the predicted AADT and locally observed truck distributions combined with truck factors. A back propagation Neural Network (BPN) method with a Generalized Delta Rule (GDR) learning algorithm is applied to estimate pavement deterioration models capable of overcoming measurement errors. Linear programming of lifecycle optimization is applied to identify M&R strategies that ensure good pavement condition while minimizing the budget. It was found that CAD 150 million is the minimum annual budget to good condition for arterial and local roads in Montreal. Montreal drivers prefer the use of public transportation for work and education purposes. Vehicle traffic is expected to double within 50 years, ESALS are expected to double the number of ESALs every 15 years. Roads in the island of Montreal need to undergo a stabilization period for about 25 years, a steady state seems to be reached after.

Keywords: pavement management system, traffic simulation, backpropagation neural network, performance modeling, measurement errors, linear programming, lifecycle optimization

Procedia PDF Downloads 450
2299 Double Liposomes Based Dual Drug Delivery System for Effective Eradication of Helicobacter pylori

Authors: Yuvraj Singh Dangi, Brajesh Kumar Tiwari, Ashok Kumar Jain, Kamta Prasad Namdeo

Abstract:

The potential use of liposomes as drug carriers by i.v. injection is limited by their low stability in blood stream. Firstly, phospholipid exchange and transfer to lipoproteins, mainly HDL destabilizes and disintegrates liposomes with subsequent loss of content. To avoid the pain associated with injection and to obtain better patient compliance studies concerning various dosage forms, have been developed. Conventional liposomes (unilamellar and multilamellar) have certain drawbacks like low entrapment efficiency, stability and release of drug after single breach in external membrane, have led to the new type of liposomal systems. The challenge has been successfully met in the form of Double Liposomes (DL). DL is a recently developed type of liposome, consisting of smaller liposomes enveloped in lipid bilayers. The outer lipid layer of DL can protect inner liposomes against various enzymes, therefore DL was thought to be more effective than ordinary liposomes. This concept was also supported by in vitro release characteristics i.e. DL formation inhibited the release of drugs encapsulated in inner liposomes. DL consists of several small liposomes encapsulated in large liposomes, i.e., multivesicular vesicles (MVV), therefore, DL should be discriminated from ordinary classification of multilamellar vesicles (MLV), large unilamellar vesicles (LUV), small unilamellar vesicles (SUV). However, for these liposomes, the volume of inner phase is small and loading volume of water-soluble drugs is low. In the present study, the potential of phosphatidylethanolamine (PE) lipid anchored double liposomes (DL) to incorporate two drugs in a single system is exploited as a tool to augment the H. pylori eradication rate. Preparation of DL involves two steps, first formation of primary (inner) liposomes by thin film hydration method containing one drug, then addition of suspension of inner liposomes on thin film of lipid containing the other drug. The success of formation of DL was characterized by optical and transmission electron microscopy. Quantitation of DL-bacterial interaction was evaluated in terms of percent growth inhibition (%GI) on reference strain of H. pylori ATCC 26695. To confirm specific binding efficacy of DL to H. pylori PE surface receptor we performed an agglutination assay. Agglutination in DL treated H. pylori suspension suggested selectivity of DL towards the PE surface receptor of H. pylori. Monotherapy is generally not recommended for treatment of a H. pylori infection due to the danger of development of resistance and unacceptably low eradication rates. Therefore, combination therapy with amoxicillin trihydrate (AMOX) as anti-H. pylori agent and ranitidine bismuth citrate (RBC) as antisecretory agent were selected for the study with an expectation that this dual-drug delivery approach will exert acceptable anti-H. pylori activity.

Keywords: Helicobacter pylorI, amoxicillin trihydrate, Ranitidine Bismuth citrate, phosphatidylethanolamine, multi vesicular systems

Procedia PDF Downloads 195
2298 Photophysics and Torsional Dynamics of Thioflavin T in Deep Eutectic Solvents

Authors: Rajesh Kumar Gautam, Debabrata Seth

Abstract:

Thioflavin-T (ThT) play a key role of an important biologically active fluorescent sensor for amyloid fibrils. ThT molecule has been developed a method to detect the analysis of different type of diseases such as neurodegenerative disorders, Alzheimer’s, Parkinson’s, and type II diabetes. ThT was used as a fluorescent marker to detect the formation of amyloid fibril. In the presence of amyloid fibril, ThT becomes highly fluorescent. ThT undergoes twisting motion around C-C bonds of the two adjacent benzothiazole and dimethylaniline aromatic rings, which is predominantly affected by the micro-viscosity of the local environment. The present study articulates photophysics and torsional dynamics of biologically active molecule ThT in the presence of deep-eutectic solvents (DESs). DESs are environment-friendly, low cost and biodegradable alternatives to the ionic liquids. DES resembles ionic liquids, but the constituents of a DES include a hydrogen bond donor and acceptor species, in addition to ions. Due to the presence of the H-bonding network within a DES, it exhibits structural heterogeneity. Herein, we have prepared two different DESs by mixing urea with choline chloride and N, N-diethyl ethanol ammonium chloride at ~ 340 K. It was reported that deep eutectic mixture of choline chloride with urea gave a liquid with a freezing point of 12°C. We have experimented by taking two different concentrations of ThT. It was observed that at higher concentration of ThT (50 µM) it forms aggregates in DES. The photophysics of ThT as a function of temperature have been explored by using steady-state, and picoseconds time-resolved fluorescence emission spectroscopic techniques. From the spectroscopic analysis, we have observed that with rising temperature the fluorescence quantum yields and lifetime values of ThT molecule gradually decreases; this is the cumulative effect of thermal quenching and increase in the rate of the torsional rate constant. The fluorescence quantum yield and fluorescence lifetime decay values were always higher for DES-II (urea & N, N-diethyl ethanol ammonium chloride) than those for DES-I (urea & choline chloride). This was mainly due to the presence of structural heterogeneity of the medium. This was further confirmed by comparison with the activation energy of viscous flow with the activation energy of non-radiative decay. ThT molecule in less viscous media undergoes a very fast twisting process and leads to deactivation from the photoexcited state. In this system, the torsional motion increases with increasing temperature. We have concluded that beside bulk viscosity of the media, structural heterogeneity of the medium play crucial role to guide the photophysics of ThT in DESs. The analysis of the experimental data was carried out in the temperature range 288 ≤ T = 333K. The present articulate is to obtain an insight into the DESs as media for studying various photophysical processes of amyloid fibrils sensing molecule of ThT.

Keywords: deep eutectic solvent, photophysics, Thioflavin T, the torsional rate constant

Procedia PDF Downloads 154
2297 Enhancer: An Effective Transformer Architecture for Single Image Super Resolution

Authors: Pitigalage Chamath Chandira Peiris

Abstract:

A widely researched domain in the field of image processing in recent times has been single image super-resolution, which tries to restore a high-resolution image from a single low-resolution image. Many more single image super-resolution efforts have been completed utilizing equally traditional and deep learning methodologies, as well as a variety of other methodologies. Deep learning-based super-resolution methods, in particular, have received significant interest. As of now, the most advanced image restoration approaches are based on convolutional neural networks; nevertheless, only a few efforts have been performed using Transformers, which have demonstrated excellent performance on high-level vision tasks. The effectiveness of CNN-based algorithms in image super-resolution has been impressive. However, these methods cannot completely capture the non-local features of the data. Enhancer is a simple yet powerful Transformer-based approach for enhancing the resolution of images. A method for single image super-resolution was developed in this study, which utilized an efficient and effective transformer design. This proposed architecture makes use of a locally enhanced window transformer block to alleviate the enormous computational load associated with non-overlapping window-based self-attention. Additionally, it incorporates depth-wise convolution in the feed-forward network to enhance its ability to capture local context. This study is assessed by comparing the results obtained for popular datasets to those obtained by other techniques in the domain.

Keywords: single image super resolution, computer vision, vision transformers, image restoration

Procedia PDF Downloads 93
2296 Differentiating Morphological Patterns of the Common Benthic Anglerfishes from the Indian Waters

Authors: M. P. Rajeeshkumar, K. V. Aneesh Kumar, J. L. Otero-Ferrer, A. Lombarte, M. Hashim, N. Saravanane, V. N.Sanjeevan, V. M. Tuset

Abstract:

The anglerfishes are widely distributed from shallow to deep-water habitats and are highly diverse in morphology, behaviour, and niche occupancy patterns. To understand this interspecific variability and degree of niche overlap, we performed a functional analysis of five species inhabiting Indian waters where diversity of deep-sea anglerfishes is very high. The sensory capacities (otolith shape and eye size) were also studied to improve the understanding of coexistence of species. The analyses of fish body and otolith shape clustered species in two morphotypes related to phylogenetic lineages: i) Malthopsis lutea, Lophiodes lugubri and Halieutea coccinea were characterized by a dorso-ventrally flattened body with high swimming ability and relative small otoliths, and ii) Chaunax spp. were distinguished by their higher body depth, lower swimming efficiency, and relative big otoliths. The sensory organs did not show a pattern linked to depth distribution of species. However, the larger eye size in M. lutea suggested a nocturnal feeding activity, whereas Chaunax spp. had a large mouth and deeper body in response to different ecological niches. Therefore, the present study supports the hypothesis of spatial and temporal segregation of anglerfishes in the Indian waters, which can be explained from a functional approach and understanding from sensory capabilities.

Keywords: functional traits, otoliths, niche overlap, fishes, Indian waters

Procedia PDF Downloads 120
2295 GA3C for Anomalous Radiation Source Detection

Authors: Chia-Yi Liu, Bo-Bin Xiao, Wen-Bin Lin, Hsiang-Ning Wu, Liang-Hsun Huang

Abstract:

In order to reduce the risk of radiation damage that personnel may suffer during operations in the radiation environment, the use of automated guided vehicles to assist or replace on-site personnel in the radiation environment has become a key technology and has become an important trend. In this paper, we demonstrate our proof of concept for autonomous self-learning radiation source searcher in an unknown environment without a map. The research uses GPU version of Asynchronous Advantage Actor-Critic network (GA3C) of deep reinforcement learning to search for radiation sources. The searcher network, based on GA3C architecture, has self-directed learned and improved how search the anomalous radiation source by training 1 million episodes under three simulation environments. In each episode of training, the radiation source position, the radiation source intensity, starting position, are all set randomly in one simulation environment. The input for searcher network is the fused data from a 2D laser scanner and a RGB-D camera as well as the value of the radiation detector. The output actions are the linear and angular velocities. The searcher network is trained in a simulation environment to accelerate the learning process. The well-performance searcher network is deployed to the real unmanned vehicle, Dashgo E2, which mounts LIDAR of YDLIDAR G4, RGB-D camera of Intel D455, and radiation detector made by Institute of Nuclear Energy Research. In the field experiment, the unmanned vehicle is enable to search out the radiation source of the 18.5MBq Na-22 by itself and avoid obstacles simultaneously without human interference.

Keywords: deep reinforcement learning, GA3C, source searching, source detection

Procedia PDF Downloads 101
2294 A Critical Review of Assessments of Geological CO2 Storage Resources in Pennsylvania and the Surrounding Region

Authors: Levent Taylan Ozgur Yildirim, Qihao Qian, John Yilin Wang

Abstract:

A critical review of assessments of geological carbon dioxide (CO2) storage resources in Pennsylvania and the surrounding region was completed with a focus on the studies of Midwest Regional Carbon Sequestration Partnership (MRCSP), United States Department of Energy (US-DOE), and United States Geological Survey (USGS). Pennsylvania Geological Survey participated in the MRCSP Phase I research to characterize potential storage formations in Pennsylvania. The MRCSP’s volumetric method estimated ~89 gigatonnes (Gt) of total CO2 storage resources in deep saline formations, depleted oil and gas reservoirs, coals, and shales in Pennsylvania. Meanwhile, the US-DOE calculated storage efficiency factors using log-odds normal distribution and Monte Carlo sampling, revealing contingent storage resources of ~18 Gt to ~20 Gt in deep saline formations, depleted oil and gas reservoirs, and coals in Pennsylvania. Additionally, the USGS employed Beta-PERT distribution and Monte Carlo sampling to determine buoyant and residual storage efficiency factors, resulting in 20 Gt of contingent storage resources across four storage assessment units in Appalachian Basin. However, few studies have explored CO2 storage resources in shales in the region, yielding inconclusive findings. This article provides a critical and most up to date review and analysis of geological CO2 storage resources in Pennsylvania and the region.

Keywords: carbon capture and storage, geological CO2 storage, pennsylvania, appalachian basin

Procedia PDF Downloads 32
2293 Numerical Methodology to Support the Development of a Double Chamber Syringe

Authors: Lourenço Bastos, Filipa Carneiro, Bruno Vale, Rita Marques Joana Silva, Ricardo Freitas, Ângelo Marques, Sara Cortez, Alberta Coelho, Pedro Parreira, Liliana Sousa, Anabela Salgueiro, Bruno Silva

Abstract:

The process of flushing is considered to be an adequate technique to reduce the risk of infection during the clinical practice of venous catheterization. Nonetheless, there is still a lack of adhesion to this method, in part due to the complexity of this procedure. The project SeringaDuo aimed to develop an innovative double-chamber syringe for intravenous sequential administration of drugs and serums. This device served the purpose of improving the adherence to the practice, through the reduction of manipulations needed, which also improves patient safety, and though the promotion of flushing practice by health professionals, by simplifying this task. To assist on the development of this innovative syringe, a numerical methodology was developed and validated in order to predict the syringe’s mechanical and flow behavior during the fluids’ loading and administration phases, as well as to allow the material behavior evaluation during its production. For this, three commercial numerical simulation software was used, namely ABAQUS, ANSYS/FLUENT, and MOLDFLOW. This methodology aimed to evaluate the concepts feasibility and to optimize the geometries of the syringe’s components, creating this way an iterative process for product development based on numerical simulations, validated by the production of prototypes. Through this methodology, it was possible to achieve a final design that fulfils all the characteristics and specifications defined. This iterative process based on numerical simulations is a powerful tool for product development that allows obtaining fast and accurate results without the strict need for prototypes. An iterative process can be implemented, consisting of consecutive constructions and evaluations of new concepts, to obtain an optimized solution, which fulfils all the predefined specifications and requirements.

Keywords: Venous catheterization, flushing, syringe, numerical simulation

Procedia PDF Downloads 154
2292 Analyzing Natural and Social Resources for the Planning of Complex Development Based on Ecotourism: A Case Study from Hungary and Slovakia

Authors: Barnabás Körmöndi

Abstract:

The recent crises have affected societies worldwide, resulting in the irresponsible exploitation of natural resources and the unattainability of sustainability. Regions that are economically underdeveloped, such as the Bodrogköz in Eastern Hungary and Slovakia, experience these issues more severely. The aim of this study is to analyze the natural and social resources of the Bodrogköz area for the planning of complex development based on ecotourism. The objective is to develop ecotourism opportunities in this least developed area of the borderland of Hungary and Slovakia. The study utilizes desk research, deep interviews, focus group meetings, and remote sensing methods. Desk research is aimed at providing a comprehensive understanding of the area, while deep interviews and focus group meetings were conducted to understand the stakeholders' perspectives on the potential for ecotourism. Remote sensing methods were used to better understand changes in the natural environment. The study identified the potential for ecotourism development in the Bodrogköz area due to its near-natural habitats along its bordering rivers and rich cultural heritage. The analysis revealed that ecotourism could promote the region's sustainable development, which is essential for its economic growth. Additionally, the study identified the possible threats to the natural environment during ecotourism development and suggested strategies to mitigate these threats. This study highlights the significance of ecotourism in promoting sustainable development in underdeveloped areas such as the Bodrogköz. It provides a basis for future research on ecotourism development and sustainable planning in similar regions. The analysis is based on the data collected through desk research, deep interviews, focus group meetings, and remote sensing. The assessment was conducted through content analysis, which allowed for the identification of themes and patterns in the data. The study addressed the question of how to develop ecotourism in the least developed area of the borderland of Hungary and Slovakia and promote sustainable development in the region. In conclusion, the study highlights the potential for ecotourism development in Bodrogköz and identifies the natural and social resources that contribute to its development. The study emphasizes the need for sustainable development to promote economic growth and mitigate any environmental threats. The findings can inform the development of future strategic plans for ecotourism, promoting sustainable development in underdeveloped regions.

Keywords: ecotourism, natural resources, remote sensing, social development

Procedia PDF Downloads 53
2291 Enhancing Temporal Extrapolation of Wind Speed Using a Hybrid Technique: A Case Study in West Coast of Denmark

Authors: B. Elshafei, X. Mao

Abstract:

The demand for renewable energy is significantly increasing, major investments are being supplied to the wind power generation industry as a leading source of clean energy. The wind energy sector is entirely dependable and driven by the prediction of wind speed, which by the nature of wind is very stochastic and widely random. This s0tudy employs deep multi-fidelity Gaussian process regression, used to predict wind speeds for medium term time horizons. Data of the RUNE experiment in the west coast of Denmark were provided by the Technical University of Denmark, which represent the wind speed across the study area from the period between December 2015 and March 2016. The study aims to investigate the effect of pre-processing the data by denoising the signal using empirical wavelet transform (EWT) and engaging the vector components of wind speed to increase the number of input data layers for data fusion using deep multi-fidelity Gaussian process regression (GPR). The outcomes were compared using root mean square error (RMSE) and the results demonstrated a significant increase in the accuracy of predictions which demonstrated that using vector components of the wind speed as additional predictors exhibits more accurate predictions than strategies that ignore them, reflecting the importance of the inclusion of all sub data and pre-processing signals for wind speed forecasting models.

Keywords: data fusion, Gaussian process regression, signal denoise, temporal extrapolation

Procedia PDF Downloads 128
2290 Innovative Preparation Techniques: Boosting Oral Bioavailability of Phenylbutyric Acid Through Choline Salt-Based API-Ionic Liquids and Therapeutic Deep Eutectic Systems

Authors: Lin Po-Hsi, Sheu Ming-Thau

Abstract:

Urea cycle disorders (UCD) are rare genetic metabolic disorders that compromise the body's urea cycle. Sodium phenylbutyrate (SPB) is a medication commonly administered in tablet or powder form to lower ammonia levels. Nonetheless, its high sodium content poses risks to sodium-sensitive UCD patients. This necessitates the creation of an alternative drug formulation to mitigate sodium load and optimize drug delivery for UCD patients. This study focused on crafting a novel oral drug formulation for UCD, leveraging choline bicarbonate and phenylbutyric acid. The active pharmaceutical ingredient-ionic liquids (API-ILs) and therapeutic deep eutectic systems (THEDES) were formed by combining these with choline chloride. These systems display characteristics like maintaining a liquid state at room temperature and exhibiting enhanced solubility. This in turn amplifies drug dissolution rate, permeability, and ultimately oral bioavailability. Incorporating choline-based phenylbutyric acid as a substitute for traditional SPB can effectively curtail the sodium load in UCD patients. Our in vitro dissolution experiments revealed that the ILs and DESs, synthesized using choline bicarbonate and choline chloride with phenylbutyric acid, surpassed commercial tablets in dissolution speed. Pharmacokinetic evaluations in SD rats indicated a notable uptick in the oral bioavailability of phenylbutyric acid, underscoring the efficacy of choline salt ILs in augmenting its bioavailability. Additional in vitro intestinal permeability tests on SD rats authenticated that the ILs, formulated with choline bicarbonate and phenylbutyric acid, demonstrate superior permeability compared to their sodium and acid counterparts. To conclude, choline salt ILs developed from choline bicarbonate and phenylbutyric acid present a promising avenue for UCD treatment, with the added benefit of reduced sodium load. They also hold merit in formulation engineering. The sustained-release capabilities of DESs position them favorably for drug delivery, while the low toxicity and cost-effectiveness of choline chloride signal potential in formulation engineering. Overall, this drug formulation heralds a prospective therapeutic avenue for UCD patients.

Keywords: phenylbutyric acid, sodium phenylbutyrate, choline salt, ionic liquids, deep eutectic systems, oral bioavailability

Procedia PDF Downloads 97
2289 Big Data in Telecom Industry: Effective Predictive Techniques on Call Detail Records

Authors: Sara ElElimy, Samir Moustafa

Abstract:

Mobile network operators start to face many challenges in the digital era, especially with high demands from customers. Since mobile network operators are considered a source of big data, traditional techniques are not effective with new era of big data, Internet of things (IoT) and 5G; as a result, handling effectively different big datasets becomes a vital task for operators with the continuous growth of data and moving from long term evolution (LTE) to 5G. So, there is an urgent need for effective Big data analytics to predict future demands, traffic, and network performance to full fill the requirements of the fifth generation of mobile network technology. In this paper, we introduce data science techniques using machine learning and deep learning algorithms: the autoregressive integrated moving average (ARIMA), Bayesian-based curve fitting, and recurrent neural network (RNN) are employed for a data-driven application to mobile network operators. The main framework included in models are identification parameters of each model, estimation, prediction, and final data-driven application of this prediction from business and network performance applications. These models are applied to Telecom Italia Big Data challenge call detail records (CDRs) datasets. The performance of these models is found out using a specific well-known evaluation criteria shows that ARIMA (machine learning-based model) is more accurate as a predictive model in such a dataset than the RNN (deep learning model).

Keywords: big data analytics, machine learning, CDRs, 5G

Procedia PDF Downloads 128
2288 Mechanical and Optical Properties of Doped Aluminum Nitride Thin Films

Authors: Padmalochan Panda, R. Ramaseshan

Abstract:

Aluminum nitride (AlN) is a potential candidate for semiconductor industry due to its wide band gap (6.2 eV), high thermal conductivity and low thermal coefficient of expansion. A-plane oriented AlN film finds an important role in deep UV-LED with higher isotropic light extraction efficiency. Also, Cr-doped AlN films exhibit dilute magnetic semiconductor property with high Curie temperature (300 K), and thus compatible with modern day microelectronics. In this work, highly a-axis oriented wurtzite AlN and Al1-xMxN (M = Cr, Ti) films have synthesized by reactive co-sputtering technique at different concentration. Crystal structure of these films is studied by Grazing incidence X-ray diffraction (GIXRD) and Transmission electron microscopy (TEM). Identification of binding energy and concentration (x) in these films is carried out by X-ray photoelectron spectroscopy (XPS). Local crystal structure around the Cr and Ti atom of these films are investigated by X-ray absorption spectroscopy (XAS). It is found that Cr and Ti replace the Al atom in AlN lattice and the bond lengths in first and second coordination sphere with N and Al, respectively, decrease concerning doping concentration due to strong p-d hybridization. The nano-indentation hardness of Cr and Ti-doped AlN films seems to increase from 17.5 GPa (AlN) to around 23 and 27.5 GPa, respectively. An-isotropic optical properties of these films are studied by the Spectroscopic Ellipsometry technique. Refractive index and extinction coefficient of these films are enhanced in normal dispersion region as compared to the parent AlN film. The optical band gap energies also seem to vary between deep UV to UV regions with the addition of Cr, thus by bringing out the usefulness of these films in the area of optoelectronic device applications.

Keywords: ellipsometry, GIXRD, hardness, XAS

Procedia PDF Downloads 106
2287 Optimized Deep Learning-Based Facial Emotion Recognition System

Authors: Erick C. Valverde, Wansu Lim

Abstract:

Facial emotion recognition (FER) system has been recently developed for more advanced computer vision applications. The ability to identify human emotions would enable smart healthcare facility to diagnose mental health illnesses (e.g., depression and stress) as well as better human social interactions with smart technologies. The FER system involves two steps: 1) face detection task and 2) facial emotion recognition task. It classifies the human expression in various categories such as angry, disgust, fear, happy, sad, surprise, and neutral. This system requires intensive research to address issues with human diversity, various unique human expressions, and variety of human facial features due to age differences. These issues generally affect the ability of the FER system to detect human emotions with high accuracy. Early stage of FER systems used simple supervised classification task algorithms like K-nearest neighbors (KNN) and artificial neural networks (ANN). These conventional FER systems have issues with low accuracy due to its inefficiency to extract significant features of several human emotions. To increase the accuracy of FER systems, deep learning (DL)-based methods, like convolutional neural networks (CNN), are proposed. These methods can find more complex features in the human face by means of the deeper connections within its architectures. However, the inference speed and computational costs of a DL-based FER system is often disregarded in exchange for higher accuracy results. To cope with this drawback, an optimized DL-based FER system is proposed in this study.An extreme version of Inception V3, known as Xception model, is leveraged by applying different network optimization methods. Specifically, network pruning and quantization are used to enable lower computational costs and reduce memory usage, respectively. To support low resource requirements, a 68-landmark face detector from Dlib is used in the early step of the FER system.Furthermore, a DL compiler is utilized to incorporate advanced optimization techniques to the Xception model to improve the inference speed of the FER system. In comparison to VGG-Net and ResNet50, the proposed optimized DL-based FER system experimentally demonstrates the objectives of the network optimization methods used. As a result, the proposed approach can be used to create an efficient and real-time FER system.

Keywords: deep learning, face detection, facial emotion recognition, network optimization methods

Procedia PDF Downloads 112
2286 Trajectory Optimization for Autonomous Deep Space Missions

Authors: Anne Schattel, Mitja Echim, Christof Büskens

Abstract:

Trajectory planning for deep space missions has become a recent topic of great interest. Flying to space objects like asteroids provides two main challenges. One is to find rare earth elements, the other to gain scientific knowledge of the origin of the world. Due to the enormous spatial distances such explorer missions have to be performed unmanned and autonomously. The mathematical field of optimization and optimal control can be used to realize autonomous missions while protecting recourses and making them safer. The resulting algorithms may be applied to other, earth-bound applications like e.g. deep sea navigation and autonomous driving as well. The project KaNaRiA ('Kognitionsbasierte, autonome Navigation am Beispiel des Ressourcenabbaus im All') investigates the possibilities of cognitive autonomous navigation on the example of an asteroid mining mission, including the cruise phase and approach as well as the asteroid rendezvous, landing and surface exploration. To verify and test all methods an interactive, real-time capable simulation using virtual reality is developed under KaNaRiA. This paper focuses on the specific challenge of the guidance during the cruise phase of the spacecraft, i.e. trajectory optimization and optimal control, including first solutions and results. In principle there exist two ways to solve optimal control problems (OCPs), the so called indirect and direct methods. The indirect methods are being studied since several decades and their usage needs advanced skills regarding optimal control theory. The main idea of direct approaches, also known as transcription techniques, is to transform the infinite-dimensional OCP into a finite-dimensional non-linear optimization problem (NLP) via discretization of states and controls. These direct methods are applied in this paper. The resulting high dimensional NLP with constraints can be solved efficiently by special NLP methods, e.g. sequential quadratic programming (SQP) or interior point methods (IP). The movement of the spacecraft due to gravitational influences of the sun and other planets, as well as the thrust commands, is described through ordinary differential equations (ODEs). The competitive mission aims like short flight times and low energy consumption are considered by using a multi-criteria objective function. The resulting non-linear high-dimensional optimization problems are solved by using the software package WORHP ('We Optimize Really Huge Problems'), a software routine combining SQP at an outer level and IP to solve underlying quadratic subproblems. An application-adapted model of impulsive thrusting, as well as a model of an electrically powered spacecraft propulsion system, is introduced. Different priorities and possibilities of a space mission regarding energy cost and flight time duration are investigated by choosing different weighting factors for the multi-criteria objective function. Varying mission trajectories are analyzed and compared, both aiming at different destination asteroids and using different propulsion systems. For the transcription, the robust method of full discretization is used. The results strengthen the need for trajectory optimization as a foundation for autonomous decision making during deep space missions. Simultaneously they show the enormous increase in possibilities for flight maneuvers by being able to consider different and opposite mission objectives.

Keywords: deep space navigation, guidance, multi-objective, non-linear optimization, optimal control, trajectory planning.

Procedia PDF Downloads 402
2285 Influence of Thermal History on the Undrained Shear Strength of the Bentonite-Sand Mixture

Authors: K. Ravi, Sabu Subhash

Abstract:

Densely compacted bentonite or bentonite–sand mixture has been identified as a suitable buffer in the deep geological repository (DGR) for the safe disposal of high-level nuclear waste (HLW) due to its favourable physicochemical and hydro-mechanical properties. The addition of sand to the bentonite enhances the thermal conductivity and compaction properties and reduces the drying shrinkage of the buffer material. The buffer material may undergo cyclic wetting and drying upon ingress of groundwater from the surrounding rock mass and from evaporation due to high temperature (50–210 °C) derived from the waste canister. The cycles of changes in temperature may result in thermal history, and the hydro-mechanical properties of the buffer material may be affected. This paper examines the influence of thermal history on the undrained shear strength of bentonite and bentonite-sand mixture. Bentonite from Rajasthan state and sand from the Assam state of India are used in this study. The undrained shear strength values are obtained by conducting unconfined compressive strength (UCS) tests on cylindrical specimens (dry densities 1.30 and 1.5 Mg/m3) of bentonite and bentonite-sand mixture consisting of 30 % bentonite+ 70 % sand. The specimens are preheated at temperatures varying from 50-150 °C for one, two and four hours in hot air oven. The results indicate that the undrained shear strength is increased by the thermal history of the buffer material. The specimens of bentonite-sand mixture exhibited more increase in strength compared to the pure bentonite specimens. This indicates that the sand content of the mixture plays a vital role in taking the thermal stresses of the bentonite buffer in DGR conditions.

Keywords: bentonite, deep geological repository, thermal history, undrained shear strength

Procedia PDF Downloads 335