Search results for: Frank Henning
20 Palladium-Catalyzed Hydrodechlorination for Water Remediation: Catalyst Deactivation and Regeneration
Authors: Dalia Angeles-Wedler, Katrin Mackenzie, Frank-Dieter Kopinke
Abstract:
Palladium-catalyzed hydrodechlorination is a promising alternative for the treatment of environmentally relevant water bodies, such as groundwater, contaminated with chlorinated organic compounds (COCs). In the aqueous phase hydrodechlorination of COCs, Pd-based catalysts were found to have a very high catalytic activity. However, the full utilization of the catalyst-s potential is impeded by the sensitivity of the catalyst to poisoning and deactivation induced by reduced sulfur compounds (e.g. sulfides). Several regenerants have been tested before to recover the performance of sulfide-fouled Pd catalyst. But these only delivered partial success with respect to re-establishment of the catalyst activity. In this study, the deactivation behaviour of Pd/Al2O3 in the presence of sulfide was investigated. Subsequent to total deactivation the catalyst was regenerated in the aqueous phase using potassium permanganate. Under neutral pH condition, oxidative regeneration with permanganate delivered a slow recovery of catalyst activity. However, changing the pH of the bulk solution to acidic resulted in the complete recovery of catalyst activity within a regeneration time of about half an hour. These findings suggest the superiority of permanganate as regenerant in re-activating Pd/Al2O3 by oxidizing Pd-bound sulfide.Keywords: Deactivation, hydrodechlorination, Pd catalyst, regeneration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 226619 Advanced Energy Absorbers Used in Blast Resistant Systems
Authors: Martina Drdlová, Michal Frank, Radek Řídký, Jaroslav Buchar, Josef Krátký
Abstract:
The main aim of the presented experiments is to improve behaviour of sandwich structures under dynamic loading, such as crash or explosion. This paper describes experimental investigation on the response of new advanced materials to low and high velocity load. Blast wave energy absorbers were designed using two types of porous lightweight raw particle materials based on expanded glass and ceramics with dimensions of 0.5-1 mm, combined with polymeric binder. The effect of binder amount on the static and dynamic properties of designed materials was observed. Prism shaped specimens were prepared and loaded to obtain physicomechanical parameters – bulk density, compressive and flexural strength under quasistatic load, the dynamic response was determined using Split Hopkinson Pressure bar apparatus. Numerical investigation of the material behaviour in sandwich structure was performed using implicit/explicit solver LS-Dyna. As the last step, the developed material was used as the interlayer of blast resistant litter bin, and it´s functionality was verified by real field blast tests.
Keywords: Blast energy absorber, SHPB, expanded glass, expanded ceramics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 243718 Protein Secondary Structure Prediction Using Parallelized Rule Induction from Coverings
Authors: Leong Lee, Cyriac Kandoth, Jennifer L. Leopold, Ronald L. Frank
Abstract:
Protein 3D structure prediction has always been an important research area in bioinformatics. In particular, the prediction of secondary structure has been a well-studied research topic. Despite the recent breakthrough of combining multiple sequence alignment information and artificial intelligence algorithms to predict protein secondary structure, the Q3 accuracy of various computational prediction algorithms rarely has exceeded 75%. In a previous paper [1], this research team presented a rule-based method called RT-RICO (Relaxed Threshold Rule Induction from Coverings) to predict protein secondary structure. The average Q3 accuracy on the sample datasets using RT-RICO was 80.3%, an improvement over comparable computational methods. Although this demonstrated that RT-RICO might be a promising approach for predicting secondary structure, the algorithm-s computational complexity and program running time limited its use. Herein a parallelized implementation of a slightly modified RT-RICO approach is presented. This new version of the algorithm facilitated the testing of a much larger dataset of 396 protein domains [2]. Parallelized RTRICO achieved a Q3 score of 74.6%, which is higher than the consensus prediction accuracy of 72.9% that was achieved for the same test dataset by a combination of four secondary structure prediction methods [2].Keywords: data mining, protein secondary structure prediction, parallelization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 159517 Application of the Transtheoretical Model of Exercise Behavior Change Plan in High School Students
Authors: Frank C. Pan, Chih-Hao Chen
Abstract:
The purpose of this study is to discuss the effect of the intervention of exercise behavior change plan for high school students on study subjects- social and psychological factors and exercise stages. This research uses the transtheoretical model as the research framework. One experiment group and one control group were used in a quasi-experimental design research. The experimental group accepted health-related physical fitness course and the traditional course; the control group accepted traditional physical education course. There is a significant difference before and after the intervention in the experimental group. Karl-s test shows the experimental group gained a better improvement than that in the control group. The Analysis of Covariance had shown the exercise stages (F=7.62, p<0.05), and the perceived exercise benefit (F=16.91, p<0.05) of the experimental group were obviously larger than those of the control group. There was no major effect on the perceived exercise barrier (F=0.61, p>0.05) after the intervention of exercise behavior change plan. However, the rate of high school students the practical implementation of the exercise is not high, obviously influenced by the exercise barrier. The academic and examination pressure may be the major causes of the exercise barrier in the high school.Keywords: Transtheoretical model (TTM), High school students, exercise behavior
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 397816 Maternal Health Outcome and Economic Growth in Sub-Saharan Africa: A Dynamic Panel Analysis
Authors: Okwan Frank
Abstract:
Maternal health outcome is one of the major population development challenges in Sub-Saharan Africa. The region has the highest maternal mortality ratio, despite the progressive economic growth in the region during the global economic crisis. It has been hypothesized that increase in economic growth will reduce the level of maternal mortality. The purpose of this study is to investigate the existence of the negative relationship between health outcome proxy by maternal mortality ratio and economic growth in Sub-Saharan Africa. The study used the Pooled Mean Group estimator of ARDL Autoregressive Distributed Lag (ARDL) and the Kao test for cointegration to examine the short-run and long-run relationship between maternal mortality and economic growth. The results of the cointegration test showed the existence of a long-run relationship between the variables considered for the study. The long-run result of the Pooled Mean group estimates confirmed the hypothesis of an inverse relationship between maternal health outcome proxy by maternal mortality ratio and economic growth proxy by Gross Domestic Product (GDP) per capita. Thus increasing economic growth by investing in the health care systems to reduce pregnancy and childbirth complications will help reduce maternal mortality in the sub-region.
Keywords: Economic growth, maternal mortality, pool mean group, Sub-Saharan Africa.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 58515 A Systems Approach to Gene Ranking from DNA Microarray Data of Cervical Cancer
Authors: Frank Emmert Streib, Matthias Dehmer, Jing Liu, Max Mühlhauser
Abstract:
In this paper we present a method for gene ranking from DNA microarray data. More precisely, we calculate the correlation networks, which are unweighted and undirected graphs, from microarray data of cervical cancer whereas each network represents a tissue of a certain tumor stage and each node in the network represents a gene. From these networks we extract one tree for each gene by a local decomposition of the correlation network. The interpretation of a tree is that it represents the n-nearest neighbor genes on the n-th level of a tree, measured by the Dijkstra distance, and, hence, gives the local embedding of a gene within the correlation network. For the obtained trees we measure the pairwise similarity between trees rooted by the same gene from normal to cancerous tissues. This evaluates the modification of the tree topology due to progression of the tumor. Finally, we rank the obtained similarity values from all tissue comparisons and select the top ranked genes. For these genes the local neighborhood in the correlation networks changes most between normal and cancerous tissues. As a result we find that the top ranked genes are candidates suspected to be involved in tumor growth and, hence, indicates that our method captures essential information from the underlying DNA microarray data of cervical cancer.Keywords: Graph similarity, DNA microarray data, cancer.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 175514 Vehicle Risk Evaluation in Low Speed Accidents: Consequences for Relevant Test Scenarios
Authors: Philip Feig, Klaus Gschwendtner, Julian Schatz, Frank Diermeyer
Abstract:
Projects of accident research analysis are mostly focused on accidents involving personal damage. Property damage only has a high frequency of occurrence combined with high economic impact. This paper describes main influencing parameters for the extent of damage and presents a repair cost model. For a prospective evaluation method of the monetary effect of advanced driver assistance systems (ADAS), it is necessary to be aware of and quantify all influencing parameters. Furthermore, this method allows the evaluation of vehicle concepts in combination with an ADAS at an early point in time of the product development process. In combination with a property damage database and the introduced repair cost model relevant test scenarios for specific vehicle configurations and their individual property damage risk may be determined. Currently, equipment rates of ADAS are low and a purchase incentive for customers would be beneficial. The next ADAS generation will prevent property damage to a large extent or at least reduce damage severity. Both effects may be a purchasing incentive for the customer and furthermore contribute to increased traffic safety.Keywords: Property damage analysis, effectiveness, ADAS, damage risk, accident research, accident scenarios.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 138713 Towards Clustering of Web-based Document Structures
Authors: Matthias Dehmer, Frank Emmert Streib, Jürgen Kilian, Andreas Zulauf
Abstract:
Methods for organizing web data into groups in order to analyze web-based hypertext data and facilitate data availability are very important in terms of the number of documents available online. Thereby, the task of clustering web-based document structures has many applications, e.g., improving information retrieval on the web, better understanding of user navigation behavior, improving web users requests servicing, and increasing web information accessibility. In this paper we investigate a new approach for clustering web-based hypertexts on the basis of their graph structures. The hypertexts will be represented as so called generalized trees which are more general than usual directed rooted trees, e.g., DOM-Trees. As a important preprocessing step we measure the structural similarity between the generalized trees on the basis of a similarity measure d. Then, we apply agglomerative clustering to the obtained similarity matrix in order to create clusters of hypertext graph patterns representing navigation structures. In the present paper we will run our approach on a data set of hypertext structures and obtain good results in Web Structure Mining. Furthermore we outline the application of our approach in Web Usage Mining as future work.Keywords: Clustering methods, graph-based patterns, graph similarity, hypertext structures, web structure mining
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 150512 Application of a Similarity Measure for Graphs to Web-based Document Structures
Authors: Matthias Dehmer, Frank Emmert Streib, Alexander Mehler, Jürgen Kilian, Max Mühlhauser
Abstract:
Due to the tremendous amount of information provided by the World Wide Web (WWW) developing methods for mining the structure of web-based documents is of considerable interest. In this paper we present a similarity measure for graphs representing web-based hypertext structures. Our similarity measure is mainly based on a novel representation of a graph as linear integer strings, whose components represent structural properties of the graph. The similarity of two graphs is then defined as the optimal alignment of the underlying property strings. In this paper we apply the well known technique of sequence alignments for solving a novel and challenging problem: Measuring the structural similarity of generalized trees. In other words: We first transform our graphs considered as high dimensional objects in linear structures. Then we derive similarity values from the alignments of the property strings in order to measure the structural similarity of generalized trees. Hence, we transform a graph similarity problem to a string similarity problem for developing a efficient graph similarity measure. We demonstrate that our similarity measure captures important structural information by applying it to two different test sets consisting of graphs representing web-based document structures.Keywords: Graph similarity, hierarchical and directed graphs, hypertext, generalized trees, web structure mining.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 189011 Measuring the Structural Similarity of Web-based Documents: A Novel Approach
Authors: Matthias Dehmer, Frank Emmert Streib, Alexander Mehler, Jürgen Kilian
Abstract:
Most known methods for measuring the structural similarity of document structures are based on, e.g., tag measures, path metrics and tree measures in terms of their DOM-Trees. Other methods measures the similarity in the framework of the well known vector space model. In contrast to these we present a new approach to measuring the structural similarity of web-based documents represented by so called generalized trees which are more general than DOM-Trees which represent only directed rooted trees.We will design a new similarity measure for graphs representing web-based hypertext structures. Our similarity measure is mainly based on a novel representation of a graph as strings of linear integers, whose components represent structural properties of the graph. The similarity of two graphs is then defined as the optimal alignment of the underlying property strings. In this paper we apply the well known technique of sequence alignments to solve a novel and challenging problem: Measuring the structural similarity of generalized trees. More precisely, we first transform our graphs considered as high dimensional objects in linear structures. Then we derive similarity values from the alignments of the property strings in order to measure the structural similarity of generalized trees. Hence, we transform a graph similarity problem to a string similarity problem. We demonstrate that our similarity measure captures important structural information by applying it to two different test sets consisting of graphs representing web-based documents.
Keywords: Graph similarity, hierarchical and directed graphs, hypertext, generalized trees, web structure mining.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 255610 Ranking Genes from DNA Microarray Data of Cervical Cancer by a local Tree Comparison
Authors: Frank Emmert-Streib, Matthias Dehmer, Jing Liu, Max Muhlhauser
Abstract:
The major objective of this paper is to introduce a new method to select genes from DNA microarray data. As criterion to select genes we suggest to measure the local changes in the correlation graph of each gene and to select those genes whose local changes are largest. More precisely, we calculate the correlation networks from DNA microarray data of cervical cancer whereas each network represents a tissue of a certain tumor stage and each node in the network represents a gene. From these networks we extract one tree for each gene by a local decomposition of the correlation network. The interpretation of a tree is that it represents the n-nearest neighbor genes on the n-th level of a tree, measured by the Dijkstra distance, and, hence, gives the local embedding of a gene within the correlation network. For the obtained trees we measure the pairwise similarity between trees rooted by the same gene from normal to cancerous tissues. This evaluates the modification of the tree topology due to tumor progression. Finally, we rank the obtained similarity values from all tissue comparisons and select the top ranked genes. For these genes the local neighborhood in the correlation networks changes most between normal and cancerous tissues. As a result we find that the top ranked genes are candidates suspected to be involved in tumor growth. This indicates that our method captures essential information from the underlying DNA microarray data of cervical cancer.
Keywords: Graph similarity, generalized trees, graph alignment, DNA microarray data, cervical cancer.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17529 Understanding and Designing Situation-Aware Mobile and Ubiquitous Computing Systems
Authors: Kai Häussermann, Christoph Hubig, Paul Levi, Frank Leymann, Oliver Siemoneit, Matthias Wieland, Oliver Zweigle
Abstract:
Using spatial models as a shared common basis of information about the environment for different kinds of contextaware systems has been a heavily researched topic in the last years. Thereby the research focused on how to create, to update, and to merge spatial models so as to enable highly dynamic, consistent and coherent spatial models at large scale. In this paper however, we want to concentrate on how context-aware applications could use this information so as to adapt their behavior according to the situation they are in. The main idea is to provide the spatial model infrastructure with a situation recognition component based on generic situation templates. A situation template is – as part of a much larger situation template library – an abstract, machinereadable description of a certain basic situation type, which could be used by different applications to evaluate their situation. In this paper, different theoretical and practical issues – technical, ethical and philosophical ones – are discussed important for understanding and developing situation dependent systems based on situation templates. A basic system design is presented which allows for the reasoning with uncertain data using an improved version of a learning algorithm for the automatic adaption of situation templates. Finally, for supporting the development of adaptive applications, we present a new situation-aware adaptation concept based on workflows.Keywords: context-awareness, ethics, facilitation of system use through workflows, situation recognition and learning based on situation templates and situation ontology's, theory of situationaware systems
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17588 Fault Tolerant (n, k)-Star Power Network Topology for Multi-Agent Communication in Automated Power Distribution Systems
Authors: Ning Gong, Michael Korostelev, Qiangguo Ren, Li Bai, Saroj Biswas, Frank Ferrese
Abstract:
This paper investigates the joint effect of the interconnected (n,k)-star network topology and Multi-Agent automated control on restoration and reconfiguration of power systems. With the increasing trend in development in Multi-Agent control technologies applied to power system reconfiguration in presence of faulty components or nodes. Fault tolerance is becoming an important challenge in the design processes of the distributed power system topology. Since the reconfiguration of a power system is performed by agent communication, the (n,k)-star interconnected network topology is studied and modeled in this paper to optimize the process of power reconfiguration. In this paper, we discuss the recently proposed (n,k)-star topology and examine its properties and advantages as compared to the traditional multi-bus power topologies. We design and simulate the topology model for distributed power system test cases. A related lemma based on the fault tolerance and conditional diagnosability properties is presented and proved both theoretically and practically. The conclusion is reached that (n,k)-star topology model has measurable advantages compared to standard bus power systems while exhibiting fault tolerance properties in power restoration, as well as showing efficiency when applied to power system route discovery.
Keywords: (n, k)-star Topology, Fault Tolerance, Conditional Diagnosability, Multi-Agent System, Automated Power System.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24487 Degradation of Heating, Ventilation, and Air Conditioning Components across Locations
Authors: Timothy E. Frank, Josh R. Aldred, Sophie B. Boulware, Michelle K. Cabonce, Justin H. White
Abstract:
Materials degrade at different rates in different environments depending on factors such as temperature, aridity, salinity, and solar radiation. Therefore, predicting asset longevity depends, in part, on the environmental conditions to which the asset is exposed. Heating, ventilation, and air conditioning (HVAC) systems are critical to building operations yet are responsible for a significant proportion of their energy consumption. HVAC energy use increases substantially with slight operational inefficiencies. Understanding the environmental influences on HVAC degradation in detail will inform maintenance schedules and capital investment, reduce energy use, and increase lifecycle management efficiency. HVAC inspection records spanning 14 years from 21 locations across the United States were compiled and associated with the climate conditions to which they were exposed. Three environmental features were explored in this study: average high temperature, average low temperature, and annual precipitation, as well as four non-environmental features. Initial insights showed no correlations between individual features and the rate of HVAC component degradation. Using neighborhood component analysis, however, the most critical features related to degradation were identified. Two models were considered, and results varied between them. However, longitude and latitude emerged as potentially the best predictors of average HVAC component degradation. Further research is needed to evaluate additional environmental features, increase the resolution of the environmental data, and develop more robust models to achieve more conclusive results.
Keywords: Climate, infrastructure degradation, HVAC, neighborhood component analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1706 Development of a Health Literacy Scale for Chinese-Speaking Adults in Taiwan
Authors: Frank C. Pan, Che-Long Su, Ching-Hsuen Chen
Abstract:
Background, measuring an individual-s Health Literacy is gaining attention, yet no appropriate instrument is available in Taiwan. Measurement tools that were developed and used in western countries may not be appropriate for use in Taiwan due to a different language system. Purpose of this research was to develop a Health Literacy measurement instrument specific for Taiwan adults. Methods, several experts of clinic physicians; healthcare administrators and scholars identified 125 common used health related Chinese phrases from major medical knowledge sources that easy accessible to the public. A five-point Likert scale is used to measure the understanding level of the target population. Such measurement is then used to compare with the correctness of their answers to a health knowledge test for validation. Samples, samples under study were purposefully taken from four groups of people in the northern Pingtung, OPD patients, university students, community residents, and casual visitors to the central park. A set of health knowledge index with 10 questions is used to screen those false responses. A sample size of 686 valid cases out of 776 was then included to construct this scale. An independent t-test was used to examine each individual phrase. The phrases with the highest significance are then identified and retained to compose this scale. Result, a Taiwan Health Literacy Scale (THLS) was finalized with 66 health-related phrases under nine divisions. Cronbach-s alpha of each division is at a satisfactory level of 89% and above. Conclusions, factors significantly differentiate the levels of health literacy are education, female gender, age, family members of stroke victims, experience with patient care, and healthcare professionals in the initial application in this study..Keywords: Health literacy, health knowledge, REALM, THLS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25245 A Posterior Predictive Model-Based Control Chart for Monitoring Healthcare
Authors: Yi-Fan Lin, Peter P. Howley, Frank A. Tuyl
Abstract:
Quality measurement and reporting systems are used in healthcare internationally. In Australia, the Australian Council on Healthcare Standards records and reports hundreds of clinical indicators (CIs) nationally across the healthcare system. These CIs are measures of performance in the clinical setting, and are used as a screening tool to help assess whether a standard of care is being met. Existing analysis and reporting of these CIs incorporate Bayesian methods to address sampling variation; however, such assessments are retrospective in nature, reporting upon the previous six or twelve months of data. The use of Bayesian methods within statistical process control for monitoring systems is an important pursuit to support more timely decision-making. Our research has developed and assessed a new graphical monitoring tool, similar to a control chart, based on the beta-binomial posterior predictive (BBPP) distribution to facilitate the real-time assessment of health care organizational performance via CIs. The BBPP charts have been compared with the traditional Bernoulli CUSUM (BC) chart by simulation. The more traditional “central” and “highest posterior density” (HPD) interval approaches were each considered to define the limits, and the multiple charts were compared via in-control and out-of-control average run lengths (ARLs), assuming that the parameter representing the underlying CI rate (proportion of cases with an event of interest) required estimation. Preliminary results have identified that the BBPP chart with HPD-based control limits provides better out-of-control run length performance than the central interval-based and BC charts. Further, the BC chart’s performance may be improved by using Bayesian parameter estimation of the underlying CI rate.
Keywords: Average run length, Bernoulli CUSUM chart, beta binomial posterior predictive distribution, clinical indicator, health care organization, highest posterior density interval.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8744 Exergetic Optimization on Solid Oxide Fuel Cell Systems
Authors: George N. Prodromidis, Frank A. Coutelieris
Abstract:
Biogas can be currently considered as an alternative option for electricity production, mainly due to its high energy content (hydrocarbon-rich source), its renewable status and its relatively low utilization cost. Solid Oxide Fuel Cell (SOFC) stacks convert fuel’s chemical energy to electricity with high efficiencies and reveal significant advantages on fuel flexibility combined with lower emissions rate, especially when utilize biogas. Electricity production by biogas constitutes a composite problem which incorporates an extensive parametric analysis on numerous dynamic variables. The main scope of the presented study is to propose a detailed thermodynamic model on the optimization of SOFC-based power plants’ operation based on fundamental thermodynamics, energy and exergy balances. This model named THERMAS (THERmodynamic MAthematical Simulation model) incorporates each individual process, during electricity production, mathematically simulated for different case studies that represent real life operational conditions. Also, THERMAS offers the opportunity to choose a great variety of different values for each operational parameter individually, thus allowing for studies within unexplored and experimentally impossible operational ranges. Finally, THERMAS innovatively incorporates a specific criterion concluded by the extensive energy analysis to identify the most optimal scenario per simulated system in exergy terms. Therefore, several dynamical parameters as well as several biogas mixture compositions have been taken into account, to cover all the possible incidents. Towards the optimization process in terms of an innovative OPF (OPtimization Factor), presented here, this research study reveals that systems supplied by low methane fuels can be comparable to these supplied by pure methane. To conclude, such an innovative simulation model indicates a perspective on the optimal design of a SOFC stack based system, in the direction of the commercialization of systems utilizing biogas.
Keywords: Biogas, Exergy, Optimization, SOFC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11983 Miniature Fast Steering Mirrors for Space Optical Communication on NanoSats and CubeSats
Authors: Sylvain Chardon, Timotéo Payre, Hugo Grardel, Yann Quentel, Mathieu Thomachot, Gérald Aigouy, Frank Claeyssen
Abstract:
With the increasing digitalization of society, access to data has become vital and strategic for individuals and nations. In this context, the number of satellite constellation projects is growing drastically worldwide and is a next-generation challenge of the New Space industry. So far, existing satellite constellations have been using radio frequencies (RF) for satellite-to-ground communications, inter-satellite communications, and feeder link communication. However, RF has several limitations, such as limited bandwidth and low protection level. To address these limitations, space optical communication will be the new trend, addressing both very high-speed and secured encrypted communication. Fast Steering Mirrors (FSM) are key components used in optical communication as well as space imagery and for a large field of functions such as Point Ahead Mechanisms (PAM), Raster Scanning, Beam Steering Mirrors (BSM), Fine Pointing Mechanisms (FPM) and Line of Sight stabilization (LOS). The main challenges of space FSM development for optical communication are to propose both a technology and a supply chain relevant for high quantities New Space approach, which requires secured connectivity for high-speed internet, Earth planet observation and monitoring, and mobility applications. CTEC proposes a mini-FSM technology offering a stroke of +/-6 mrad and a resonant frequency of 1700 Hz, with a mass of 50 g. This FSM mechanism is a good candidate for giant constellations and all applications on board NanoSats and CubeSats, featuring a very high level of miniaturization and optimized for New Space high quantities cost efficiency. The use of piezo actuators offers a high resonance frequency for optimal control, with almost zero power consumption in step and stay pointing, and with very high-reliability figures > 0,995 demonstrated over years of recurrent manufacturing for Optronics applications at CTEC.
Keywords: Fast steering mirror, feeder link, line of sight stabilization, optical communication, pointing ahead mechanism, raster scan.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1752 Influence of Transverse Steel and Casting Direction on Shear Response and Ductility of Reinforced Ultra-High Performance Concrete Beams
Authors: Timothy E. Frank, Peter J. Amaddio, Elizabeth D. Decko, Alexis M. Tri, Darcy A. Farrell, Cole M. Landes
Abstract:
Ultra-high performance concrete (UHPC) is a class of cementitious composites with a relatively large percentage of cement generating high compressive strength. Additionally, UHPC contains disbursed fibers, which control crack width, carry the tensile load across narrow cracks, and limit spalling. These characteristics lend themselves to a wide range of structural applications when UHPC members are reinforced with longitudinal steel. Efficient use of fibers and longitudinal steel is required to keep lifecycle cost competitive in reinforced UHPC members; this requires full utilization of both the compressive and tensile qualities of the reinforced cementitious composite. The objective of this study is to investigate the shear response of steel-reinforced UHPC beams to guide design decisions that keep initial costs reasonable, limit serviceability crack widths, and ensure a ductile structural response and failure path. Five small-scale, reinforced UHPC beams were experimentally tested. Longitudinal steel, transverse steel, and casting direction were varied. Results indicate that an increase in transverse steel in short-spanned reinforced UHPC beams provided additional shear capacity and increased the peak load achieved. Beams with very large longitudinal steel reinforcement ratios did not achieve yield and fully utilized the tension properties of the longitudinal steel. Casting the UHPC beams from the end or from the middle affected load-carrying capacity and ductility, but image analysis determined that the fiber orientation was not significantly different. It is believed that the presence of transverse and longitudinal steel reinforcement minimized the effect of different UHPC casting directions. Results support recent recommendations in the literature suggesting that a 1% fiber volume fraction is sufficient within UHPC to prevent spalling and provide compressive fracture toughness under extreme loading conditions.
Keywords: Fiber orientation, reinforced ultra-high performance concrete beams, shear, transverse steel.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2121 Estimation of Relative Subsidence of Collapsible Soils Using Electromagnetic Measurements
Authors: Henok Hailemariam, Frank Wuttke
Abstract:
Collapsible soils are weak soils that appear to be stable in their natural state, normally dry condition, but rapidly deform under saturation (wetting), thus generating large and unexpected settlements which often yield disastrous consequences for structures unwittingly built on such deposits. In this study, a prediction model for the relative subsidence of stressed collapsible soils based on dielectric permittivity measurement is presented. Unlike most existing methods for soil subsidence prediction, this model does not require moisture content as an input parameter, thus providing the opportunity to obtain accurate estimation of the relative subsidence of collapsible soils using dielectric measurement only. The prediction model is developed based on an existing relative subsidence prediction model (which is dependent on soil moisture condition) and an advanced theoretical frequency and temperature-dependent electromagnetic mixing equation (which effectively removes the moisture content dependence of the original relative subsidence prediction model). For large scale sub-surface soil exploration purposes, the spatial sub-surface soil dielectric data over wide areas and high depths of weak (collapsible) soil deposits can be obtained using non-destructive high frequency electromagnetic (HF-EM) measurement techniques such as ground penetrating radar (GPR). For laboratory or small scale in-situ measurements, techniques such as an open-ended coaxial line with widely applicable time domain reflectometry (TDR) or vector network analysers (VNAs) are usually employed to obtain the soil dielectric data. By using soil dielectric data obtained from small or large scale non-destructive HF-EM investigations, the new model can effectively predict the relative subsidence of weak soils without the need to extract samples for moisture content measurement. Some of the resulting benefits are the preservation of the undisturbed nature of the soil as well as a reduction in the investigation costs and analysis time in the identification of weak (problematic) soils. The accuracy of prediction of the presented model is assessed by conducting relative subsidence tests on a collapsible soil at various initial soil conditions and a good match between the model prediction and experimental results is obtained.
Keywords: Collapsible soil, relative subsidence, dielectric permittivity, moisture content.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1115