Search results for: Juergen Markus Lackner
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 78

Search results for: Juergen Markus Lackner

48 Linearly Polarized Single Photon Emission from Nonpolar, Semipolar and Polar Quantum Dots in GaN/InGaN Nanowires

Authors: Snezana Lazic, Zarko Gacevic, Mark Holmes, Ekaterina Chernysheva, Marcus Müller, Peter Veit, Frank Bertram, Juergen Christen, Yasuhiko Arakawa, Enrique Calleja

Abstract:

The study reports how the pencil-like morphology of a homoepitaxially grown GaN nanowire can be exploited for the fabrication of a thin conformal InGaN nanoshell, hosting nonpolar, semipolar and polar single photon sources (SPSs). All three SPS types exhibit narrow emission lines (FWHM~0.35 - 2 meV) and high degrees of linear optical polarization (P > 70%) in the low-temperature micro-photoluminescence (µ-PL) experiments and are characterized by a pronounced antibunching in the photon correlation measurements (gcorrected(2)(0) < 0.3). The quantum-dot-like exciton localization centers induced by compositional fluctuations within the InGaN nanoshell are identified as the driving mechanism for the single photon emission. As confirmed by the low-temperature transmission electron microscopy combined with cathodoluminescence (TEM-CL) study, the crystal region (i.e. non-polar m-, semi-polar r- and polar c-facets) hosting the single photon emitters strongly affects their emission wavelength, which ranges from ultra-violet for the non-polar to visible for the polar SPSs. The photon emission lifetime is also found to be facet-dependent and varies from sub-nanosecond time scales for the non- and semi-polar SPSs to a few nanoseconds for the polar ones. These differences are mainly attributed to facet-dependent indium content and electric field distribution across the hosting InGaN nanoshell. The hereby reported pencil-like InGaN nanoshell is the first single nanostructure able to host all three types of single photon emitters and is thus a promising building block for tunable quantum light devices integrated into future photonic and optoelectronic circuits.

Keywords: GaN nanowire, InGaN nanoshell, linear polarization, nonpolar, semipolar, polar quantum dots, single-photon sources

Procedia PDF Downloads 368
47 High Power Thermal Energy Storage for Industrial Applications Using Phase Change Material Slurry

Authors: Anastasia Stamatiou, Markus Odermatt, Dominic Leemann, Ludger J. Fischer, Joerg Worlitschek

Abstract:

The successful integration of thermal energy storage in industrial processes is expected to play an important role in the energy turnaround. Latent heat storage technologies can offer more compact thermal storage at a constant temperature level, in comparison to conventional, sensible thermal storage technologies. The focus of this study is the development of latent heat storage solutions based on the Phase Change Slurry (PCS) concept. Such systems promise higher energy densities both as refrigerants and as storage media while presenting better heat transfer characteristics than conventional latent heat storage technologies. This technology is expected to deliver high thermal power and high-temperature stability which makes it ideal for storage of process heat. An evaluation of important batch processes in industrial applications set the focus on materials with a melting point in the range of 55 - 90 °C. Aluminium ammonium sulfate dodecahydrate (NH₄Al(SO₄)₂·12H₂O) was chosen as the first interesting PCM for the next steps of this study. The ability of this material to produce slurries at the relevant temperatures was demonstrated in a continuous mode in a laboratory test-rig. Critical operational and design parameters were identified.

Keywords: esters, latent heat storage, phase change materials, thermal properties

Procedia PDF Downloads 276
46 The Evaluation of Superiority of Foot Local Anesthesia Method in Dairy Cows

Authors: Samaneh Yavari, Christiane Pferrer, Elisabeth Engelke, Alexander Starke, Juergen Rehage

Abstract:

Background: Nowadays, bovine limb interventions, especially any claw surgeries, raises selection of the most qualified and appropriate local anesthesia technique applicable for any superficial or deep interventions of the limbs. Currently, two local anesthesia methods of Intravenous Regional Anesthesia (IVRA), as well as Nerve Blocks, have been routine to apply. However, the lack of studies investigating the quality and duration as well as quantity and onset of full (complete) local anesthesia, is noticeable. Therefore, the aim of our study was comparing the onset and quality of both IVRA and our modified NBA at the hind limb of dairy cows. For this abstract, only the onset of full local anesthesia would be consider. Materials and Methods: For that reason, we used six healthy non pregnant non lactating Holestein Frisian cows in a cross-over study design. Those cows divided into two groups to receive IVRA and our modified four-point NBA. For IVRA, 20 ml procaine without epinephrine was injected into the vein digitalis dorsalis communis III and for our modified four-point NBA, 10-15 ml procaine without epinephrine preneurally to the nerves, superficial and deep peroneal as well as lateral and medial branches of metatarsal nerves. For pain stimulation, electrical stimulator Grass S48 was applied. Results: The results of electrical stimuli revealed the faster onset of full local anesthesia (p < 0.05) by application of our modified NBA in comparison to IVRA about 10 minutes. Conclusion and discussion: Despite of available references showing faster onset of foot local anesthesia of IVRA, our study demonstrated that our modified four point NBA not only can be well known as a standard foot local anesthesia method applicable to desensitize the hind limb of dairy cows, but also, selection of this modified validated local anesthesia method can lead to have a faster start of complete desensitization of distal hind limb that is remarkable in any bovine limb interventions under time constraint.

Keywords: IVRA, four point NBA, dairy cow, hind limb, full onset

Procedia PDF Downloads 132
45 Compact Optical Sensors for Harsh Environments

Authors: Branislav Timotijevic, Yves Petremand, Markus Luetzelschwab, Dara Bayat, Laurent Aebi

Abstract:

Optical miniaturized sensors with remote readout are required devices for the monitoring in harsh electromagnetic environments. As an example, in turbo and hydro generators, excessively high vibrations of the end-windings can lead to dramatic damages, imposing very high, additional service costs. A significant change of the generator temperature can also be an indicator of the system failure. Continuous monitoring of vibrations, temperature, humidity, and gases is therefore mandatory. The high electromagnetic fields in the generators impose the use of non-conductive devices in order to prevent electromagnetic interferences and to electrically isolate the sensing element to the electronic readout. Metal-free sensors are good candidates for such systems since they are immune to very strong electromagnetic fields and given the fact that they are non-conductive. We have realized miniature optical accelerometer and temperature sensors for a remote sensing of the harsh environments using the common, inexpensive silicon Micro Electro-Mechanical System (MEMS) platform. Both devices show highly linear response. The accelerometer has a deviation within 1% from the linear fit when tested in a range 0 – 40 g. The temperature sensor can provide the measurement accuracy better than 1 °C in a range 20 – 150 °C. The design of other type of sensors for the environments with high electromagnetic interferences has also been discussed.

Keywords: optical MEMS, temperature sensor, accelerometer, remote sensing, harsh environment

Procedia PDF Downloads 344
44 Defining a Reference Architecture for Predictive Maintenance Systems: A Case Study Using the Microsoft Azure IoT-Cloud Components

Authors: Walter Bernhofer, Peter Haber, Tobias Mayer, Manfred Mayr, Markus Ziegler

Abstract:

Current preventive maintenance measures are cost intensive and not efficient. With the available sensor data of state of the art internet of things devices new possibilities of automated data processing emerge. Current advances in data science and in machine learning enable new, so called predictive maintenance technologies, which empower data scientists to forecast possible system failures. The goal of this approach is to cut expenses in preventive maintenance by automating the detection of possible failures and to improve efficiency and quality of maintenance measures. Additionally, a centralization of the sensor data monitoring can be achieved by using this approach. This paper describes the approach of three students to define a reference architecture for a predictive maintenance solution in the internet of things domain with a connected smartphone app for service technicians. The reference architecture is validated by a case study. The case study is implemented with current Microsoft Azure cloud technologies. The results of the case study show that the reference architecture is valid and can be used to achieve a system for predictive maintenance execution with the cloud components of Microsoft Azure. The used concepts are technology platform agnostic and can be reused in many different cloud platforms. The reference architecture is valid and can be used in many use cases, like gas station maintenance, elevator maintenance and many more.

Keywords: case study, internet of things, predictive maintenance, reference architecture

Procedia PDF Downloads 226
43 Modeling and Experimental Verification of Crystal Growth Kinetics in Glass Forming Alloys

Authors: Peter K. Galenko, Stefanie Koch, Markus Rettenmayr, Robert Wonneberger, Evgeny V. Kharanzhevskiy, Maria Zamoryanskaya, Vladimir Ankudinov

Abstract:

We analyze the structure of undercooled melts, crystal growth kinetics and amorphous/crystalline microstructure of rapidly solidifying glass-forming Pd-based and CuZr-based alloys. A dendrite growth model is developed using a combination of the kinetic phase-field model and mesoscopic sharp interface model. The model predicts features of crystallization kinetics in alloys from thermodynamically controlled growth (governed by the Gibbs free energy change on solidification) to the kinetically limited regime (governed by atomic attachment-detachment processes at the solid/liquid interface). Comparing critical undercoolings observed in the crystallization kinetics with experimental data on melt viscosity, atomistic simulation's data on liquid microstructure and theoretically predicted dendrite growth velocity allows us to conclude that the dendrite growth kinetics strongly depends on the cluster structure changes of the melt. The obtained data of theoretical and experimental investigations are used for interpretation of microstructure of samples processed in electro-magnetic levitator on board International Space Station in the frame of the project "MULTIPHAS" (European Space Agency and German Aerospace Center, 50WM1941) and "KINETIKA" (ROSKOSMOS).

Keywords: dendrite, kinetics, model, solidification

Procedia PDF Downloads 97
42 Formulation of Hybrid Nanopowder-Molecular Ink for Fabricating Critical Material-Free Cu₂ZnSnS₄ Thin Film Solar Absorber

Authors: Anies Mutiari, Neha Bansal, Martin Artner, Veronika Mayer, Juergen Roth, Mathias Weil, Rachmat Adhi Wibowo

Abstract:

Cu₂ZnSnS₄ (CZTS) compound (mineral name kesterite) has attracted considerable interests for photovoltaic application owing to its optoelectrical properties. Moreover, its elemental abundance in Earth’s crust offers a comparative advantage for envisaged large-scale photovoltaic deployment without any material shortage issues. In this contribution, we present an innovative route to prepare CZTS solar absorber layer for photovoltaic application from low-cost and up-scalable process. CZTS layers were spin coated on the Molybdenum-coated glass from two inks composed of different solvents; dimethylsulfoxide (DMSO) and ultrapure water. Into each solvent; 0.57M CuCl₂, 0.39M ZnCl₂, 0.53M SnCl₂, and 1.85M Thiourea or Na₂S₂O₃, as well as pre-synthesized CZTS nanopowder, were added as sources of Cu, Zn, Sn and S in the ink. The crystallisation of ink into CZTS dense layers was carried out by firstly annealing the as-deposited CZTS layer in open air at 300°C for 1 minute, followed by sulfurisation at 560–620°C under atmospheric pressure for 120 minutes. Complementary electron microscopy, grazing incidence X-ray diffraction and Raman spectroscopy investigations suggest that both solvents can be used for preparing high quality and device relevant CZTS solar absorber layers. The sulphurisation crystallizes the as-deposited CZTS into highly polycrystalline CZTS layer with tetragonal structure demonstrated by the presence of tetrahedrally-shaped grains with the size of 1 µm. An advancement of the CZTS layer preparation was made by gradual substitution of volatile organic compound solvent of DMSO with ultrapure water. It is revealed that by using similar air annealing and sulphurisation process, dense and compact CZTS layers can also be fabricated from an ink with reduced volatile organic compound content.

Keywords: kesterite, solar ink, spin coating, photovoltaics

Procedia PDF Downloads 148
41 Polymer Patterning by Dip Pen Nanolithography

Authors: Ayse Cagil Kandemir, Derya Erdem, Markus Niederberger, Ralph Spolenak

Abstract:

Dip Pen nanolithography (DPN), which is a tip based method, serves a novel approach to produce nano and micro-scaled patterns due to its high resolution and pattern flexibility. It is introduced as a new constructive scanning probe lithography (SPL) technique. DPN delivers materials in the form of an ink by using the tip of a cantilever as pen and substrate as paper in order to form surface architectures. First studies rely on delivery of small organic molecules on gold substrate in ambient conditions. As time passes different inks such as; polymers, colloidal particles, oligonucleotides, metallic salts were examined on a variety of surfaces. Discovery of DPN also enabled patterning with multiple inks by using multiple cantilevers for the first time in SPL history. Specifically, polymer inks, which constitute a flexible matrix for various materials, can have a potential in MEMS, NEMS and drug delivery applications. In our study, it is aimed to construct polymer patterns using DPN by studying wetting behavior of polymer on semiconductor, metal and polymer surfaces. The optimum viscosity range of polymer and effect of environmental conditions such as humidity and temperature are examined. It is observed that there is an inverse relation with ink viscosity and depletion time. This study also yields the optimal writing conditions to produce consistent patterns with DPN. It is shown that written dot sizes increase with dwell time, indicating that the examined writing conditions yield repeatable patterns.

Keywords: dip pen nanolithography, polymer, surface patterning, surface science

Procedia PDF Downloads 377
40 PaSA: A Dataset for Patent Sentiment Analysis to Highlight Patent Paragraphs

Authors: Renukswamy Chikkamath, Vishvapalsinhji Ramsinh Parmar, Christoph Hewel, Markus Endres

Abstract:

Given a patent document, identifying distinct semantic annotations is an interesting research aspect. Text annotation helps the patent practitioners such as examiners and patent attorneys to quickly identify the key arguments of any invention, successively providing a timely marking of a patent text. In the process of manual patent analysis, to attain better readability, recognising the semantic information by marking paragraphs is in practice. This semantic annotation process is laborious and time-consuming. To alleviate such a problem, we proposed a dataset to train machine learning algorithms to automate the highlighting process. The contributions of this work are: i) we developed a multi-class dataset of size 150k samples by traversing USPTO patents over a decade, ii) articulated statistics and distributions of data using imperative exploratory data analysis, iii) baseline Machine Learning models are developed to utilize the dataset to address patent paragraph highlighting task, and iv) future path to extend this work using Deep Learning and domain-specific pre-trained language models to develop a tool to highlight is provided. This work assists patent practitioners in highlighting semantic information automatically and aids in creating a sustainable and efficient patent analysis using the aptitude of machine learning.

Keywords: machine learning, patents, patent sentiment analysis, patent information retrieval

Procedia PDF Downloads 67
39 External Noise Distillation in Quantum Holography with Undetected Light

Authors: Sebastian Töpfer, Jorge Fuenzalida, Marta Gilaberte Basset, Juan P. Torres, Markus Gräfe

Abstract:

This work presents an experimental and theoretical study about the noise resilience of quantum holography with undetected photons. Quantum imaging has become an important research topic in the recent years after its first publication in 2014. Following this research, advances towards different spectral ranges in detection and different optical geometries have been made. Especially an interest in the field of near infrared to mid infrared measurements has developed, because of the unique characteristic, that allows to sample a probe with photons in a different wavelength than the photons arriving at the detector. This promising effect can be used for medical applications, to measure in the so-called molecule fingerprint region, while using broadly available detectors for the visible spectral range. Further advance the development of quantum imaging methods have been made by new measurement and detection schemes. One of which is quantum holography with undetected light. It combines digital phase shifting holography with quantum imaging to extent the obtainable sample information, by measuring not only the object transmission, but also its influence on the phase shift experienced by the transmitted light. This work will present extended research for the quantum holography with undetected light scheme regarding the influence of external noise. It is shown experimentally and theoretically that the samples information can still be at noise levels of 250 times higher than the signal level, because of its information being transmitted by the interferometric pattern. A detailed theoretic explanation is also provided.

Keywords: distillation, quantum holography, quantum imaging, quantum metrology

Procedia PDF Downloads 46
38 Ethnomedicinal Assets of Plants Collected from Nasarawa State, North Central Nigeria

Authors: Enock E. Goler, Emmanuel H. Kwon-Ndung, Gbenga F. Akomolafe, Terna T. Paul, Markus Musa, Joshua I. Waya, James H. Okogbaa

Abstract:

An ethno-medicinal survey of plants used in treating various diseases and ailments was carried out in the study area of Nasarawa State, North Central Nigeria to obtain information on their uses and potentials. The ethno-medicinal survey was administered through structured questionnaires among local inhabitants from areas with high plant density and diversity within the various Local Government Areas of the State. A total of 84 (Eighty four) plant species belonging to 45 (Forty five) families were found to be useful in treatment of various ailments such as diabetes, measles, fever, asthma, jaundice, pneumonia, sexually transmitted diseases (STDs), aches, diarrhea, cough, arthritis, yellow fever, typhoid, erectile dysfunction and excessive bleeding. Different parts of the plant such as the roots, leaves and stems are used in preparing herbal remedies which could be from dry or freshly collected plants. The main methods of preparation are decoction or infusion, while in some cases the plant parts used are consumed directly. Residents in the study areas find the herbal remedy cheaper and more accessible and claimed that there are no side effects compared to orthodox medicine. This study has confirmed the need towards the conscious conservation of plant genetic resources in order to ensure sustained access to these ethno-medicinal plant materials.

Keywords: ethno-medicinal, Nasarawa, plants, survey

Procedia PDF Downloads 259
37 Agile Smartphone Porting and App Integration of Signal Processing Algorithms Obtained through Rapid Development

Authors: Marvin Chibuzo Offiah, Susanne Rosenthal, Markus Borschbach

Abstract:

Certain research projects in Computer Science often involve research on existing signal processing algorithms and developing improvements on them. Research budgets are usually limited, hence there is limited time for implementing the algorithms from scratch. It is therefore common practice, to use implementations provided by other researchers as a template. These are most commonly provided in a rapid development, i.e. 4th generation, programming language, usually Matlab. Rapid development is a common method in Computer Science research for quickly implementing and testing new developed algorithms, which is also a common task within agile project organization. The growing relevance of mobile devices in the computer market also gives rise to the need to demonstrate the successful executability and performance measurement of these algorithms on a mobile device operating system and processor, particularly on a smartphone. Open mobile systems such as Android, are most suitable for this task, which is to be performed most efficiently. Furthermore, efficiently implementing an interaction between the algorithm and a graphical user interface (GUI) that runs exclusively on the mobile device is necessary in cases where the project’s goal statement also includes such a task. This paper examines different proposed solutions for porting computer algorithms obtained through rapid development into a GUI-based smartphone Android app and evaluates their feasibilities. Accordingly, the feasible methods are tested and a short success report is given for each tested method.

Keywords: SMARTNAVI, Smartphone, App, Programming languages, Rapid Development, MATLAB, Octave, C/C++, Java, Android, NDK, SDK, Linux, Ubuntu, Emulation, GUI

Procedia PDF Downloads 463
36 Development of a Telemedical Network Supporting an Automated Flow Cytometric Analysis for the Clinical Follow-up of Leukaemia

Authors: Claude Takenga, Rolf-Dietrich Berndt, Erling Si, Markus Diem, Guohui Qiao, Melanie Gau, Michael Brandstoetter, Martin Kampel, Michael Dworzak

Abstract:

In patients with acute lymphoblastic leukaemia (ALL), treatment response is increasingly evaluated with minimal residual disease (MRD) analyses. Flow Cytometry (FCM) is a fast and sensitive method to detect MRD. However, the interpretation of these multi-parametric data requires intensive operator training and experience. This paper presents a pipeline-software, as a ready-to-use FCM-based MRD-assessment tool for the daily clinical practice for patients with ALL. The new tool increases accuracy in assessment of FCM-MRD in samples which are difficult to analyse by conventional operator-based gating since computer-aided analysis potentially has a superior resolution due to utilization of the whole multi-parametric FCM-data space at once instead of step-wise, two-dimensional plot-based visualization. The system developed as a telemedical network reduces the work-load and lab-costs, staff-time needed for training, continuous quality control, operator-based data interpretation. It allows dissemination of automated FCM-MRD analysis to medical centres which have no established expertise for the benefit of an even larger community of diseased children worldwide. We established a telemedical network system for analysis and clinical follow-up and treatment monitoring of Leukaemia. The system is scalable and adapted to link several centres and laboratories worldwide.

Keywords: data security, flow cytometry, leukaemia, telematics platform, telemedicine

Procedia PDF Downloads 961
35 A Comparative Time-Series Analysis and Deep Learning Projection of Innate Radon Gas Risk in Canadian and Swedish Residential Buildings

Authors: Selim M. Khan, Dustin D. Pearson, Tryggve Rönnqvist, Markus E. Nielsen, Joshua M. Taron, Aaron A. Goodarzi

Abstract:

Accumulation of radioactive radon gas in indoor air poses a serious risk to human health by increasing the lifetime risk of lung cancer and is classified by IARC as a category one carcinogen. Radon exposure risks are a function of geologic, geographic, design, and human behavioural variables and can change over time. Using time series and deep machine learning modelling, we analyzed long-term radon test outcomes as a function of building metrics from 25,489 Canadian and 38,596 Swedish residential properties constructed between 1945 to 2020. While Canadian and Swedish properties built between 1970 and 1980 are comparable (96–103 Bq/m³), innate radon risks subsequently diverge, rising in Canada and falling in Sweden such that 21st Century Canadian houses show 467% greater average radon (131 Bq/m³) relative to Swedish equivalents (28 Bq/m³). These trends are consistent across housing types and regions within each country. The introduction of energy efficiency measures within Canadian and Swedish building codes coincided with opposing radon level trajectories in each nation. Deep machine learning modelling predicts that, without intervention, average Canadian residential radon levels will increase to 176 Bq/m³ by 2050, emphasizing the importance and urgency of future building code intervention to achieve systemic radon reduction in Canada.

Keywords: radon health risk, time-series, deep machine learning, lung cancer, Canada, Sweden

Procedia PDF Downloads 64
34 Enhanced Acquisition Time of a Quantum Holography Scheme within a Nonlinear Interferometer

Authors: Sergio Tovar-Pérez, Sebastian Töpfer, Markus Gräfe

Abstract:

The work proposes a technique that decreases the detection acquisition time of quantum holography schemes down to one-third; this allows the possibility to image moving objects. Since its invention, quantum holography with undetected photon schemes has gained interest in the scientific community. This is mainly due to its ability to tailor the detected wavelengths according to the needs of the scheme implementation. Yet this wavelength flexibility grants the scheme a wide range of possible applications; an important matter was yet to be addressed. Since the scheme uses digital phase-shifting techniques to retrieve the information of the object out of the interference pattern, it is necessary to acquire a set of at least four images of the interference pattern along with well-defined phase steps to recover the full object information. Hence, the imaging method requires larger acquisition times to produce well-resolved images. As a consequence, the measurement of moving objects remains out of the reach of the imaging scheme. This work presents the use and implementation of a spatial light modulator along with a digital holographic technique called quasi-parallel phase-shifting. This technique uses the spatial light modulator to build a structured phase image consisting of a chessboard pattern containing the different phase steps for digitally calculating the object information. Depending on the reduction in the number of needed frames, the acquisition time reduces by a significant factor. This technique opens the door to the implementation of the scheme for moving objects. In particular, the application of this scheme in imaging alive specimens comes one step closer.

Keywords: quasi-parallel phase shifting, quantum imaging, quantum holography, quantum metrology

Procedia PDF Downloads 91
33 Comparison of FNTD and OSLD Detectors' Responses to Light Ion Beams Using Monte Carlo Simulations and Exprimental Data

Authors: M. R. Akbari, H. Yousefnia, A. Ghasemi

Abstract:

Al2O3:C,Mg fluorescent nuclear track detector (FNTD) and Al2O3:C optically stimulated luminescence detector (OSLD) are becoming two of the applied detectors in ion dosimetry. Therefore, the response of these detectors to hadron beams is highly of interest in radiation therapy (RT) using ion beams. In this study, these detectors' responses to proton and Helium-4 ion beams were compared using Monte Carlo simulations. The calculated data for proton beams were compared with Markus ionization chamber (IC) measurement (in water phantom) from M.D. Anderson proton therapy center. Monte Carlo simulations were performed via the FLUKA code (version 2011.2-17). The detectors were modeled in cylindrical shape at various depths of the water phantom without shading each other for obtaining relative depth dose in the phantom. Mono-energetic parallel ion beams in different incident energies (100 MeV/n to 250 MeV/n) were collided perpendicularly on the phantom surface. For proton beams, the results showed that the simulated detectors have over response relative to IC measurements in water phantom. In all cases, there were good agreements between simulated ion ranges in the water with calculated and experimental results reported by the literature. For proton, maximum peak to entrance dose ratio in the simulated water phantom was 4.3 compared with about 3 obtained from IC measurements. For He-4 ion beams, maximum peak to entrance ratio calculated by both detectors was less than 3.6 in all energies. Generally, it can be said that FLUKA is a good tool to calculate Al2O3:C,Mg FNTD and Al2O3:C OSLD detectors responses to therapeutic proton and He-4 ion beams. It can also calculate proton and He-4 ion ranges with a reasonable accuracy.

Keywords: comparison, FNTD and OSLD detectors response, light ion beams, Monte Carlo simulations

Procedia PDF Downloads 321
32 Human Leukocyte Antigen Class 1 Phenotype Distribution and Analysis in Persons from Central Uganda with Active Tuberculosis and Latent Mycobacterium tuberculosis Infection

Authors: Helen K. Buteme, Rebecca Axelsson-Robertson, Moses L. Joloba, Henry W. Boom, Gunilla Kallenius, Markus Maeurer

Abstract:

Background: The Ugandan population is heavily affected by infectious diseases and Human leukocyte antigen (HLA) diversity plays a crucial role in the host-pathogen interaction and affects the rates of disease acquisition and outcome. The identification of HLA class 1 alleles and determining which alleles are associated with tuberculosis (TB) outcomes would help in screening individuals in TB endemic areas for susceptibility to TB and to predict resistance or progression to TB which would inevitably lead to better clinical management of TB. Aims: To be able to determine the HLA class 1 phenotype distribution in a Ugandan TB cohort and to establish the relationship between these phenotypes and active and latent TB. Methods: Blood samples were drawn from 32 HIV negative individuals with active TB and 45 HIV negative individuals with latent MTB infection. DNA was extracted from the blood samples and the DNA samples HLA typed by the polymerase chain reaction-sequence specific primer method. The allelic frequencies were determined by direct count. Results: HLA-A*02, A*01, A*74, A*30, B*15, B*58, C*07, C*03 and C*04 were the dominant phenotypes in this Ugandan cohort. There were differences in the distribution of HLA types between the individuals with active TB and the individuals with LTBI with only HLA-A*03 allele showing a statistically significant difference (p=0.0136). However, after FDR computation the corresponding q-value is above the expected proportion of false discoveries (q-value 0.2176). Key findings: We identified a number of HLA class I alleles in a population from Central Uganda which will enable us to carry out a functional characterization of CD8+ T-cell mediated immune responses to MTB. Our results also suggest that there may be a positive association between the HLA-A*03 allele and TB implying that individuals with the HLA-A*03 allele are at a higher risk of developing active TB.

Keywords: HLA, phenotype, tuberculosis, Uganda

Procedia PDF Downloads 384
31 Numerical Investigation of Multiphase Flow in Pipelines

Authors: Gozel Judakova, Markus Bause

Abstract:

We present and analyze reliable numerical techniques for simulating complex flow and transport phenomena related to natural gas transportation in pipelines. Such kind of problems are of high interest in the field of petroleum and environmental engineering. Modeling and understanding natural gas flow and transformation processes during transportation is important for the sake of physical realism and the design and operation of pipeline systems. In our approach a two fluid flow model based on a system of coupled hyperbolic conservation laws is considered for describing natural gas flow undergoing hydratization. The accurate numerical approximation of two-phase gas flow remains subject of strong interest in the scientific community. Such hyperbolic problems are characterized by solutions with steep gradients or discontinuities, and their approximation by standard finite element techniques typically gives rise to spurious oscillations and numerical artefacts. Recently, stabilized and discontinuous Galerkin finite element techniques have attracted researchers’ interest. They are highly adapted to the hyperbolic nature of our two-phase flow model. In the presentation a streamline upwind Petrov-Galerkin approach and a discontinuous Galerkin finite element method for the numerical approximation of our flow model of two coupled systems of Euler equations are presented. Then the efficiency and reliability of stabilized continuous and discontinous finite element methods for the approximation is carefully analyzed and the potential of the either classes of numerical schemes is investigated. In particular, standard benchmark problems of two-phase flow like the shock tube problem are used for the comparative numerical study.

Keywords: discontinuous Galerkin method, Euler system, inviscid two-fluid model, streamline upwind Petrov-Galerkin method, twophase flow

Procedia PDF Downloads 304
30 Investigations on the Influence of Web Openings on the Load Bearing Behavior of Steel Beams

Authors: Felix Eyben, Simon Schaffrath, Markus Feldmann

Abstract:

A building should maximize the potential for use through its design. Therefore, flexible use is always important when designing a steel structure. To create flexibility, steel beams with web openings are increasingly used, because these offer the advantage that cables, pipes and other technical equipment can easily be routed through without detours, allowing for more space-saving and aesthetically pleasing construction. This can also significantly reduce the height of ceiling systems. Until now, beams with web openings were not explicitly considered in the European standard. However, this is to be done with the new EN 1993-1-13, in which design rules for different opening forms are defined. In order to further develop the design concepts, beams with web openings under bending are therefore to be investigated in terms of damage mechanics as part of a German national research project aiming to optimize the verifications for steel structures based on a wider database and a validated damage prediction. For this purpose, first, fundamental factors influencing the load-bearing behavior of girders with web openings under bending load were investigated numerically without taking material damage into account. Various parameter studies were carried out for this purpose. For example, the factors under study were the opening shape, size and position as well as structural aspects as the span length, arrangement of stiffeners and loading situation. The load-bearing behavior is evaluated using resulting load-deformation curves. These results are compared with the design rules and critically analyzed. Experimental tests are also planned based on these results. Moreover, the implementation of damage mechanics in the form of the modified Bai-Wierzbicki model was examined. After the experimental tests will have been carried out, the numerical models are validated and further influencing factors will be investigated on the basis of parametric studies.

Keywords: damage mechanics, finite element, steel structures, web openings

Procedia PDF Downloads 146
29 Wear Progress and -Mechanisms in Torpedo Ladles in Steel Industry

Authors: Mattahias Maj, Fabio Tatzgern, Karl Adam, Damir Kahrimanovic, Markus Varga

Abstract:

Torpedo ladles are necessary transport carriages in steel production to move the molten crude iron from the blast furnace to the steel refining plant. This requires the ladles to be high temperature resistant and insulate well to preserve the temperature and hold the risk of solidification at bay. Therefore, the involved refractories lining the inside of the torpedo ladles are chosen mostly according to their thermal properties, although wear of the materials by the liquid iron is also of major importance. In this work, we combined investigations of the thermal behaviour with wear studies of the lining over the whole lifetime of a torpedo ladle. Additional numerical simulations enabled a detailed model of the mechanical loads and temperature propagation at the various stations (heating, filling, emptying, cooling). The core of the investigation were detailed 3D measurements of the ladle’s cavity and thereby quantitative information of the wear progress at different time intervals during the lifetime of the ladles. The measurements allowed for a separation of different wear zones according to severity, namely the “splash zone” where the melt directly hits the ladle, the “melt zone” where during transport always liquid melt is present, and the “slag zone”, where the slag floats on the melt causing the most severe wear loss. Numerical simulations of the filling process were taken to calculate stress levels and temperature gradients, which led to the different onset of wear on those zones. Thermal imaging and punctual temperature measurements allowed for a study of the thermal consequences entailed by the wear onset. Additional “classical” damage analysis of the worn refractories complete the investigation. Thereby the wear mechanisms leading to the substantial wear loss were disclosed.

Keywords: high temperature, tribology, liquid-solid interaction, refractories, thermography

Procedia PDF Downloads 200
28 Promises versus Realities: A Critical Assessment of the Integrated Design Process

Authors: Firdous Nizar, Carmela Cucuzzella

Abstract:

This paper explores how the integrated design process (IDP) was adopted for an architectural project. The IDP is a relatively new approach to collaborative design in architectural design projects in Canada. It has gained much traction recently as the closest possible approach to the successful management of low energy building projects and has been advocated as a productive method for multi-disciplinary collaboration within complex projects. This study is based on the premise that there are explicit and implicit dimensions of power within the integrated design process (IDP) in the green building industry that may or may not lead to irreconcilable differences in a process that demands consensus. To gain insight on the potential gap between the theoretical promises and practical realities of the IDP, a review of existing IDP literature is compared with a case study analysis of a competition-based architectural project in Canada, a first to incorporate the IDP in its overall design format. This paper aims to address the undertheorized power relations of the IDP in a real project. It presents a critical assessment through the lens of the combined theories of deliberative democracy by Jürgen Habermas, with that of agonistic pluralism by political theorist Chantal Mouffe. These two theories are intended to more appropriately embrace the conflictual situations in collaborative environments, and shed light on the relationships of power, between engineers, city officials, architects, and designers in this conventional consensus-based model. In addition, propositions for a shift in approach that embraces conflictual differences among its participants are put forth based on concepts of critical spatial practice by Markus Meissen. As IDP is a relatively new design process, it requires much deliberation on its structure from the theoretical framework built in this paper in order to unlock its true potential.

Keywords: agonistic pluralism, critical spatial practice, deliberative democracy, integrated design process

Procedia PDF Downloads 158
27 Principles for the Realistic Determination of the in-situ Concrete Compressive Strength under Consideration of Rearrangement Effects

Authors: Rabea Sefrin, Christian Glock, Juergen Schnell

Abstract:

The preservation of existing structures is of great economic interest because it contributes to higher sustainability and resource conservation. In the case of existing buildings, in addition to repair and maintenance, modernization or reconstruction works often take place in the course of adjustments or changes in use. Since the structural framework and the associated load level are usually changed in the course of the structural measures, the stability of the structure must be verified in accordance with the currently valid regulations. The concrete compressive strength of the existing structures concrete and the derived mechanical parameters are of central importance for the recalculation and verification. However, the compressive strength of the existing concrete is usually set comparatively low and thus underestimated. The reasons for this are too small numbers, and large scatter of material properties of the drill cores, which are used for the experimental determination of the design value of the compressive strength. Within a structural component, the load is usually transferred over the area with higher stiffness and consequently with higher compressive strength. Therefore, existing strength variations within a component only play a subordinate role due to rearrangement effects. This paper deals with the experimental and numerical determination of such rearrangement effects in order to calculate the concrete compressive strength of existing structures more realistic and economical. The influence of individual parameters such as the specimen geometry (prism or cylinder) or the coefficient of variation of the concrete compressive strength is analyzed in experimental small-part tests. The coefficients of variation commonly used in practice are adjusted by dividing the test specimens into several layers consisting of different concretes, which are monolithically connected to each other. From each combination, a sufficient number of the test specimen is produced and tested to enable evaluation on a statistical basis. Based on the experimental tests, FE simulations are carried out to validate the test results. In the frame of a subsequent parameter study, a large number of combinations is considered, which had not been investigated in the experimental tests yet. Thus, the influence of individual parameters on the size and characteristic of the rearrangement effect is determined and described more detailed. Based on the parameter study and the experimental results, a calculation model for a more realistic determination of the in situ concrete compressive strength is developed and presented. By considering rearrangement effects in concrete during recalculation, a higher number of existing structures can be maintained without structural measures. The preservation of existing structures is not only decisive from an economic, sustainable, and resource-saving point of view but also represents an added value for cultural and social aspects.

Keywords: existing structures, in-situ concrete compressive strength, rearrangement effects, recalculation

Procedia PDF Downloads 95
26 Online Versus Face-To-Face – How Do Video Consultations Change The Doctor-Patient-Interaction

Authors: Markus Feufel, Friederike Kendel, Caren Hilger, Selamawit Woldai

Abstract:

Since the corona pandemic, the use of video consultation has increased remarkably. For vulnerable groups such as oncological patients, the advantages seem obvious. But how does video consultation potentially change the doctor-patient relationship compared to face-to-face consultation? Which barriers may hinder the effective use of this consultation format in practice? We are presenting first results from a mixed-methods field study, funded by Federal Ministry of Health, which will provide the basis for a hands-on guide for both physicians and patients on how to improve the quality of video consultations. We use a quasi-experimental design to analyze qualitative and quantitative differences between face-to-face and video consultations based on video recordings of N = 64 actual counseling sessions (n = 32 for each consultation format). Data will be recorded from n = 32 gynecological and n = 32 urological cancer patients at two clinics. After the consultation, all patients will be asked to fill out a questionnaire about their consultation experience. For quantitative analyses, the counseling sessions will be systematically compared in terms of verbal and nonverbal communication patterns. Relative frequencies of eye contact and the information exchanged will be compared using 𝝌2 -tests. The validated questionnaire MAPPIN'Obsdyad will be used to assess the expression of shared decision-making parameters. In addition, semi-structured interviews will be conducted with n = 10 physicians and n = 10 patients experienced with video consultation, for which a qualitative content analysis will be conducted. We will elaborate the comprehensive methodological approach we used to compare video vs. face-to-face consultations and present first evidence on how video consultations change the doctor-patient interaction. We will also outline possible barriers of video consultations and best practices on how they may be overcome. Based on the results, we will present and discuss recommendations outlining best practices for how to prepare and conduct high-quality video consultations from the perspective of both physicians and patients.

Keywords: video consultation, patient-doctor-relationship, digital applications, technical barriers

Procedia PDF Downloads 121
25 Lateral Torsional Buckling: Tests on Glued Laminated Timber Beams

Authors: Vera Wilden, Benno Hoffmeister, Markus Feldmann

Abstract:

Glued laminated timber (glulam) is a preferred choice for long span girders, e.g., for gyms or storage halls. While the material provides sufficient strength to resist the bending moments, large spans lead to increased slenderness of such members and to a higher susceptibility to stability issues, in particular to lateral torsional buckling (LTB). Rules for the determination of the ultimate LTB resistance are provided by Eurocode 5. The verifications of the resistance may be performed using the so called equivalent member method or by means of theory 2nd order calculations (direct method), considering equivalent imperfections. Both methods have significant limitations concerning their applicability; the equivalent member method is limited to rather simple cases; the direct method is missing detailed provisions regarding imperfections and requirements for numerical modeling. In this paper, the results of a test series on slender glulam beams in three- and four-point bending are presented. The tests were performed in an innovative, newly developed testing rig, allowing for a very precise definition of loading and boundary conditions. The load was introduced by a hydraulic jack, which follows the lateral deformation of the beam by means of a servo-controller, coupled with the tested member and keeping the load direction vertically. The deformation-controlled tests allowed for the identification of the ultimate limit state (governed by elastic stability) and the corresponding deformations. Prior to the tests, the structural and geometrical imperfections were determined and used later in the numerical models. After the stability tests, the nearly undamaged members were tested again in pure bending until reaching the ultimate moment resistance of the cross-section. These results, accompanied by numerical studies, were compared to resistance values obtained using both methods according to Eurocode 5.

Keywords: experimental tests, glued laminated timber, lateral torsional buckling, numerical simulation

Procedia PDF Downloads 207
24 The Effects of Cardiovascular Risk on Age-Related Cognitive Decline in Healthy Older Adults

Authors: A. Badran, M. Hollocks, H. Markus

Abstract:

Background: Common risk factors for cardiovascular disease are associated with age-related cognitive decline. There has been much interest in treating modifiable cardiovascular risk factors in the hope of reducing cognitive decline. However, there is currently no validated neuropsychological test to assess the subclinical cognitive effects of vascular risk. The Brief Memory and Executive Test (BMET) is a clinical screening tool, which was originally designed to be sensitive and specific to Vascular Cognitive Impairment (VCI), an impairment characterised by decline in frontally-mediated cognitive functions (e.g. Executive Function and Processing Speed). Objective: To cross-sectionally assess the validity of the BMET as a measure of the subclinical effects of vascular risk on cognition, in an otherwise healthy elderly cohort. Methods: Data from 346 participants (57 ± 10 years) without major neurological or psychiatric disorders were included in this study, gathered as part of a previous multicentre validation study for the BMET. Framingham Vascular Age was used as a surrogate measure of vascular risk, incorporating several established risk factors. Principal Components Analysis of the subtests was used to produce common constructs: an index for Memory and another for Executive Function/Processing Speed. Univariate General Linear models were used to relate Vascular Age to performance on Executive Function/Processing Speed and Memory subtests of the BMET, adjusting for Age, Premorbid Intelligence and Ethnicity. Results: Adverse vascular risk was associated with poorer performance on both the Memory and Executive Function/Processing Speed indices, adjusted for Age, Premorbid Intelligence and Ethnicity (p=0.011 and p<0.001, respectively). Conclusions: Performance on the BMET reflects the subclinical effects of vascular risk on cognition, in age-related cognitive decline. Vascular risk is associated with decline in both Executive Function/Processing Speed and Memory groups of subtests. Future studies are needed to explore whether treating vascular risk factors can effectively reduce age-related cognitive decline.

Keywords: age-related cognitive decline, vascular cognitive impairment, subclinical cerebrovascular disease, cognitive aging

Procedia PDF Downloads 447
23 Impacting the Processes of Freight Logistics at Upper Austrian Companies by the Use of Mobility Management

Authors: Theresa Steiner, Markus Pajones, Christian Haider

Abstract:

Traffic is being induced by companies due to their economic behavior. Basically, two different types of traffic occur at company sites: freight traffic and commuting traffic. Due to the fact that these traffic types are connected to each other in different kinds, an integrated approach to manage them is useful. Mobility management is a proved method for companies, to handle the traffic processes caused by their business activities. According to recent trend analysis in Austria, the freight traffic as well as the individual traffic, as part of the commuting traffic, will continue to increase. More traffic jams, as well as negative environmental impacts, are expected impacts for the future. Mobility management is a tool to control the traffic behavior with the scope to reduce emissions and other negative effects which are caused by traffic. Until now, mobility management is mainly used for optimizing commuting traffic without taking the freight logistics processes into consideration. However, the method of mobility management can be used to improve the freight traffic area of a company as well. The focus of this paper will be particularly laid on analyzing to what extent companies are already using mobility management to influence not only the commuting traffic they produce but also their processes of freight logistics. A further objective is to acquire knowledge about the motivating factors which persuade companies to introduce and apply mobility management. Additionally, advantages and disadvantages of this tool will be defined as well as limitations and factors of success, with a special focus on freight logistics, will be depicted. The first step of this paper is to conduct a literature review on the issue of mobility management with a special focus on freight logistics processes. To compare the theoretical findings with the practice, interviews, following a structured interview guidline, with mobility managers of different companies in Upper Austria will be undertaken. A qualitative analysis of these surveys will in a first step show the motivation behind using mobility management to improve traffic processes and how far this approach is already being used to especially influence the freight traffic of the companies. An evaluation to what extent the method of mobility management is already being approached at Upper Austrian companies to regulate freight logistics processes will be one outcome of this publication. Furthermore, the results of the theoretical and practical analysis will reveal not only the possibilities but also the limitations of using mobility management to influence the processes of freight logistics.

Keywords: freight logistics processes, freight traffic, mobility management, passenger traffic

Procedia PDF Downloads 294
22 Systematic Analysis of Logistics Location Search Methods under Aspects of Sustainability

Authors: Markus Pajones, Theresa Steiner, Matthias Neubauer

Abstract:

Selecting a logistics location is vital for logistics providers, food retailing and other trading companies since the selection poses an essential factor for economic success. Therefore various location search methods like cost-benefit analysis and others are well known and under usage. The development of a logistics location can be related to considerable negative effects for the eco system such as sealing the surface, wrecking of biodiversity or CO2 and noise emissions generated by freight and commuting traffic. The increasing importance of sustainability demands for taking an informed decision when selecting a logistics location for the future. Sustainability considers economic, ecologic and social aspects which should be equally integrated in the process of location search. Objectives of this paper are to define various methods which support the selection of sustainable logistics locations and to generate knowledge about the suitability, assets and limitations of the methods within the selection process. This paper investigates the role of economical, ecological and social aspects when searching for new logistics locations. Thereby, related work targeted towards location search is analyzed with respect to encoded sustainability aspects. In addition, this research aims to gain knowledge on how to include aspects of sustainability and take an informed decision when searching for a logistics location. As a result, a decomposition of the various location search methods in there components leads to a comparative analysis in form of a matrix. The comparison within a matrix enables a transparent overview about the mentioned assets and limitations of the methods and their suitability for selecting sustainable logistics locations. A further result is to generate knowledge on how to combine the separate methods to a new method for a more efficient selection of logistics locations in the context of sustainability. Future work will especially investigate the above mentioned combination of various location search methods. The objective is to develop an innovative instrument, which supports the search for logistics locations with a focus on a balanced sustainability (economy, ecology, social). Because of an ideal selection of logistics locations, induced traffic should be reduced and a mode shift to rail and public transport should be facilitated.

Keywords: commuting traffic, freight traffic, logistics location search, location search method

Procedia PDF Downloads 299
21 Flexible Integration of Airbag Weakening Lines in Interior Components: Airbag Weakening with Jenoptik Laser Technology

Authors: Markus Remm, Sebastian Dienert

Abstract:

Vehicle interiors are not only changing in terms of design and functionality but also due to new driving situations in which, for example, autonomous operating modes are possible. Flexible seating positions are changing the requirements for passive safety system behavior and location in the interior of a vehicle. With fully autonomous driving, the driver can, for example, leave the position behind the steering wheel and take a seated position facing backward. Since autonomous and non-autonomous vehicles will share the same road network for the foreseeable future, accidents cannot be avoided, which makes the use of passive safety systems indispensable. With JENOPTIK-VOTAN® A technology, the trend towards flexible predetermined airbag weakening lines is enabled. With the help of laser beams, the predetermined weakening lines are introduced from the backside of the components so that they are absolutely invisible. This machining process is sensor-controlled and guarantees that a small residual wall thickness remains for the best quality and reliability for airbag weakening lines. Due to the wide processing range of the laser, the processing of almost all materials is possible. A CO₂ laser is used for many plastics, natural fiber materials, foams, foils and material composites. A femtosecond laser is used for natural materials and textiles that are very heat-sensitive. This laser type has extremely short laser pulses with very high energy densities. Supported by a high-precision and fast movement of the laser beam by a laser scanner system, the so-called cold ablation is enabled to predetermine weakening lines layer by layer until the desired residual wall thickness remains. In that way, for example, genuine leather can be processed in a material-friendly and process-reliable manner without design implications to the components A-Side. Passive safety in the vehicle is increased through the interaction of modern airbag technology and high-precision laser airbag weakening. The JENOPTIK-VOTAN® A product family has been representing this for more than 25 years and is pointing the way to the future with new and innovative technologies.

Keywords: design freedom, interior material processing, laser technology, passive safety

Procedia PDF Downloads 96
20 Hard and Soft Skills in Marketing Education: Using Serious Games to Engage Higher Order Processing

Authors: Ann Devitt, Mairead Brady, Markus Lamest, Stephen Gomez

Abstract:

This study set out to explore the use of an online collaborative serious game for student learning in a postgraduate introductory marketing module. The simulation game aimed to bridge the theory-practice divide in marketing by allowing students to apply theory in a safe, simulated marketplace. This study addresses the following research questions: Does an online marketing simulation game engage students higher order cognitive skills? Does collaborative activity required develop students’ “soft” skills, such as communication and negotiation? What specific affordances of the online simulation promote learning? This qualitative case study took place in 2014 with 40 postgraduate students on a Business Masters Programme. The two-week intensive module combined lectures with collaborative activity on a marketing simulation game, MMX from Pearsons. The game requires student teams to compete against other teams in a marketplace and design a marketing plan to maximize key performance indicators. The data for this study comprise essays written by students after the module reflecting on their learning on the module. A thematic analysis was conducted of the essays using the following a priori theme sets: 6 levels of the cognitive domain of Blooms taxonomy; 5 principles of Cooperative Learning; affordances of simulation environments including experiential learning; motivation and engagement; goal orientation. Preliminary findings would strongly suggest that the game facilitated students identifying the value of theory in practice, in particular for future employment; enhanced their understanding of group dynamics and their role within that; and impacted very strongly, both positively and negatively on motivation. In particular the game mechanics of MMX, which hinges on the correct identification of a target consumer group, was identified as a key determinant of extrinsic and intrinsic motivation for learners. The findings also suggest that the situation of the simulation game within a broader module which required post-game reflection was valuable in identifying key learning of marketing concepts in both the positive and the negative experiences of the game.

Keywords: simulation, marketing, serious game, cooperative learning, bloom's taxonomy

Procedia PDF Downloads 534
19 Fast Estimation of Fractional Process Parameters in Rough Financial Models Using Artificial Intelligence

Authors: Dávid Kovács, Bálint Csanády, Dániel Boros, Iván Ivkovic, Lóránt Nagy, Dalma Tóth-Lakits, László Márkus, András Lukács

Abstract:

The modeling practice of financial instruments has seen significant change over the last decade due to the recognition of time-dependent and stochastically changing correlations among the market prices or the prices and market characteristics. To represent this phenomenon, the Stochastic Correlation Process (SCP) has come to the fore in the joint modeling of prices, offering a more nuanced description of their interdependence. This approach has allowed for the attainment of realistic tail dependencies, highlighting that prices tend to synchronize more during intense or volatile trading periods, resulting in stronger correlations. Evidence in statistical literature suggests that, similarly to the volatility, the SCP of certain stock prices follows rough paths, which can be described using fractional differential equations. However, estimating parameters for these equations often involves complex and computation-intensive algorithms, creating a necessity for alternative solutions. In this regard, the Fractional Ornstein-Uhlenbeck (fOU) process from the family of fractional processes offers a promising path. We can effectively describe the rough SCP by utilizing certain transformations of the fOU. We employed neural networks to understand the behavior of these processes. We had to develop a fast algorithm to generate a valid and suitably large sample from the appropriate process to train the network. With an extensive training set, the neural network can estimate the process parameters accurately and efficiently. Although the initial focus was the fOU, the resulting model displayed broader applicability, thus paving the way for further investigation of other processes in the realm of financial mathematics. The utility of SCP extends beyond its immediate application. It also serves as a springboard for a deeper exploration of fractional processes and for extending existing models that use ordinary Wiener processes to fractional scenarios. In essence, deploying both SCP and fractional processes in financial models provides new, more accurate ways to depict market dynamics.

Keywords: fractional Ornstein-Uhlenbeck process, fractional stochastic processes, Heston model, neural networks, stochastic correlation, stochastic differential equations, stochastic volatility

Procedia PDF Downloads 87