Search results for: classical mechanics
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1358

Search results for: classical mechanics

308 Total Plaque Area in Chronic Renal Failure

Authors: Hernán A. Perez, Luis J. Armando, Néstor H. García

Abstract:

Background and aims Cardiovascular disease rates are very high in patients with renal failure (CRF), but the underlying mechanisms are incompletely understood. Traditional cardiovascular risk factors do not explain the increased risk, and observational studies have observed paradoxical or absent associations between classical risk factors and mortality in dialysis patients. A large randomized controlled trial, the 4D Study, the AURORA and the ALERT study found that statin therapy in CRF do not reduce cardiovascular events. These results may be the results of ‘accelerated atherosclerosis’ observed on these patients. The objective of this study was to investigate if carotid total plaque area (TPA), a measure of carotid plaque burden growth is increased at progressively lower creatinine clearance in patients with CRF. We studied a cohort of patients with CRF not on dialysis, reasoning that risk factor associations might be more easily discerned before end stage renal disease. Methods: The Blossom DMO Argentina ethics committee approved the study and informed consent from each participant was obtained. We performed a cohort study in 412 patients with Stage 1, 2 and 3 CRF. Clinical and laboratory data were obtained. TPA was determined using bilateral carotid ultrasonography. Modification of Diet in Renal Disease estimation formula was used to determine renal function. ANOVA was used when appropriate. Results: Stage 1 CRF group (n= 16, 43±2yo) had a blood pressure of 123±2/78±2 mmHg, BMI 30±1, LDL col 145±10 mg/dl, HbA1c 5.8±0.4% and had the lowest TPA 25.8±6.9 mm2. Stage 2 CRF (n=231, 50±1 yo) had a blood pressure of 132±1/81±1 mmHg, LDL col 125±2 mg/dl, HbA1c 6±0.1% and TPA 48±10mm2 ( p< 0.05 vs CRF stage 1) while Stage 3 CRF (n=165, 59±1 yo) had a blood pressure of 134±1/81±1, LDL col 125±3 mg/dl, HbA1c 6±0.1% and TPA 71±6mm2 (p < 0.05 vs CRF stage 1 and 2). Conclusion: Our data indicate that TPA increases along the renal function deterioration, and it is not related with the LDL cholesterol and triglycerides levels. We suggest that mechanisms other than the classics are responsible for the observed excess of cardiovascular disease in CKD patients and finally, determination of total plaque area should be used to measure effects of antiatherosclerotic therapy.

Keywords: hypertension, chronic renal failure, atherosclerosis, cholesterol

Procedia PDF Downloads 257
307 A Review on Development of Pedicle Screws and Characterization of Biomaterials for Fixation in Lumbar Spine

Authors: Shri Dubey, Jamal Ghorieshi

Abstract:

Instability of the lumbar spine is caused by various factors that include degenerative disc, herniated disc, traumatic injuries, and other disorders. Pedicle screws are widely used as a main fixation device to construct rigid linkages of vertebrae to provide a fully functional and stable spine. Various technologies and methods have been used to restore the stabilization. However, loosening of pedicle screws is the main cause of concerns for neurosurgeons. This could happen due to poor bone quality with osteoporosis as well as types of pedicle screw used. Compatibilities and stabilities of pedicle screws with bone depend on design (thread design, length, and diameter) and material. Grip length and pullout strength affect the motion and stability of the spine when it goes through different phases such as extension, flexion, and rotation. Pullout strength of augmented pedicle screws is increased in both primary and salvage procedures by 119% (p = 0.001) and 162% (p = 0.01), respectively. Self-centering pedicle screws at different trajectories (0°, 10°, 20°, and 30°) show the same pullout strength as insertion in a straight-ahead trajectory. The outer cylindrical and inner conical shape of pedicle screws show the highest pullout strength in Grades 5 and 15 foams (synthetic bone). An outer cylindrical and inner conical shape with a V-shape thread exhibit the highest pullout strength in all foam grades. The maximum observed pullout strength is at axial pullout configuration at 0°. For Grade 15 (240 kg/m³) foam, there is a decline in pull out strength. The largest decrease in pullout strength is reported for Grade 10 (160 kg/m³) foam. The maximum pullout strength of 2176 N (0.32-g/cm³ Sawbones) on all densities. Type 1 Pedicle screw shows the best fixation due to smaller conical core diameter and smaller thread pitch (Screw 2 with 2 mm; Screws 1 and 3 with 3 mm).

Keywords: polymethylmethacrylate, PMMA, classical pedicle screws, CPS, expandable poly-ether-ether-ketone shell, EPEEKS, includes translaminar facet screw, TLFS, poly-ether-ether-ketone, PEEK, transfacetopedicular screw, TFPS

Procedia PDF Downloads 141
306 A Kierkegaardian Reading of Iqbal's Poetry as a Communicative Act

Authors: Sevcan Ozturk

Abstract:

The overall aim of this paper is to present a Kierkegaardian approach to Iqbal’s use of literature as a form of communication. Despite belonging to different historical, cultural, and religious backgrounds, the philosophical approaches of Soren Kierkegaard, ‘the father of existentialism,' and Muhammad Iqbal ‘the spiritual father of Pakistan’ present certain parallels. Both Kierkegaard and Iqbal take human existence as the starting point for their reflections, emphasise the subject of becoming genuine religious personalities, and develop a notion of the self. While doing these they both adopt parallel methods, employ literary techniques and poetical forms, and use their literary works as a form of communication. The problem is that Iqbal does not provide a clear account of his method as Kierkegaard does in his works. As a result, Iqbal’s literary approach appears to be a collection of contradictions. This is mainly because despite he writes most of his works in the poetical form, he condemns all kinds of art including poetry. Moreover, while attacking on Islamic mysticism, he, at the same time, uses classical literary forms, and a number of traditional mystical, poetic symbols. This paper will argue that the contradictions found in Iqbal’s approach are actually a significant part of Iqbal’s way of communicating his reader. It is the contention of this paper that with the help of the parallels between the literary and philosophical theories of Kierkegaard and Iqbal, the application of Kierkegaard’s method to Iqbal’s use of poetry as a communicative act will make it possible to dispel the seeming ambiguities in Iqbal’s literary approach. The application of Kierkegaard’s theory to Iqbal’s literary method will include an analysis of the main principles of Kierkegaard’s own literary technique of ‘indirect communication,' which is a crucial term of his existentialist philosophy. Second, the clash between what Iqbal’s says about art and poetry and what he does will be highlighted in the light of Kierkegaardian theory of indirect communication. It will be argued that Iqbal’s literary technique can be considered as a form of ‘indirect communication,' and that reading his technique in this way helps on dispelling the contradictions in his approach. It is hoped that this paper will cultivate a dialogue between those who work in the fields of comparative philosophy Kierkegaard studies, existentialism, contemporary Islamic thought, Iqbal studies, and literary criticism.

Keywords: comparative philosophy, existentialism, indirect communication, intercultural philosophy, literary communication, Muhammad Iqbal, Soren Kierkegaard

Procedia PDF Downloads 308
305 Computation of Residual Stresses in Human Face Due to Growth

Authors: M. A. Askari, M. A. Nazari, P. Perrier, Y. Payan

Abstract:

Growth and remodeling of biological structures have gained lots of attention over the past decades. Determining the response of the living tissues to the mechanical loads is necessary for a wide range of developing fields such as, designing of prosthetics and optimized surgery operations. It is a well-known fact that biological structures are never stress-free, even when externally unloaded. The exact origin of these residual stresses is not clear, but theoretically growth and remodeling is one of the main sources. Extracting body organs from medical imaging, does not produce any information regarding the existing residual stresses in that organ. The simplest cause of such stresses is the gravity since an organ grows under its influence from its birth. Ignoring such residual stresses might cause erroneous results in numerical simulations. Accounting for residual stresses due to tissue growth can improve the accuracy of mechanical analysis results. In this paper, we have implemented a computational framework based on fixed-point iteration to determine the residual stresses due to growth. Using nonlinear continuum mechanics and the concept of fictitious configuration we find the unknown stress-free reference configuration which is necessary for mechanical analysis. To illustrate the method, we apply it to a finite element model of healthy human face whose geometry has been extracted from medical images. We have computed the distribution of residual stress in facial tissues, which can overcome the effect of gravity and cause that tissues remain firm. Tissue wrinkles caused by aging could be a consequence of decreasing residual stress and not counteracting the gravity. Considering these stresses has important application in maxillofacial surgery. It helps the surgeons to predict the changes after surgical operations and their consequences.

Keywords: growth, soft tissue, residual stress, finite element method

Procedia PDF Downloads 337
304 Comparison of Data Reduction Algorithms for Image-Based Point Cloud Derived Digital Terrain Models

Authors: M. Uysal, M. Yilmaz, I. Tiryakioğlu

Abstract:

Digital Terrain Model (DTM) is a digital numerical representation of the Earth's surface. DTMs have been applied to a diverse field of tasks, such as urban planning, military, glacier mapping, disaster management. In the expression of the Earth' surface as a mathematical model, an infinite number of point measurements are needed. Because of the impossibility of this case, the points at regular intervals are measured to characterize the Earth's surface and DTM of the Earth is generated. Hitherto, the classical measurement techniques and photogrammetry method have widespread use in the construction of DTM. At present, RADAR, LiDAR, and stereo satellite images are also used for the construction of DTM. In recent years, especially because of its superiorities, Airborne Light Detection and Ranging (LiDAR) has an increased use in DTM applications. A 3D point cloud is created with LiDAR technology by obtaining numerous point data. However recently, by the development in image mapping methods, the use of unmanned aerial vehicles (UAV) for photogrammetric data acquisition has increased DTM generation from image-based point cloud. The accuracy of the DTM depends on various factors such as data collection method, the distribution of elevation points, the point density, properties of the surface and interpolation methods. In this study, the random data reduction method is compared for DTMs generated from image based point cloud data. The original image based point cloud data set (100%) is reduced to a series of subsets by using random algorithm, representing the 75, 50, 25 and 5% of the original image based point cloud data set. Over the ANS campus of Afyon Kocatepe University as the test area, DTM constructed from the original image based point cloud data set is compared with DTMs interpolated from reduced data sets by Kriging interpolation method. The results show that the random data reduction method can be used to reduce the image based point cloud datasets to 50% density level while still maintaining the quality of DTM.

Keywords: DTM, Unmanned Aerial Vehicle (UAV), uniform, random, kriging

Procedia PDF Downloads 141
303 Mental Wellbeing Using Music Intervention: A Case Study of Therapeutic Role of Music, From Both Psychological and Neurocognitive Perspectives

Authors: Medha Basu, Kumardeb Banerjee, Dipak Ghosh

Abstract:

After the massive blow of the COVID-19 pandemic, several health hazards have been reported all over the world. Serious cases of Major Depressive Disorder (MDD) are seen to be common in about 15% of the global population, making depression one of the leading mental health diseases, as reported by the World Health Organization. Various psychological and pharmacological treatment techniques are regularly being reported. Music, a globally accepted mode of entertainment, is often used as a therapeutic measure to treat various health conditions. We have tried to understand how Indian Classical Music can affect the overall well-being of the human brain. A case study has been reported here, where a Flute-rendition has been chosen from a detailed audience response survey, and the effects of that clip on human brain conditions have been studied from both psychological and neural perspectives. Taking help from internationally-accepted depression-rating scales, two questionnaires have been designed to understand both the prolonged and immediate effect of music on various emotional states of human lives. Thereafter, from EEG experiments on 5 participants using the same clip, the parameter ‘ALAY’, alpha frontal asymmetry (alpha power difference of right and left frontal hemispheres), has been calculated. Works of Richard Davidson show that an increase in the ‘ALAY’ value indicates a decrease in depressive symptoms. Using the non-linear technique of MFDFA on EEG analysis, we have also calculated frontal asymmetry using the complexity values of alpha-waves in both hemispheres. The results show a positive correlation between both the psychological survey and the EEG findings, revealing the prominent role of music on the human brain, leading to a decrease in mental unrest and an increase in overall well-being. In this study, we plan to propose the scientific foundation of music therapy, especially from a neurocognition perspective, with appropriate neural bio-markers to understand the positive and remedial effects of music on the human brain.

Keywords: music therapy, EEG, psychological survey, frontal alpha asymmetry, wellbeing

Procedia PDF Downloads 14
302 Exploring Regularity Results in the Context of Extremely Degenerate Elliptic Equations

Authors: Zahid Ullah, Atlas Khan

Abstract:

This research endeavors to explore the regularity properties associated with a specific class of equations, namely extremely degenerate elliptic equations. These equations hold significance in understanding complex physical systems like porous media flow, with applications spanning various branches of mathematics. The focus is on unraveling and analyzing regularity results to gain insights into the smoothness of solutions for these highly degenerate equations. Elliptic equations, fundamental in expressing and understanding diverse physical phenomena through partial differential equations (PDEs), are particularly adept at modeling steady-state and equilibrium behaviors. However, within the realm of elliptic equations, the subset of extremely degenerate cases presents a level of complexity that challenges traditional analytical methods, necessitating a deeper exploration of mathematical theory. While elliptic equations are celebrated for their versatility in capturing smooth and continuous behaviors across different disciplines, the introduction of degeneracy adds a layer of intricacy. Extremely degenerate elliptic equations are characterized by coefficients approaching singular behavior, posing non-trivial challenges in establishing classical solutions. Still, the exploration of extremely degenerate cases remains uncharted territory, requiring a profound understanding of mathematical structures and their implications. The motivation behind this research lies in addressing gaps in the current understanding of regularity properties within solutions to extremely degenerate elliptic equations. The study of extreme degeneracy is prompted by its prevalence in real-world applications, where physical phenomena often exhibit characteristics defying conventional mathematical modeling. Whether examining porous media flow or highly anisotropic materials, comprehending the regularity of solutions becomes crucial. Through this research, the aim is to contribute not only to the theoretical foundations of mathematics but also to the practical applicability of mathematical models in diverse scientific fields.

Keywords: elliptic equations, extremely degenerate, regularity results, partial differential equations, mathematical modeling, porous media flow

Procedia PDF Downloads 49
301 Application of Particle Image Velocimetry in the Analysis of Scale Effects in Granular Soil

Authors: Zuhair Kadhim Jahanger, S. Joseph Antony

Abstract:

The available studies in the literature which dealt with the scale effects of strip footings on different sand packing systematically still remain scarce. In this research, the variation of ultimate bearing capacity and deformation pattern of soil beneath strip footings of different widths under plane-strain condition on the surface of loose, medium-dense and dense sand have been systematically studied using experimental and noninvasive methods for measuring microscopic deformations. The presented analyses are based on model scale compression test analysed using Particle Image Velocimetry (PIV) technique. Upper bound analysis of the current study shows that the maximum vertical displacement of the sand under the ultimate load increases for an increase in the width of footing, but at a decreasing rate with relative density of sand, whereas the relative vertical displacement in the sand decreases for an increase in the width of the footing. A well agreement is observed between experimental results for different footing widths and relative densities. The experimental analyses have shown that there exists pronounced scale effect for strip surface footing. The bearing capacity factors rapidly decrease up to footing widths B=0.25 m, 0.35 m, and 0.65 m for loose, medium-dense and dense sand respectively, after that there is no significant decrease in . The deformation modes of the soil as well as the ultimate bearing capacity values have been affected by the footing widths. The obtained results could be used to improve settlement calculation of the foundation interacting with granular soil.

Keywords: DPIV, granular mechanics, scale effect, upper bound analysis

Procedia PDF Downloads 133
300 Exploring the Applications of Neural Networks in the Adaptive Learning Environment

Authors: Baladitya Swaika, Rahul Khatry

Abstract:

Computer Adaptive Tests (CATs) is one of the most efficient ways for testing the cognitive abilities of students. CATs are based on Item Response Theory (IRT) which is based on item selection and ability estimation using statistical methods of maximum information selection/selection from posterior and maximum-likelihood (ML)/maximum a posteriori (MAP) estimators respectively. This study aims at combining both classical and Bayesian approaches to IRT to create a dataset which is then fed to a neural network which automates the process of ability estimation and then comparing it to traditional CAT models designed using IRT. This study uses python as the base coding language, pymc for statistical modelling of the IRT and scikit-learn for neural network implementations. On creation of the model and on comparison, it is found that the Neural Network based model performs 7-10% worse than the IRT model for score estimations. Although performing poorly, compared to the IRT model, the neural network model can be beneficially used in back-ends for reducing time complexity as the IRT model would have to re-calculate the ability every-time it gets a request whereas the prediction from a neural network could be done in a single step for an existing trained Regressor. This study also proposes a new kind of framework whereby the neural network model could be used to incorporate feature sets, other than the normal IRT feature set and use a neural network’s capacity of learning unknown functions to give rise to better CAT models. Categorical features like test type, etc. could be learnt and incorporated in IRT functions with the help of techniques like logistic regression and can be used to learn functions and expressed as models which may not be trivial to be expressed via equations. This kind of a framework, when implemented would be highly advantageous in psychometrics and cognitive assessments. This study gives a brief overview as to how neural networks can be used in adaptive testing, not only by reducing time-complexity but also by being able to incorporate newer and better datasets which would eventually lead to higher quality testing.

Keywords: computer adaptive tests, item response theory, machine learning, neural networks

Procedia PDF Downloads 160
299 Validity of Universe Structure Conception as Nested Vortexes

Authors: Khaled M. Nabil

Abstract:

This paper introduces the Nested Vortexes conception of the universe structure and interprets all the physical phenomena according this conception. The paper first reviews recent physics theories, either in microscopic scale or macroscopic scale, to collect evidence that the space is not empty. But, these theories describe the property of the space medium without determining its structure. Determining the structure of space medium is essential to understand the mechanism that leads to its properties. Without determining the space medium structure, many phenomena; such as electric and magnetic fields, gravity, or wave-particle duality remain uninterpreted. Thus, this paper introduces a conception about the structure of the universe. It assumes that the universe is a medium of ultra-tiny homogeneous particles which are still undiscovered. Like any medium with certain movements, possibly because of a great asymmetric explosion, vortexes have occurred. A vortex condenses the ultra-tiny particles in its center forming a bigger particle, the bigger particles, in turn, could be trapped in a bigger vortex and condense in its center forming a much bigger particle and so on. This conception describes galaxies, stars, protons as particles at different levels. Existing of the particle’s vortexes make the consistency of the speed of light postulate is not true. This conception shows that the vortex motion dynamic agrees with the motion of all the universe particles at any level. An experiment has been carried out to detect the orbiting effect of aggregated vortexes of aligned atoms of a permanent magnet. Based on the described particle’s structure, the gravity force of a particle and attraction between particles as well as charge, electric and magnetic fields and quantum mechanics characteristics are interpreted. All augmented physics phenomena are solved.

Keywords: astrophysics, cosmology, particles’ structure model, particles’ forces

Procedia PDF Downloads 106
298 Close-Range Remote Sensing Techniques for Analyzing Rock Discontinuity Properties

Authors: Sina Fatolahzadeh, Sergio A. Sepúlveda

Abstract:

This paper presents advanced developments in close-range, terrestrial remote sensing techniques to enhance the characterization of rock masses. The study integrates two state-of-the-art laser-scanning technologies, the HandySCAN and GeoSLAM laser scanners, to extract high-resolution geospatial data for rock mass analysis. These instruments offer high accuracy, precision, low acquisition time, and high efficiency in capturing intricate geological features in small to medium size outcrops and slope cuts. Using the HandySCAN and GeoSLAM laser scanners facilitates real-time, three-dimensional mapping of rock surfaces, enabling comprehensive assessments of rock mass characteristics. The collected data provide valuable insights into structural complexities, surface roughness, and discontinuity patterns, which are essential for geological and geotechnical analyses. The synergy of these advanced remote sensing technologies contributes to a more precise and straightforward understanding of rock mass behavior. In this case, the main parameters of RQD, joint spacing, persistence, aperture, roughness, infill, weathering, water condition, and joint orientation in a slope cut along the Sea-to-Sky Highway, BC, were remotely analyzed to calculate and evaluate the Rock Mass Rating (RMR) and Geological Strength Index (GSI) classification systems. Automatic and manual analyses of the acquired data are then compared with field measurements. The results show the usefulness of the proposed remote sensing methods and their appropriate conformity with the actual field data.

Keywords: remote sensing, rock mechanics, rock engineering, slope stability, discontinuity properties

Procedia PDF Downloads 45
297 Robustness of the Deep Chroma Extractor and Locally-Normalized Quarter Tone Filters in Automatic Chord Estimation under Reverberant Conditions

Authors: Luis Alvarado, Victor Poblete, Isaac Gonzalez, Yetzabeth Gonzalez

Abstract:

In MIREX 2016 (http://www.music-ir.org/mirex), the deep neural network (DNN)-Deep Chroma Extractor, proposed by Korzeniowski and Wiedmer, reached the highest score in an audio chord recognition task. In the present paper, this tool is assessed under acoustic reverberant environments and distinct source-microphone distances. The evaluation dataset comprises The Beatles and Queen datasets. These datasets are sequentially re-recorded with a single microphone in a real reverberant chamber at four reverberation times (0 -anechoic-, 1, 2, and 3 s, approximately), as well as four source-microphone distances (32, 64, 128, and 256 cm). It is expected that the performance of the trained DNN will dramatically decrease under these acoustic conditions with signals degraded by room reverberation and distance to the source. Recently, the effect of the bio-inspired Locally-Normalized Cepstral Coefficients (LNCC), has been assessed in a text independent speaker verification task using speech signals degraded by additive noise at different signal-to-noise ratios with variations of recording distance, and it has also been assessed under reverberant conditions with variations of recording distance. LNCC showed a performance so high as the state-of-the-art Mel Frequency Cepstral Coefficient filters. Based on these results, this paper proposes a variation of locally-normalized triangular filters called Locally-Normalized Quarter Tone (LNQT) filters. By using the LNQT spectrogram, robustness improvements of the trained Deep Chroma Extractor are expected, compared with classical triangular filters, and thus compensating the music signal degradation improving the accuracy of the chord recognition system.

Keywords: chord recognition, deep neural networks, feature extraction, music information retrieval

Procedia PDF Downloads 216
296 Integration of Agile Philosophy and Scrum Framework to Missile System Design Processes

Authors: Misra Ayse Adsiz, Selim Selvi

Abstract:

In today's world, technology is competing with time. In order to catch up with the world's companies and adapt quickly to the changes, it is necessary to speed up the processes and keep pace with the rate of change of the technology. The missile system design processes, which are handled with classical methods, keep behind in this race. Because customer requirements are not clear, and demands are changing again and again in the design process. Therefore, in the system design process, a methodology suitable for the missile system design dynamics has been investigated and the processes used for catching up the era are examined. When commonly used design processes are analyzed, it is seen that any one of them is dynamic enough for today’s conditions. So a hybrid design process is established. After a detailed review of the existing processes, it is decided to focus on the Scrum Framework and Agile Philosophy. Scrum is a process framework. It is focused on to develop software and handling change management with rapid methods. In addition, agile philosophy is intended to respond quickly to changes. In this study, it is aimed to integrate Scrum framework and agile philosophy, which are the most appropriate ways for rapid production and change adaptation, into the missile system design process. With this approach, it is aimed that the design team, involved in the system design processes, is in communication with the customer and provide an iterative approach in change management. These methods, which are currently being used in the software industry, have been integrated with the product design process. A team is created for system design process. The roles of Scrum Team are realized with including the customer. A scrum team consists of the product owner, development team and scrum master. Scrum events, which are short, purposeful and time-limited, are organized to serve for coordination rather than long meetings. Instead of the classic system design methods used in product development studies, a missile design is made with this blended method. With the help of this design approach, it is become easier to anticipate changing customer demands, produce quick solutions to demands and combat uncertainties in the product development process. With the feedback of the customer who included in the process, it is worked towards marketing optimization, design and financial optimization.

Keywords: agile, design, missile, scrum

Procedia PDF Downloads 150
295 Electric Field-Induced Deformation of Particle-Laden Drops and Structuring of Surface Particles

Authors: Alexander Mikkelsen, Khobaib Khobaib, Zbigniew Rozynek

Abstract:

Drops covered by particles have found important uses in various fields, ranging from stabilization of emulsions to production of new advanced materials. Particles at drop interfaces can be interlocked to form solid capsules with properties tailored for a myriad of applications. Despite the huge potential of particle-laden drops and capsules, the knowledge of their deformation and stability are limited. In this regard, we contribute with experimental studies on the deformation and manipulation of silicone oil drops covered with micrometer-sized particles subjected to electric fields. A mixture of silicone oil and particles were immersed in castor oil using a mechanical pipette, forming millimeter sized drops. The particles moved and adsorbed at the drop interfaces by sedimentation, and were structured at the interface by electric field-induced electrohydrodynamic flows. When applying a direct current electric field, free charges accumulated at the drop interfaces, yielding electric stress that deformed the drops. In our experiments, we investigated how particle properties affected drop deformation, break-up, and particle structuring. We found that by increasing the size of weakly-conductive clay particles, the drop shape can go from compressed to stretched out in the direction of the electric field. Increasing the particle size and electrical properties were also found to weaken electrohydrodynamic flows, induce break-up of drops at weaker electric field strengths and structure particles in chains. These particle parameters determine the dipolar force between the interfacial particles, which can yield particle chaining. We conclude that the balance between particle chaining and electrohydrodynamic flows governs the observed drop mechanics.

Keywords: drop deformation, electric field induced stress, electrohydrodynamic flows, particle structuring at drop interfaces

Procedia PDF Downloads 183
294 Impact of Expressive Writing on Creativity

Authors: Małgorzata Osowiecka

Abstract:

Negative emotions are rather seen as creativity inhibitor. On the other hand, it is worth noting that negative emotions may be good for our functioning. Negative emotions enhance cognitive resources and improve evaluative processes. Moreover maintaining a negative emotional state allow for cognitive reinterpretation of the emotional stimuli, what is good for our creativity, especially cognitive flexibility. Writing a diary or writing about difficult emotional experiences in general can be the way to not only improve psychical health, but also – enhance creative behaviors. Thanks to translating difficult emotions to the verbal level and giving them ‘a name’ or ‘a label’, we can get easier access to both emotional content of an experience and to the semantic content, without the need of speaking out loud. Expressive writing improves academic results and the efficiency of working memory. The classical method of writing about emotions consists in a long-term process of describing negative experiences. Present research demonstrate the efficiency of this process over a shorter period of time - one writing session, on school children sample. Participants performed writing task. Writing task had two different topics: emotions connected with their negative emotions (expressive writing) and content not connected with negative emotional state (writing about one’s typical day). Creativity was measured by Guilford’s Alternative Uses Task. Results have shown that writing about negative emotions results in the higher level of divergent thinking in all three parameters: fluency, flexibility and originality. After the writing task mood of expressive writing participants remained negative more than the mood of the controls. Taking an expressive action after a difficult emotional experience can support functioning, which can be observed in enhancement of divergent thinking. Writing about emotions connected with negative experience makes one more creative, than writing about something unrelated with difficult emotional moments. Research has shown that young people should not demonize negative emotions. Sometimes, properly applied, negative emotions can be the basis of creation. Preparation was supported by a The Young Scientist University grant titled ‘Dynamics of emotions in the creative process’ from The Polish Ministry of Science and Higher Education.

Keywords: creativity, divergent thinking, emotions, expressive writing

Procedia PDF Downloads 174
293 Discerning Divergent Nodes in Social Networks

Authors: Mehran Asadi, Afrand Agah

Abstract:

In data mining, partitioning is used as a fundamental tool for classification. With the help of partitioning, we study the structure of data, which allows us to envision decision rules, which can be applied to classification trees. In this research, we used online social network dataset and all of its attributes (e.g., Node features, labels, etc.) to determine what constitutes an above average chance of being a divergent node. We used the R statistical computing language to conduct the analyses in this report. The data were found on the UC Irvine Machine Learning Repository. This research introduces the basic concepts of classification in online social networks. In this work, we utilize overfitting and describe different approaches for evaluation and performance comparison of different classification methods. In classification, the main objective is to categorize different items and assign them into different groups based on their properties and similarities. In data mining, recursive partitioning is being utilized to probe the structure of a data set, which allow us to envision decision rules and apply them to classify data into several groups. Estimating densities is hard, especially in high dimensions, with limited data. Of course, we do not know the densities, but we could estimate them using classical techniques. First, we calculated the correlation matrix of the dataset to see if any predictors are highly correlated with one another. By calculating the correlation coefficients for the predictor variables, we see that density is strongly correlated with transitivity. We initialized a data frame to easily compare the quality of the result classification methods and utilized decision trees (with k-fold cross validation to prune the tree). The method performed on this dataset is decision trees. Decision tree is a non-parametric classification method, which uses a set of rules to predict that each observation belongs to the most commonly occurring class label of the training data. Our method aggregates many decision trees to create an optimized model that is not susceptible to overfitting. When using a decision tree, however, it is important to use cross-validation to prune the tree in order to narrow it down to the most important variables.

Keywords: online social networks, data mining, social cloud computing, interaction and collaboration

Procedia PDF Downloads 131
292 Understanding Inhibitory Mechanism of the Selective Inhibitors of Cdk5/p25 Complex by Molecular Modeling Studies

Authors: Amir Zeb, Shailima Rampogu, Minky Son, Ayoung Baek, Sang H. Yoon, Keun W. Lee

Abstract:

Neurotoxic insults activate calpain, which in turn produces truncated p25 from p35. p25 forms hyperactivated Cdk5/p25 complex, and thereby induces severe neuropathological aberrations including hyperphosphorylated tau, neuroinflammation, apoptosis, and neuronal death. Inhibition of Cdk5/p25 complex alleviates aberrant phosphorylation of tau to mitigate AD pathology. PHA-793887 and Roscovitine have been investigated as selective inhibitors of Cdk5/p25 with IC50 values 5nM and 160nM, respectively, but their mechanistic studies remain unknown. Herein, computational simulations have explored the binding mode and interaction mechanism of PHA-793887 and Roscovitine with Cdk5/p25. Docking results suggested that PHA-793887 and Rsocovitine have occupied the ATP-binding site of Cdk5 and obtained highest docking (GOLD) score of 66.54 and 84.03, respectively. Furthermore, molecular dynamics (MD) simulation demonstrated that PHA-793887 and Roscovitine established stable RMSD of 1.09 Å and 1.48 Å with Cdk5/p25, respectively. Profiling of polar interactions suggested that each inhibitor formed hydrogen bonds (H-bond) with catalytic residues of Cdk5 and could remain stable throughout the molecular dynamics simulation. Additionally, binding free energy calculation by molecular mechanics/Poisson–Boltzmann surface area (MM/PBSA) suggested that PHA-793887 and Roscovitine had lowest binding free energies of -150.05 kJ/mol and -113.14 kJ/mol, respectively with Cdk5/p25. Free energy decomposition demonstrated that polar energy by H-bond between the Glu81 of Cdk5 and PHA-793887 is the essential factor to make PHA-793887 highly selective towards Cdk5/p25. Overall, this study provided substantial evidences to explore mechanistic interactions of the selective inhibitors of Cdk5/p25 and could be used as fundamental considerations in the development of structure-based selective inhibitors of Cdk5/p25.

Keywords: Cdk5/p25 inhibition, molecular modeling of Cdk5/p25, PHA-793887 and roscovitine, selective inhibition of Cdk5/p25

Procedia PDF Downloads 123
291 Triplet Shear Tests on Retrofitted Brickwork Masonry Walls

Authors: Berna Istegun, Erkan Celebi

Abstract:

The main objective of this experimental study is to assess the shear strength and the crack behavior of the triplets built of perforated brickwork masonry elements. In order to observe the influence of shear resistance and energy dissipating before and after retrofitting applications by using the reinforcing system, static-cyclic shear tests were employed in the structural mechanics laboratory of Sakarya University. The reinforcing system is composed of hybrid multiaxial seismic fabric consisting of alkali resistant glass and polypropylene fibers. The plaster as bonding material used in the specimen’s retrofitting consists of expanded glass granular. In order to acquire exact measuring data about the failure behavior of the two mortar joints under shear stressing, vertical load-controlled cylinder having force capacity of 50 kN and loading rate of 1.5 mm/min. with an internal inductive displacement transducers is carried out perpendicular to the triplet specimens. In this study, a total of six triplet specimens with textile reinforcement were prepared for these shear bond tests. The three of them were produced as single-sided reinforced triplets with seismic fabric, while the others were strengthened on both sides. In addition, three triplet specimens without retrofitting and plaster were also tested as reference samples. The obtained test results were given in the manner of force-displacement relationships, ductility coefficients and shear strength parameters comparatively. It is concluded that two-side seismic textile applications on masonry elements with relevant plaster have considerably increased the sheer force resistance and the ductility capacity.

Keywords: expanded glass granular, perforated brickwork, retrofitting, seismic fabric, triplet shear tests

Procedia PDF Downloads 190
290 Therapeutic Drug Monitoring by Dried Blood Spot and LC-MS/MS: Novel Application to Carbamazepine and Its Metabolite in Paediatric Population

Authors: Giancarlo La Marca, Engy Shokry, Fabio Villanelli

Abstract:

Epilepsy is one of the most common neurological disorders, with an estimated prevalence of 50 million people worldwide. Twenty five percent of the epilepsy population is represented in children under the age of 15 years. For antiepileptic drugs (AED), there is a poor correlation between plasma concentration and dose especially in children. This was attributed to greater pharmacokinetic variability than adults. Hence, therapeutic drug monitoring (TDM) is recommended in controlling toxicity while drug exposure is maintained. Carbamazepine (CBZ) is a first-line AED and the drug of first choice in trigeminal neuralgia. CBZ is metabolised in the liver into carbamazepine-10,11-epoxide (CBZE), its major metabolite which is equipotent. This develops the need for an assay able to monitor the levels of both CBZ and CBZE. The aim of the present study was to develop and validate a LC-MS/MS method for simultaneous quantification of CBZ and CBZE in dried blood spots (DBS). DBS technique overcomes many logistical problems, ethical issues and technical challenges faced by classical plasma sampling. LC-MS/MS has been regarded as superior technique over immunoassays and HPLC/UV methods owing to its better specificity and sensitivity, lack of interference or matrix effects. Our method combines advantages of DBS technique and LC-MS/MS in clinical practice. The extraction process was done using methanol-water-formic acid (80:20:0.1, v/v/v). The chromatographic elution was achieved by using a linear gradient with a mobile phase consisting of acetonitrile-water-0.1% formic acid at a flow rate of 0.50 mL/min. The method was linear over the range 1-40 mg/L and 0.25-20 mg/L for CBZ and CBZE respectively. The limit of quantification was 1.00 mg/L and 0.25 mg/L for CBZ and CBZE, respectively. Intra-day and inter-day assay precisions were found to be less than 6.5% and 11.8%. An evaluation of DBS technique was performed, including effect of extraction solvent, spot homogeneity and stability in DBS. Results from a comparison with the plasma assay are also presented. The novelty of the present work lies in being the first to quantify CBZ and its metabolite from only one 3.2 mm DBS disc finger-prick sample (3.3-3.4 µl blood) by LC-MS/MS in a 10 min. chromatographic run.

Keywords: carbamazepine, carbamazepine-10, 11-epoxide, dried blood spots, LC-MS/MS, therapeutic drug monitoring

Procedia PDF Downloads 393
289 A Comprehensive Survey on Machine Learning Techniques and User Authentication Approaches for Credit Card Fraud Detection

Authors: Niloofar Yousefi, Marie Alaghband, Ivan Garibay

Abstract:

With the increase of credit card usage, the volume of credit card misuse also has significantly increased, which may cause appreciable financial losses for both credit card holders and financial organizations issuing credit cards. As a result, financial organizations are working hard on developing and deploying credit card fraud detection methods, in order to adapt to ever-evolving, increasingly sophisticated defrauding strategies and identifying illicit transactions as quickly as possible to protect themselves and their customers. Compounding on the complex nature of such adverse strategies, credit card fraudulent activities are rare events compared to the number of legitimate transactions. Hence, the challenge to develop fraud detection that are accurate and efficient is substantially intensified and, as a consequence, credit card fraud detection has lately become a very active area of research. In this work, we provide a survey of current techniques most relevant to the problem of credit card fraud detection. We carry out our survey in two main parts. In the first part, we focus on studies utilizing classical machine learning models, which mostly employ traditional transnational features to make fraud predictions. These models typically rely on some static physical characteristics, such as what the user knows (knowledge-based method), or what he/she has access to (object-based method). In the second part of our survey, we review more advanced techniques of user authentication, which use behavioral biometrics to identify an individual based on his/her unique behavior while he/she is interacting with his/her electronic devices. These approaches rely on how people behave (instead of what they do), which cannot be easily forged. By providing an overview of current approaches and the results reported in the literature, this survey aims to drive the future research agenda for the community in order to develop more accurate, reliable and scalable models of credit card fraud detection.

Keywords: Credit Card Fraud Detection, User Authentication, Behavioral Biometrics, Machine Learning, Literature Survey

Procedia PDF Downloads 94
288 How Social Support, Interaction with Clients and Work-Family Conflict Contribute to Mental Well-Being for Employees in the Human Service System

Authors: Uwe C. Fischer

Abstract:

Mental health and well-being for employees working in the human service system are getting more and more important given the increasing rate of absenteeism at work. Besides individual capacities, social and community factors seem to be important in the working setting. Starting from a demand resource framework including the classical demand control aspects, social support systems, specific demands and resources of the client work, and work-family conflict were considered in the present study. We state hypothetically, that these factors have a meaningful association with the mental quality of life of employees working in the field of social, educational and health sectors. 1140 employees, working in human service organizations (education, youth care, nursing etc.) were asked for strains and resources at work (selected scales from Salutogenetic Subjective Work Assessment SALSA and own new scales for client work), work-family conflict, and mental quality of life from the German Short Form Health Survey. Considering the complex influences of the variables, we conducted a multiple hierarchical regression analysis. One third of the whole variance of the mental quality of life can be declared by the different variables of the model. When the variables concerning social influences were included in the hierarchical regression, the influence of work related control resource decreased. Excessive workload, work-family conflict, social support by supervisors, co-workers and other persons outside work, as well as strains and resources associated with client work had significant regression coefficients. Conclusions: Social support systems are crucial in the social, educational and health related service sector, regarding the influence on mental well-being. Especially the work-family conflict focuses on the importance of the work-life balance. Also the specific strains and resources of the client work, measured with new constructed scales, showed great impact on mental health. Therefore occupational health promotion should focus more on the social factors within and outside the working place.

Keywords: client interaction, human service system, mental health, social support, work-family conflict

Procedia PDF Downloads 423
287 Mechanical Behavior of Laminated Glass Cylindrical Shell with Hinged Free Boundary Conditions

Authors: Ebru Dural, M. Zulfu Asık

Abstract:

Laminated glass is a kind of safety glass, which is made by 'sandwiching' two glass sheets and a polyvinyl butyral (PVB) interlayer in between them. When the glass is broken, the interlayer in between the glass sheets can stick them together. Because of this property, the hazards of sharp projectiles during natural and man-made disasters reduces. They can be widely applied in building, architecture, automotive, transport industries. Laminated glass can easily undergo large displacements even under their own weight. In order to explain their true behavior, they should be analyzed by using large deflection theory to represent nonlinear behavior. In this study, a nonlinear mathematical model is developed for the analysis of laminated glass cylindrical shell which is free in radial directions and restrained in axial directions. The results will be verified by using the results of the experiment, carried out on laminated glass cylindrical shells. The behavior of laminated composite cylindrical shell can be represented by five partial differential equations. Four of the five equations are used to represent axial displacements and radial displacements and the fifth one for the transverse deflection of the unit. Governing partial differential equations are derived by employing variational principles and minimum potential energy concept. Finite difference method is employed to solve the coupled differential equations. First, they are converted into a system of matrix equations and then iterative procedure is employed. Iterative procedure is necessary since equations are coupled. Problems occurred in getting convergent sequence generated by the employed procedure are overcome by employing variable underrelaxation factor. The procedure developed to solve the differential equations provides not only less storage but also less calculation time, which is a substantial advantage in computational mechanics problems.

Keywords: laminated glass, mathematical model, nonlinear behavior, PVB

Procedia PDF Downloads 304
286 Multi-Agent System Based Solution for Operating Agile and Customizable Micro Manufacturing Systems

Authors: Dylan Santos De Pinho, Arnaud Gay De Combes, Matthieu Steuhlet, Claude Jeannerat, Nabil Ouerhani

Abstract:

The Industry 4.0 initiative has been launched to address huge challenges related to ever-smaller batch sizes. The end-user need for highly customized products requires highly adaptive production systems in order to keep the same efficiency of shop floors. Most of the classical Software solutions that operate the manufacturing processes in a shop floor are based on rigid Manufacturing Execution Systems (MES), which are not capable to adapt the production order on the fly depending on changing demands and or conditions. In this paper, we present a highly modular and flexible solution to orchestrate a set of production systems composed of a micro-milling machine-tool, a polishing station, a cleaning station, a part inspection station, and a rough material store. The different stations are installed according to a novel matrix configuration of a 3x3 vertical shelf. The different cells of the shelf are connected through horizontal and vertical rails on which a set of shuttles circulate to transport the machined parts from a station to another. Our software solution for orchestrating the tasks of each station is based on a Multi-Agent System. Each station and each shuttle is operated by an autonomous agent. All agents communicate with a central agent that holds all the information about the manufacturing order. The core innovation of this paper lies in the path planning of the different shuttles with two major objectives: 1) reduce the waiting time of stations and thus reduce the cycle time of the entire part, and 2) reduce the disturbances like vibration generated by the shuttles, which highly impacts the manufacturing process and thus the quality of the final part. Simulation results show that the cycle time of the parts is reduced by up to 50% compared with MES operated linear production lines while the disturbance is systematically avoided for the critical stations like the milling machine-tool.

Keywords: multi-agent systems, micro-manufacturing, flexible manufacturing, transfer systems

Procedia PDF Downloads 118
285 Molecular Modeling of Structurally Diverse Compounds as Potential Therapeutics for Transmissible Spongiform Encephalopathy

Authors: Sanja O. Podunavac-Kuzmanović, Strahinja Z. Kovačević, Lidija R. Jevrić

Abstract:

Prion is a protein substance whose certain form is considered as infectious agent. It is presumed to be the cause of the transmissible spongiform encephalopathies (TSEs). The protein it is composed of, called PrP, can fold in structurally distinct ways. At least one of those 3D structures is transmissible to other prion proteins. Prions can be found in brain tissue of healthy people and have certain biological role. The structure of prions naturally occurring in healthy organisms is marked as PrPc, and the structure of infectious prion is labeled as PrPSc. PrPc may play a role in synaptic plasticity and neuronal development. Also, it may be required for neuronal myelin sheath maintenance, including a role in iron uptake and iron homeostasis. PrPSc can be considered as an environmental pollutant. The main aim of this study was to carry out the molecular modeling and calculation of molecular descriptors (lipophilicity, physico-chemical and topological descriptors) of structurally diverse compounds which can be considered as anti-prion agents. Molecular modeling was conducted applying ChemBio3D Ultra version 12.0 software. The obtained 3D models were subjected to energy minimization using molecular mechanics force field method (MM2). The cutoff for structure optimization was set at a gradient of 0.1 kcal/Åmol. The Austin Model 1 (AM-1) was used for full geometry optimization of all structures. The obtained set of molecular descriptors is applied in analysis of similarities and dissimilarities among the tested compounds. This study is an important step in further development of quantitative structure-activity relationship (QSAR) models, which can be used for prediction of anti-prion activity of newly synthesized compounds.

Keywords: chemometrics, molecular modeling, molecular descriptors, prions, QSAR

Procedia PDF Downloads 307
284 Adjustments of Mechanical and Hydraulic Properties of Wood Formed under Environmental Stresses

Authors: B. Niez, B. Moulia, J. Dlouha, E. Badel

Abstract:

Trees adjust their development to the environmental conditions they experience. Storms events of last decades showed that acclimation of trees to mechanical stresses due to wind is a very important process that allows the trees to sustain for long years. In the future, trees will experience new wind patterns, namely, more often strong winds and fewer daily moderate winds. Moreover, these patterns will go along with drought periods that may interact with the capacity of trees to adjust their growth to mechanical stresses due to wind. It is necessary to understand the mechanisms of wood functional acclimations to environmental conditions in order to predict their behaviour and in order to give foresters and breeders the relevant tools to adapt their forest management. This work aims to study how trees adjust the mechanical and hydraulic functions of their wood to environmental stresses and how this acclimation may be beneficial for the tree to resist to future stresses. In this work, young poplars were grown under controlled climatic conditions that include permanent environmental stress (daily mechanical stress of the stem by bending and/or hydric stress). Then, the properties of wood formed under these stressed conditions were characterized. First, hydraulic conductivity and sensibility to cavitation were measured at the tissue level in order to evaluate the changes in water transport capacity. Secondly, bending tests and Charpy impact tests were carried out at the millimetric scale to locally measure mechanical parameters such as elastic modulus, elastic limit or rupture energy. These experimental data allow evaluating the impacts of mechanical and water stress on the wood material. At the stem level, they will be merged in an integrative model in order to evaluate the beneficial aspect of wood acclimation for trees.

Keywords: acclimation, environmental stresses, hydraulics, mechanics, wood

Procedia PDF Downloads 188
283 Hydraulic Characteristics of Mine Tailings by Metaheuristics Approach

Authors: Akhila Vasudev, Himanshu Kaushik, Tadikonda Venkata Bharat

Abstract:

A large number of mine tailings are produced every year as part of the extraction process of phosphates, gold, copper, and other materials. Mine tailings are high in water content and have very slow dewatering behavior. The efficient design of tailings dam and economical disposal of these slurries requires the knowledge of tailings consolidation behavior. The large-strain consolidation theory closely predicts the self-weight consolidation of these slurries as the theory considers the conservation of mass and momentum conservation and considers the hydraulic conductivity as a function of void ratio. Classical laboratory techniques, such as settling column test, seepage consolidation test, etc., are expensive and time-consuming for the estimation of hydraulic conductivity variation with void ratio. Inverse estimation of the constitutive relationships from the measured settlement versus time curves is explored. In this work, inverse analysis based on metaheuristics techniques will be explored for predicting the hydraulic conductivity parameters for mine tailings from the base excess pore water pressure dissipation curve and the initial conditions of the mine tailings. The proposed inverse model uses particle swarm optimization (PSO) algorithm, which is based on the social behavior of animals searching for food sources. The finite-difference numerical solution of the forward analytical model is integrated with the PSO algorithm to solve the inverse problem. The method is tested on synthetic data of base excess pore pressure dissipation curves generated using the finite difference method. The effectiveness of the method is verified using base excess pore pressure dissipation curve obtained from a settling column experiment and further ensured through comparison with available predicted hydraulic conductivity parameters.

Keywords: base excess pore pressure, hydraulic conductivity, large strain consolidation, mine tailings

Procedia PDF Downloads 121
282 Literature Review and Approach for the Use of Digital Factory Models in an Augmented Reality Application for Decision Making in Restructuring Processes

Authors: Rene Hellmuth, Jorg Frohnmayer

Abstract:

The requirements of the factory planning and the building concerned have changed in the last years. Factory planning has the task of designing products, plants, processes, organization, areas, and the building of a factory. Regular restructuring gains more importance in order to maintain the competitiveness of a factory. Even today, the methods and process models used in factory planning are predominantly based on the classical planning principles of Schmigalla, Aggteleky and Kettner, which, however, are not specifically designed for reorganization. In addition, they are designed for a largely static environmental situation and a manageable planning complexity as well as for medium to long-term planning cycles with a low variability of the factory. Existing approaches already regard factory planning as a continuous process that makes it possible to react quickly to adaptation requirements. However, digital factory models are not yet used as a source of information for building data. Approaches which consider building information modeling (BIM) or digital factory models in general either do not refer to factory conversions or do not yet go beyond a concept. This deficit can be further substantiated. A method for factory conversion planning using a current digital building model is lacking. A corresponding approach must take into account both the existing approaches to factory planning and the use of digital factory models in practice. A literature review will be conducted first. In it, approaches to classic factory planning and approaches to conversion planning are examined. In addition, it will be investigated which approaches already contain digital factory models. In the second step, an approach is presented how digital factory models based on building information modeling can be used as a basis for augmented reality tablet applications. This application is suitable for construction sites and provides information on the costs and time required for conversion variants. Thus a fast decision making is supported. In summary, the paper provides an overview of existing factory planning approaches and critically examines the use of digital tools. Based on this preliminary work, an approach is presented, which suggests the sensible use of digital factory models for decision support in the case of conversion variants of the factory building. The augmented reality application is designed to summarize the most important information for decision-makers during a reconstruction process.

Keywords: augmented reality, digital factory model, factory planning, restructuring

Procedia PDF Downloads 124
281 Review of Life-Cycle Analysis Applications on Sustainable Building and Construction Sector as Decision Support Tools

Authors: Liying Li, Han Guo

Abstract:

Considering the environmental issues generated by the building sector for its energy consumption, solid waste generation, water use, land use, and global greenhouse gas (GHG) emissions, this review pointed out to LCA as a decision-support tool to substantially improve the sustainability in the building and construction industry. The comprehensiveness and simplicity of LCA make it one of the most promising decision support tools for the sustainable design and construction of future buildings. This paper contains a comprehensive review of existing studies related to LCAs with a focus on their advantages and limitations when applied in the building sector. The aim of this paper is to enhance the understanding of a building life-cycle analysis, thus promoting its application for effective, sustainable building design and construction in the future. Comparisons and discussions are carried out between four categories of LCA methods: building material and component combinations (BMCC) vs. the whole process of construction (WPC) LCA,attributional vs. consequential LCA, process-based LCA vs. input-output (I-O) LCA, traditional vs. hybrid LCA. Classical case studies are presented, which illustrate the effectiveness of LCA as a tool to support the decisions of practitioners in the design and construction of sustainable buildings. (i) BMCC and WPC categories of LCA researches tend to overlap with each other, as majority WPC LCAs are actually developed based on a bottom-up approach BMCC LCAs use. (ii) When considering the influence of social and economic factors outside the proposed system by research, a consequential LCA could provide a more reliable result than an attributional LCA. (iii) I-O LCA is complementary to process-based LCA in order to address the social and economic problems generated by building projects. (iv) Hybrid LCA provides a more superior dynamic perspective than a traditional LCA that is criticized for its static view of the changing processes within the building’s life cycle. LCAs are still being developed to overcome their limitations and data shortage (especially data on the developing world), and the unification of LCA methods and data can make the results of building LCA more comparable and consistent across different studies or even countries.

Keywords: decision support tool, life-cycle analysis, LCA tools and data, sustainable building design

Procedia PDF Downloads 102
280 Graphical Theoretical Construction of Discrete time Share Price Paths from Matroid

Authors: Min Wang, Sergey Utev

Abstract:

The lessons from the 2007-09 global financial crisis have driven scientific research, which considers the design of new methodologies and financial models in the global market. The quantum mechanics approach was introduced in the unpredictable stock market modeling. One famous quantum tool is Feynman path integral method, which was used to model insurance risk by Tamturk and Utev and adapted to formalize the path-dependent option pricing by Hao and Utev. The research is based on the path-dependent calculation method, which is motivated by the Feynman path integral method. The path calculation can be studied in two ways, one way is to label, and the other is computational. Labeling is a part of the representation of objects, and generating functions can provide many different ways of representing share price paths. In this paper, the recent works on graphical theoretical construction of individual share price path via matroid is presented. Firstly, a study is done on the knowledge of matroid, relationship between lattice path matroid and Tutte polynomials and ways to connect points in the lattice path matroid and Tutte polynomials is suggested. Secondly, It is found that a general binary tree can be validly constructed from a connected lattice path matroid rather than general lattice path matroid. Lastly, it is suggested that there is a way to represent share price paths via a general binary tree, and an algorithm is developed to construct share price paths from general binary trees. A relationship is also provided between lattice integer points and Tutte polynomials of a transversal matroid. Use this way of connection together with the algorithm, a share price path can be constructed from a given connected lattice path matroid.

Keywords: combinatorial construction, graphical representation, matroid, path calculation, share price, Tutte polynomial

Procedia PDF Downloads 121
279 Finite Element Analysis for Earing Prediction Incorporating the BBC2003 Material Model with Fully Implicit Integration Method: Derivation and Numerical Algorithm

Authors: Sajjad Izadpanah, Seyed Hadi Ghaderi, Morteza Sayah Irani, Mahdi Gerdooei

Abstract:

In this research work, a sophisticated yield criterion known as BBC2003, capable of describing planar anisotropic behaviors of aluminum alloy sheets, was integrated into the commercial finite element code ABAQUS/Standard via a user subroutine. The complete formulation of the implementation process using a fully implicit integration scheme, i.e., the classic backward Euler method, is presented, and relevant aspects of the yield criterion are introduced. In order to solve nonlinear differential and algebraic equations, the line-search algorithm was adopted in the user-defined material subroutine (UMAT) to expand the convergence domain of the iterative Newton-Raphson method. The developed subroutine was used to simulate a challenging computational problem with complex stress states, i.e., deep drawing of an anisotropic aluminum alloy AA3105. The accuracy and stability of the developed subroutine were confirmed by comparing the numerically predicted earing and thickness variation profiles with the experimental results, which showed an excellent agreement between numerical and experimental earing and thickness profiles. The integration of the BBC2003 yield criterion into ABAQUS/Standard represents a significant contribution to the field of computational mechanics and provides a useful tool for analyzing the mechanical behavior of anisotropic materials subjected to complex loading conditions.

Keywords: BBC2003 yield function, plastic anisotropy, fully implicit integration scheme, line search algorithm, explicit and implicit integration schemes

Procedia PDF Downloads 60