Search results for: multi linear regression
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9405

Search results for: multi linear regression

675 Comparative Review Of Models For Forecasting Permanent Deformation In Unbound Granular Materials

Authors: Shamsulhaq Amin

Abstract:

Unbound granular materials (UGMs) are pivotal in ensuring long-term quality, especially in the layers under the surface of flexible pavements and other constructions. This study seeks to better understand the behavior of the UGMs by looking at popular models for predicting lasting deformation under various levels of stresses and load cycles. These models focus on variables such as the number of load cycles, stress levels, and features specific to materials and were evaluated on the basis of their ability to accurately predict outcomes. The study showed that these factors play a crucial role in how well the models work. Therefore, the research highlights the need to look at a wide range of stress situations to more accurately predict how much the UGMs bend or shift. The research looked at important factors, like how permanent deformation relates to the number of times a load is applied, how quickly this phenomenon happens, and the shakedown effect, in two different types of UGMs: granite and limestone. A detailed study was done over 100,000 load cycles, which provided deep insights into how these materials behave. In this study, a number of factors, such as the level of stress applied, the number of load cycles, the density of the material, and the moisture present were seen as the main factors affecting permanent deformation. It is vital to fully understand these elements for better designing pavements that last long and handle wear and tear. A series of laboratory tests were performed to evaluate the mechanical properties of materials and acquire model parameters. The testing included gradation tests, CBR tests, and Repeated load triaxial tests. The repeated load triaxial tests were crucial for studying the significant components that affect deformation. This test involved applying various stress levels to estimate model parameters. In addition, certain model parameters were established by regression analysis, and optimization was conducted to improve outcomes. Afterward, the material parameters that were acquired were used to construct graphs for each model. The graphs were subsequently compared to the outcomes obtained from the repeated load triaxial testing. Additionally, the models were evaluated to determine if they demonstrated the two inherent deformation behaviors of materials when subjected to repetitive load: the initial phase, post-compaction, and the second phase volumetric changes. In this study, using log-log graphs was key to making the complex data easier to understand. This method made the analysis clearer and helped make the findings easier to interpret, adding both precision and depth to the research. This research provides important insight into picking the right models for predicting how these materials will act under expected stress and load conditions. Moreover, it offers crucial information regarding the effect of load cycle and permanent deformation as well as the shakedown effect on granite and limestone UGMs.

Keywords: permanent deformation, unbound granular materials, load cycles, stress level

Procedia PDF Downloads 20
674 Evaluation of Sequential Polymer Flooding in Multi-Layered Heterogeneous Reservoir

Authors: Panupong Lohrattanarungrot, Falan Srisuriyachai

Abstract:

Polymer flooding is a well-known technique used for controlling mobility ratio in heterogeneous reservoirs, leading to improvement of sweep efficiency as well as wellbore profile. However, low injectivity of viscous polymer solution attenuates oil recovery rate and consecutively adds extra operating cost. An attempt of this study is to improve injectivity of polymer solution while maintaining recovery factor, enhancing effectiveness of polymer flooding method. This study is performed by using reservoir simulation program to modify conventional single polymer slug into sequential polymer flooding, emphasizing on increasing of injectivity and also reduction of polymer amount. Selection of operating conditions for single slug polymer including pre-injected water, polymer concentration and polymer slug size is firstly performed for a layered-heterogeneous reservoir with Lorenz coefficient (Lk) of 0.32. A selected single slug polymer flooding scheme is modified into sequential polymer flooding with reduction of polymer concentration in two different modes: Constant polymer mass and reduction of polymer mass. Effects of Residual Resistance Factor (RRF) is also evaluated. From simulation results, it is observed that first polymer slug with the highest concentration has the main function to buffer between displacing phase and reservoir oil. Moreover, part of polymer from this slug is also sacrificed for adsorption. Reduction of polymer concentration in the following slug prevents bypassing due to unfavorable mobility ratio. At the same time, following slugs with lower viscosity can be injected easily through formation, improving injectivity of the whole process. A sequential polymer flooding with reduction of polymer mass shows great benefit by reducing total production time and amount of polymer consumed up to 10% without any downside effect. The only advantage of using constant polymer mass is slightly increment of recovery factor (up to 1.4%) while total production time is almost the same. Increasing of residual resistance factor of polymer solution yields a benefit on mobility control by reducing effective permeability to water. Nevertheless, higher adsorption results in low injectivity, extending total production time. Modifying single polymer slug into sequence of reduced polymer concentration yields major benefits on reducing production time as well as polymer mass. With certain design of polymer flooding scheme, recovery factor can even be further increased. This study shows that application of sequential polymer flooding can be certainly applied to reservoir with high value of heterogeneity since it requires nothing complex for real implementation but just a proper design of polymer slug size and concentration.

Keywords: polymer flooding, sequential, heterogeneous reservoir, residual resistance factor

Procedia PDF Downloads 459
673 Geochemical Characterization for Identification of Hydrocarbon Generation: Implication of Unconventional Gas Resources

Authors: Yousif M. Makeen

Abstract:

This research will address the processes of geochemical characterization and hydrocarbon generation process occurring within hydrocarbon source and/or reservoir rocks. The geochemical characterization includes organic-inorganic associations that influence the storage capacity of unconventional hydrocarbon resources (e.g. shale gas) and the migration process of oil/gas of the petroleum source/reservoir rocks. Kerogen i.e. the precursor of petroleum, occurs in various forms and types, may either be oil-prone, gas-prone, or both. China has a number of petroleum-bearing sedimentary basins commonly associated with shale gas, oil sands, and oil shale. Taken Sichuan basin as a selected basin in this study, the Sichuan basin has recorded notable successful discoveries of shale gas especially in the marine shale reservoirs within the area. However, a notable discoveries of lacustrine shale in the North-Este Fuling area indicate the accumulation of shale gas within non-marine source rock. The objective of this study is to evaluate the hydrocarbon storage capacity, generation, and retention processes in the rock matrix of hydrocarbon source/reservoir rocks within the Sichuan basin using an advanced X-ray tomography 3D imaging computational technology, commonly referred to as Micro-CT, SEM (Scanning Electron Microscope), optical microscope as well as organic geochemical facilities (e.g. vitrinite reflectance and UV light). The preliminary results of this study show that the lacustrine shales under investigation are acting as both source and reservoir rocks, which are characterized by very fine grains and very low permeability and porosity. Three pore structures have also been characterized in the study in the lacustrine shales, including organic matter pores, interparticle pores and intraparticle pores using x-ray Computed Tomography (CT). The benefits of this study would be a more successful oil and gas exploration and higher recovery factor, thus having a direct economic impact on China and the surrounding region. Methodologies: SRA TOC/TPH or Rock-Eval technique will be used to determine the source rock richness (S1 and S2) and Tmax. TOC analysis will be carried out using a multi N/C 3100 analyzer. The SRA and TOC results were used in calculating other parameters such as hydrogen index (HI) and production index (PI). This analysis will indicate the quantity of the organic matter. Minimum TOC limits generally accepted as essential for a source-rock are 0.5% for shales and 0.2% for carbonates. Contributions: This research could solve issues related to oil potential, provide targets, and serve as a pathfinder to future exploration activity in the Sichuan basin.

Keywords: shale gas, unconventional resources, organic chemistry, Sichuan basin

Procedia PDF Downloads 13
672 Analytical Validity Of A Tech Transfer Solution To Internalize Genetic Testing

Authors: Lesley Northrop, Justin DeGrazia, Jessica Greenwood

Abstract:

ASPIRA Labs now offers an en-suit and ready-to-implement technology transfer solution to enable labs and hospitals that lack the resources to build it themselves to offer in-house genetic testing. This unique platform employs a patented Molecular Inversion Probe (MIP) technology that combines the specificity of a hybrid capture protocol with the ease of an amplicon-based protocol and utilizes an advanced bioinformatics analysis pipeline based on machine learning. To demonstrate its efficacy, two independent genetic tests were validated on this technology transfer platform: expanded carrier screening (ECS) and hereditary cancer testing (HC). The analytical performance of ECS and HC was validated separately in a blinded manner for calling three different types of variants: SNVs, short indels (typically, <50 bp), and large indels/CNVs defined as multi-exonic del/dup events. The reference set was constructed using samples from Coriell Institute, an external clinical genetic testing laboratory, Maine Molecular Quality Controls Inc. (MMQCI), SeraCare and GIAB Consortium. Overall, the analytical performance showed a sensitivity and specificity of >99.4% for both ECS and HC in detecting SNVs. For indels, both tests reported specificity of 100%, and ECS demonstrated a sensitivity of 100%, whereas HC exhibited a sensitivity of 96.5%. The bioinformatics pipeline also correctly called all reference CNV events resulting in a sensitivity of 100% for both tests. No additional calls were made in the HC panel, leading to a perfect performance (specificity and F-measure of 100%). In the carrier panel, however, three additional positive calls were made outside the reference set. Two of these calls were confirmed using an orthogonal method and were re-classified as true positives leaving only one false positive. The pipeline also correctly identified all challenging carrier statuses, such as positive cases for spinal muscular atrophy and alpha-thalassemia, resulting in 100% sensitivity. After confirmation of additional positive calls via long-range PCR and MLPA, specificity for such cases was estimated at 99%. These performance metrics demonstrate that this tech-transfer solution can be confidently internalized by clinical labs and hospitals to offer mainstream ECS and HC as part of their test catalog, substantially increasing access to quality germline genetic testing for labs of all sizes and resources levels.

Keywords: clinical genetics, genetic testing, molecular genetics, technology transfer

Procedia PDF Downloads 163
671 Study of Mechanical Properties of Large Scale Flexible Silicon Solar Modules on the Various Substrates

Authors: M. Maleczek, Leszek Bogdan, Kazimierz Drabczyk, Agnieszka Iwan

Abstract:

Crystalline silicon (Si) solar cells are the main product in the market among the various photovoltaic technologies concerning such advantages as: material richness, high carrier mobilities, broad spectral absorption range and established technology. However, photovoltaic technology on the stiff substrates are heavier, more fragile and less cost-effective than devices on the flexible substrates to be applied in special applications. The main goal of our work was to incorporate silicon solar cells into various fabric, without any change of the electrical and mechanical parameters of devices. This work is realized for the GEKON project (No. GEKON2/O4/268473/23/2016) sponsored by The National Centre for Research and Development and The National Fund for Environmental Protection and Water Management. In our work, the polyamide or polyester fabrics were used as a flexible substrate in the created devices. Applied fabrics differ in tensile and tear strength. All investigated polyamide fabrics are resistant to weathering and UV, while polyester ones is resistant to ozone, water and ageing. The examined fabrics are tight at 100 cm water per 2 hours. In our work, commercial silicon solar cells with the size 156 × 156 mm were cut into nine parts (called single solar cells) by diamond saw and laser. Gap and edge after cutting of solar cells were checked by transmission electron microscope (TEM) to study morphology and quality of the prepared single solar cells. Modules with the size of 160 × 70 cm (containing about 80 single solar cells) were created and investigated by electrical and mechanical methods. Weight of constructed module is about 1.9 kg. Three types of solar cell architectures such as: -fabric/EVA/Si solar cell/EVA/film for lamination, -backsheet PET/EVA/Si solar cell/EVA/film for lamination, -fabric/EVA/Si solar cell/EVA/tempered glass, were investigated taking into consideration type of fabric and lamination process together with the size of solar cells. In investigated devices EVA, it is ethylene-vinyl acetate, while PET - polyethylene terephthalate. Depend on the lamination process and compatibility of textile with solar cell an efficiency of investigated flexible silicon solar cells was in the range of 9.44-16.64 %. Multi folding and unfolding of flexible module has no impact on its efficiency as was detected by Instron equipment. Power (P) of constructed solar module is 30 W, while voltage about 36 V. Finally, solar panel contains five modules with the polyamide fabric and tempered glass will be produced commercially for different applications (dual use).

Keywords: flexible devices, mechanical properties, silicon solar cells, textiles

Procedia PDF Downloads 161
670 Measuring Systems Interoperability: A Focal Point for Standardized Assessment of Regional Disaster Resilience

Authors: Joel Thomas, Alexa Squirini

Abstract:

The key argument of this research is that every element of systems interoperability is an enabler of regional disaster resilience, and arguably should become a focal point for standardized measurement of communities’ ability to work together. Few resilience research efforts have focused on the development and application of solutions that measurably improve communities’ ability to work together at a regional level, yet a majority of the most devastating and disruptive disasters are those that have had a regional impact. The key findings of the research include a unique theoretical, mathematical, and operational approach to tangibly and defensibly measure and assess systems interoperability required to support crisis information management activities performed by governments, the private sector, and humanitarian organizations. A most effective way for communities to measurably improve regional disaster resilience is through deliberately executed disaster preparedness activities. Developing interoperable crisis information management capabilities is a crosscutting preparedness activity that greatly affects a community’s readiness and ability to work together in times of crisis. Thus, improving communities’ human and technical posture to work together in advance of a crisis, with the ultimate goal of enabling information sharing to support coordination and the careful management of available resources, is a primary means by which communities may improve regional disaster resilience. This model describes how systems interoperability can be qualitatively and quantitatively assessed when characterized as five forms of capital: governance; standard operating procedures; technology; training and exercises; and usage. The unique measurement framework presented defines the relationships between systems interoperability, information sharing and safeguarding, operational coordination, community preparedness and regional disaster resilience, and offers a means by which to implement real-world solutions and measure progress over the course of a multi-year program. The model is being developed and piloted in partnership with the U.S. Department of Homeland Security (DHS) Science and Technology Directorate (S&T) and the North Atlantic Treaty Organization (NATO) Advanced Regional Civil Emergency Coordination Pilot (ARCECP) with twenty-three organizations in Bosnia and Herzegovina, Croatia, Macedonia, and Montenegro. The intended effect of the model implementation is to enable communities to answer two key questions: 'Have we measurably improved crisis information management capabilities as a result of this effort?' and, 'As a result, are we more resilient?'

Keywords: disaster, interoperability, measurement, resilience

Procedia PDF Downloads 122
669 A Temporary Shelter Proposal for Displaced People

Authors: İrem Yetkin, Feray Maden, Seda Tosun, Yenal Akgün, Özgür Kilit, Koray Korkmaz, Gökhan Kiper, Mustafa Gündüzalp

Abstract:

Forced migration, whether caused by conflicts or other factors, frequently places individuals in vulnerable situations, necessitating immediate access to shelter. To promptly address the immediate needs of affected individuals, temporary shelters are often established. These shelters are characterized by their adaptable and functional nature, encompassing lightweight and sustainable structural systems, rapid assembly capabilities, modularity, and transportability. The shelter design is contingent upon demand, resulting in distinct phases for different structural forms. A multi-phased shelter approach covers emergency response, temporary shelter, and permanent reconstruction. Emergency shelters play a critical role in providing immediate life-saving aid, while temporary and transitional shelters, which are also called “t-shelters,” offer longer-term living environments during the recovery and rebuilding phases. Among these, temporary shelters are more extensively covered in the literature due to their diverse inhabiting functions. The roles of emergency shelters and temporary shelters are inherently separate, addressing distinct aspects of sheltering processes. Given their prolonged usage, temporary shelters are built for greater durability compared to emergency shelters. Nonetheless, inadequacies in temporary shelters can lead to challenges in ensuring habitability. Issues like non-expandable structures unsuitable for accommodating large families, the use of short-term shelters that worsen conditions, non-waterproof materials providing insufficient protection against bad weather conditions, and complex installation systems contribute to these problems. Given the aforementioned problems, there arises a need to develop adaptive shelters featuring lightweight components for ease of transport, possess the ability for rapid assembly, and utilize durable materials to withstand adverse weather conditions. In this study, first, the state-of-the-art on temporary shelters is presented. Then, an adaptive temporary shelter composed of foldable plates is proposed, which can easily be assembled and transportable. The proposed shelter is deliberated upon its movement capacity, transportability, and flexibility. This study makes a valuable contribution to the literature since it not only offers a systematic analysis of temporary shelters utilizing kinetic systems but also presents a practical solution that meets the necessary design requirements.

Keywords: deployable structures, foldable plates, forced migration, temporary shelters

Procedia PDF Downloads 52
668 The Relevance of Personality Traits and Networking in New Ventures’ Success

Authors: Caterina Muzzi, Sergio Albertini, Davide Giacomini

Abstract:

The research is aimed to investigate the role of young entrepreneurs’ personality traits and their contextual background on the success of entrepreneurial initiatives. In the literature, the debate is still open about the main drivers in predicting entrepreneurial success. Classical theories are focused on looking at specific personality traits that could lead to successful start-ups initiatives, while emerging approaches are more interested in young entrepreneurs’ contextual background (such as the family of origin, the previous experience and their professional network). An online survey was submitted to the participants of an entrepreneurial training initiative organised by the Italian Young Entrepreneurs Association (Confindustria) in Brescia headquarter (AIB). At the time the authors started data collection for this research, the third edition of the initiative was just concluded and involved a total amount of 37 young future entrepreneurs. In the literature General self-efficacy (GSE) and, more specifically, entrepreneurial self-efficacy (ESE) have often been associated to positive performances, as they allow future entrepreneurs to effectively cope with entrepreneurial activities, both at an early stage and in new venture management. In a counter-intuitive manner, optimism is not always associated with entrepreneurial positive results. Too optimistic people risk taking hazardous risks and some authors suggest that moderately optimistic entrepreneurs achieve more positive results than over-optimistic ones. Indeed highly optimistic individuals often hold unrealistic expectations, discount negative information, and mentally reconstruct experiences so as to avoid contradictions The importance of context has been increasingly considered in entrepreneurship literature and its role strongly emerges starting from the earliest entrepreneurial stage and it is crucial to transform the “intention of entrepreneurship” into the actual start-up. Furthermore, coherently with the “network approach to entrepreneurship”, context embeddedness allow future entrepreneurs to leverage relationships built through previous experiences and/or thanks to the fact of belonging to families of entrepreneurs. For the purpose of this research, entrepreneurial success was measured by the fact of having or not founded a new venture after the training initiative. In this research, the authors measured GSE, ESE and optimism using already tested items that showed to be reliable also in this case. They collected 36 completed questionnaires. The t-test for independent samples run to measure significant differences in means between those that already funded the new venture and those that did not. No significant differences emerged with respect to all the tested personality traits, but a logistic regression analysis, run with contextual variables as independent ones, showed that personal and professional networking, made both before and during the master, is the most relevant variable in determining new venture success. These findings shed more light on the process of new venture foundation and could encourage national and local policy makers to invest on networking as one of the main drivers that could support the creation of new ventures.

Keywords: entrepreneurship, networking, new ventures, personality traits

Procedia PDF Downloads 127
667 An Australian Tertiary Centre Experience of Complex Endovascular Aortic Repairs

Authors: Hansraj Bookun, Rachel Xuan, Angela Tan, Kejia Wang, Animesh Singla, David Kim, Christopher Loupos, Jim Iliopoulos

Abstract:

Introduction: Complex endovascular aortic aneursymal repairs with fenestrated and branched endografts require customised devices to exclude the pathology while reducing morbidity and mortality, which was historically associated with open repair of complex aneurysms. Such endovascular procedures have predominantly been performed in a large volume dedicated tertiary centres. We present here our nine year multidisciplinary experience with this technology in an Australian tertiary centre. Method: This was a cross-sectional, single-centre observational study of 670 patients who had undergone complex endovascular aortic aneurysmal repairs with conventional endografts, fenestrated endografts, and iliac-branched devices from January 2010 to July 2019. Descriptive statistics were used to characterise our sample with regards to demographic and perioperative variables. Homogeneity of the sample was tested using multivariant regression, which did not identify any statistically significant confounding variables. Results: 670 patients of mean age 74, were included (592 males) and the comorbid burden was as follows: ischemic heart disease (55%), diabetes (18%), hypertension (90%), stage four or greater kidney impairment (8%) and current or ex-smoking (78%). The main indications for surgery were elective aneurysms (86%), symptomatic aneurysms (5%), and rupture aneurysms (5%). 106 patients (16%) underwent fenestrated or branched endograft repairs. The mean length of stay was 7.6 days. 2 patients experienced reactionary bleeds, 11 patients had access wound complications (6 lymph fistulae, 5 haematoms), 11 patients had cardiac complications (5 arrhythmias, 3 acute myocadial infarctions, 3 exacerbation of congestive cardiac failure), 10 patients had respiratory complications, 8 patients had renal impairment, 4 patients had gastrointestinal complications, 2 patients suffered from paraplegia, 1 major stroke, 1 minor stroke, and 1 acute brain syndrome. There were 4 vascular occlusions requiring further arterial surgery, 4 type I endoleaks, 4 type II endoleaks, 3 episodes of thromboembolism, and 2 patients who required further arterial operations in the setting of patient vessels. There were 9 unplanned returns to the theatre. Discussion: Our numbers of 10 years suggest that we are not a dedicated high volume centre focusing on aortic repairs. However, we have achieved significantly low complication rates. This can be attributed to our multidisciplinary approach with the intraoperative involvement of skilled interventional radiologists and vascular surgeons as well as postoperative protocols with particular attention to spinal cord protection. Additionally, we have a ratified perioperative pathway that involves multidisciplinary team discussions of patient-related factors and lesion-centered characteristics, which allows for holistic, patient-centered care.

Keywords: aneurysm, aortic, endovascular, fenestrated

Procedia PDF Downloads 99
666 Ambivilance, Denial, and Adaptive Responses to Vulnerable Suspects in Police Custody: The New Limits of the Sovereign State

Authors: Faye Cosgrove, Donna Peacock

Abstract:

This paper examines current state strategies for dealing with vulnerable people in police custody and identifies the underpinning discourses and practices which inform these strategies. It has previously been argued that the state has utilised contradictory and conflicting responses to the control of crime, by employing opposing strategies of denial and adaptation in order to simultaneously both display sovereignty and disclaim responsibility. This paper argues that these contradictory strategies are still being employed in contemporary criminal justice, although the focus and the purpose have now shifted. The focus is upon the ‘vulnerable’ suspect, whose social identity is as incongruous, complex and contradictory as his social environment, and the purpose is to redirect attention away from negative state practices, whilst simultaneously displaying a compassionate and benevolent countenance in order to appeal to the voting public. The findings presented here result from intensive qualitative research with police officers, with health care professionals, and with civilian volunteers who work within police custodial environments. The data has been gathered over a three-year period and includes observational and interview data which has been thematically analysed to expose the underpinning mechanisms from which the properties of the system emerge. What is revealed is evidence of contemporary state practices of denial relating to the harms of austerity and the structural relations of vulnerability, whilst simultaneously adapting through processes of ‘othering’ of the vulnerable, ‘responsibilisation’ of citizens, defining deviance down through diversionary practices, and managing success through redefining the aims of the system. The ‘vulnerable’ suspect is subject to individual pathologising, and yet the nature of risk is aggregated. ‘Vulnerable’ suspects are supported in police custody by private citizens, by multi-agency partnerships, and by for-profit organisations, while the state seeks to collate and control services, and thereby to retain a veneer of control. Late modern ambivalence to crime control and the associated contradictory practices of abjuration and adjustment have extended to state responses to vulnerable suspects. The support available in the custody environment operates to control and minimise operational and procedural risk, rather than for the welfare of the detained person, and in fact, the support available is discovered to be detrimental to the very people that it claims to benefit. The ‘vulnerable’ suspect is now subject to the bifurcated logics employed at the new limits of the sovereign state.

Keywords: custody, policing, sovereign state, vulnerability

Procedia PDF Downloads 152
665 Fuzzy Optimization for Identifying Anticancer Targets in Genome-Scale Metabolic Models of Colon Cancer

Authors: Feng-Sheng Wang, Chao-Ting Cheng

Abstract:

Developing a drug from conception to launch is costly and time-consuming. Computer-aided methods can reduce research costs and accelerate the development process during the early drug discovery and development stages. This study developed a fuzzy multi-objective hierarchical optimization framework for identifying potential anticancer targets in a metabolic model. First, RNA-seq expression data of colorectal cancer samples and their healthy counterparts were used to reconstruct tissue-specific genome-scale metabolic models. The aim of the optimization framework was to identify anticancer targets that lead to cancer cell death and evaluate metabolic flux perturbations in normal cells that have been caused by cancer treatment. Four objectives were established in the optimization framework to evaluate the mortality of cancer cells for treatment and to minimize side effects causing toxicity-induced tumorigenesis on normal cells and smaller metabolic perturbations. Through fuzzy set theory, a multiobjective optimization problem was converted into a trilevel maximizing decision-making (MDM) problem. The applied nested hybrid differential evolution was applied to solve the trilevel MDM problem using two nutrient media to identify anticancer targets in the genome-scale metabolic model of colorectal cancer, respectively. Using Dulbecco’s Modified Eagle Medium (DMEM), the computational results reveal that the identified anticancer targets were mostly involved in cholesterol biosynthesis, pyrimidine and purine metabolisms, glycerophospholipid biosynthetic pathway and sphingolipid pathway. However, using Ham’s medium, the genes involved in cholesterol biosynthesis were unidentifiable. A comparison of the uptake reactions for the DMEM and Ham’s medium revealed that no cholesterol uptake reaction was included in DMEM. Two additional media, i.e., a cholesterol uptake reaction was included in DMEM and excluded in HAM, were respectively used to investigate the relationship of tumor cell growth with nutrient components and anticancer target genes. The genes involved in the cholesterol biosynthesis were also revealed to be determinable if a cholesterol uptake reaction was not induced when the cells were in the culture medium. However, the genes involved in cholesterol biosynthesis became unidentifiable if such a reaction was induced.

Keywords: Cancer metabolism, genome-scale metabolic model, constraint-based model, multilevel optimization, fuzzy optimization, hybrid differential evolution

Procedia PDF Downloads 63
664 Cultural Omnivorousness in Fikirtepe Urban Regenaration Area

Authors: Burcin Basyazici

Abstract:

The history of urban regeneration in Istanbul dates to the 1980s and has been generated by various reasons, from political state-based decisions to cultural migrations and/or translocations and economical private sector-based reasons. However, one of the latest regeneration areas in Fikirtepe that have been still under construction for ten years becomes dissimilar to other regeneration areas in Istanbul. The region is located very close to Kadıkoy's downtown area but was still considered a slum due to its inhabitants -who mostly belong to lower-income immigrants. The process begun in 2011 with the decision of the government, and the settlement has been emptied and demolished also by the government -together with the investors and construction companies. Although there has been much research on the process of deconstruction and the relocation of landlords, there hasn’t been any research on what happened after the regeneration. While many high-rise luxurious gated communities were constructed and inhabited in five years, many constructions have stopped due to the latest economic devastation in Turkey. Then the region stayed as an unfinished construction area with its new upper-income and upper-middle-income residents and old low-income landlords. This situation has also changed the commercial activities in Fikirtepe. While some new retail facilities have been offered for new residents, some of the oldest ones have also survived in new-Fikirtepe. This study aims to investigate the urban everyday life of Fikirtepe with relation to its retail-based regeneration with the help of the theories of Bourdieu called cultural capital and cultural omnivorousness. To achieve this aim, after presenting the historical background of urban regeneration in Istanbul, Bourdieu’s conceptualism of cultural capital, habitus, and the consumption tendencies related to those are introduced and discussed within the scope of the Fikirtepe case. To represent the retail-based regeneration in the area, the current situation of retail premises is mapped by comparing to its pre-situation before urban gentrification. To better understand the change of cultural capital and the consumption tendencies of the new residents, eighteen semi-structured in-depth interviews have been conducted with twelve inhabitants from three different luxurious gated communities and six shop owners containing the new ones after regeneration and old ones before it. The interview questions have been structured to understand the motivation of change and/or inhibition of retail premises and the consumption tendencies of the new residents. In conclusion, the study shows that even though the cultural capital has been changed in Fikirtepe, the new residents also tend to act as culturally omnivorous by referring to Bourdieu's theories on multi-cultural tendencies of the upper-class and upper-middle-class societies, that should be questioned regarding the cultural regeneration in in-town urban regeneration areas in metropoles.

Keywords: bourdieu, cultural omnivorousness, fikirtepe, urban regeneration in istanbul

Procedia PDF Downloads 148
663 Genetic Dissection of QTLs in Intraspecific Hybrids Derived from Muskmelon (Cucumis Melo L.) and Mangalore Melon (Cucumis Melo Var Acidulus) for Shelflife and Fruit Quality Traits

Authors: Virupakshi Hiremata, Ratnakar M. Shet, Raghavendra Gunnaiah, Prashantha A.

Abstract:

Muskmelon is a health-beneficial and refreshing dessert vegetable with a low shelf life. Mangalore melon, a genetic homeologue of muskmelon, has a shelf life of more than six months and is mostly used for culinary purposes. Understanding the genetics of shelf life, yield and yield-related traits and identification of markers linked to such traits is helpful in transfer of extended shelf life from Mangalore melon to the muskmelon through intra-specific hybridization. For QTL mapping, 276 F2 mapping population derived from the cross Arka Siri × SS-17 was genotyped with 40 polymorphic markers distributed across 12 chromosomes. The same population was also phenotyped for yield, shelf life and fruit quality traits. One major QTL (R2 >10) and fourteen minor QTLs (R2 <10) localized on four linkage groups, governing different traits were mapped in F2 mapping population developed from the intraspecific cross with a LOD > 5.5. The phenotypic varience explained by each locus varied from 3.63 to 10.97 %. One QTL was linked to shelf-life (qSHL-3-1), five QTLs were linked to TSS (qTSS-1-1, qTSS-3-3, qTSS-3-1, qTSS-3-2 and qTSS-1-2), two QTLs for flesh thickness (qFT-3-1, and qFT-3-2) and seven QTLs for fruit yield per vine (qFYV-3-1, qFYV-1-1, qFYV-3-1, qFYV1-1, qFYV-1-3, qFYV2-1 and qFYV6-1). QTL flanking markers may be used for marker assisted introgression of shelf life into muskmelon. Important QTL will be further fine-mapped for identifying candidate genes by QTLseq and RNAseq analysis. Fine-mapping of Important Quantitative Trait Loci (QTL) holds immense promise in elucidating the genetic basis of complex traits. Leveraging advanced techniques like QTLseq and RNA sequencing (RNA seq) is crucial for this endeavor. QTLseq combines next-generation sequencing with traditional QTL mapping, enabling precise identification of genomic regions associated with traits of interest. Through high-throughput sequencing, QTLseq provides a detailed map of genetic variations linked to phenotypic variations, facilitating targeted investigations. Moreover, RNA seq analysis offers a comprehensive view of gene expression patterns in response to specific traits or conditions. By comparing transcriptomes between contrasting phenotypes, RNA seq aids in pinpointing candidate genes underlying QTL regions. Integrating QTLseq with RNA seq allows for a multi-dimensional approach, coupling genetic variation with gene expression dynamics.

Keywords: QTL, shelf life, TSS, muskmelon and Mangalore melon

Procedia PDF Downloads 30
662 The Complementary Effect of Internal Control System and Whistleblowing Policy on Prevention and Detection of Fraud in Nigerian Deposit Money Banks

Authors: Dada Durojaye Joshua

Abstract:

The study examined the combined effect of internal control system and whistle blowing policy while it pursues the following specific objectives, which are to: examine the relationship between monitoring activities and fraud’s detection and prevention; investigate the effect of control activities on fraud’s detection and prevention in Nigerian Deposit Money Banks (DMBs). The population of the study comprises the 89,275 members of staff in the 20 DMBs in Nigeria as at June 2019. Purposive and convenient sampling techniques were used in the selection of the 80 members of staff at the supervisory level of the Internal Audit Departments of the head offices of the sampled banks, that is, selecting 4 respondents (Audit Executive/Head, Internal Control; Manager, Operation Risk Management; Head, Financial Crime Control; the Chief Compliance Officer) from each of the 20 DMBs in Nigeria. A standard questionnaire was adapted from 2017/2018 Internal Control Questionnaire and Assessment, Bureau of Financial Monitoring and Accountability Florida Department of Economic Opportunity. It was modified to serve the purpose for which it was meant to serve. It was self-administered to gather data from the 80 respondents at the respective headquarters of the sampled banks at their respective locations across Nigeria. Two likert-scales was used in achieving the stated objectives. A logit regression was used in analysing the stated hypotheses. It was found that effect of monitoring activities using the construct of conduct of ongoing or separate evaluation (COSE), evaluation and communication of deficiencies (ECD) revealed that monitoring activities is significant and positively related to fraud’s detection and prevention in Nigerian DMBS. So also, it was found that control activities using selection and development of control activities (SDCA), selection and development of general controls over technology to prevent financial fraud (SDGCTF), development of control activities that gives room for transparency through procedures that put policies into actions (DCATPPA) contributed to influence fraud detection and prevention in the Nigerian DMBs. In addition, it was found that transparency, accountability, reliability, independence and value relevance have significant effect on fraud detection and prevention ibn Nigerian DMBs. The study concluded that the board of directors demonstrated independence from management and exercises oversight of the development and performance of internal control. Part of the conclusion was that there was accountability on the part of the owners and preparers of the financial reports and that the system gives room for the members of staff to account for their responsibilities. Among the recommendations was that the management of Nigerian DMBs should create and establish a standard Internal Control System strong enough to deter fraud in order to encourage continuity of operations by ensuring liquidity, solvency and going concern of the banks. It was also recommended that the banks create a structure that encourages whistleblowing to complement the internal control system.

Keywords: internal control, whistleblowing, deposit money banks, fraud prevention, fraud detection

Procedia PDF Downloads 62
661 Plotting of an Ideal Logic versus Resource Outflow Graph through Response Analysis on a Strategic Management Case Study Based Questionnaire

Authors: Vinay A. Sharma, Shiva Prasad H. C.

Abstract:

The initial stages of any project are often observed to be in a mixed set of conditions. Setting up the project is a tough task, but taking the initial decisions is rather not complex, as some of the critical factors are yet to be introduced into the scenario. These simple initial decisions potentially shape the timeline and subsequent events that might later be plotted on it. Proceeding towards the solution for a problem is the primary objective in the initial stages. The optimization in the solutions can come later, and hence, the resources deployed towards attaining the solution are higher than what they would have been in the optimized versions. A ‘logic’ that counters the problem is essentially the core of the desired solution. Thus, if the problem is solved, the deployment of resources has led to the required logic being attained. As the project proceeds along, the individuals working on the project face fresh challenges as a team and are better accustomed to their surroundings. The developed, optimized solutions are then considered for implementation, as the individuals are now experienced, and know better of the consequences and causes of possible failure, and thus integrate the adequate tolerances wherever required. Furthermore, as the team graduates in terms of strength, acquires prodigious knowledge, and begins its efficient transfer, the individuals in charge of the project along with the managers focus more on the optimized solutions rather than the traditional ones to minimize the required resources. Hence, as time progresses, the authorities prioritize attainment of the required logic, at a lower amount of dedicated resources. For empirical analysis of the stated theory, leaders and key figures in organizations are surveyed for their ideas on appropriate logic required for tackling a problem. Key-pointers spotted in successfully implemented solutions are noted from the analysis of the responses and a metric for measuring logic is developed. A graph is plotted with the quantifiable logic on the Y-axis, and the dedicated resources for the solutions to various problems on the X-axis. The dedicated resources are plotted over time, and hence the X-axis is also a measure of time. In the initial stages of the project, the graph is rather linear, as the required logic will be attained, but the consumed resources are also high. With time, the authorities begin focusing on optimized solutions, since the logic attained through them is higher, but the resources deployed are comparatively lower. Hence, the difference between consecutive plotted ‘resources’ reduces and as a result, the slope of the graph gradually increases. On an overview, the graph takes a parabolic shape (beginning on the origin), as with each resource investment, ideally, the difference keeps on decreasing, and the logic attained through the solution keeps increasing. Even if the resource investment is higher, the managers and authorities, ideally make sure that the investment is being made on a proportionally high logic for a larger problem, that is, ideally the slope of the graph increases with the plotting of each point.

Keywords: decision-making, leadership, logic, strategic management

Procedia PDF Downloads 92
660 Complementary Effect of Wistleblowing Policy and Internal Control System on Prevention and Detection of Fraud in Nigerian Deposit Money Banks

Authors: Dada Durojaye Joshua

Abstract:

The study examined the combined effect of internal control system and whistle blowing policy while it pursues the following specific objectives, which are to: examine the relationship between monitoring activities and fraud’s detection and prevention; investigate the effect of control activities on fraud’s detection and prevention in Nigerian Deposit Money Banks (DMBs). The population of the study comprises the 89,275 members of staff in the 20 DMBs in Nigeria as at June 2019. Purposive and convenient sampling techniques were used in the selection of the 80 members of staff at the supervisory level of the Internal Audit Departments of the head offices of the sampled banks, that is, selecting 4 respondents (Audit Executive/Head, Internal Control; Manager, Operation Risk Management; Head, Financial Crime Control; the Chief Compliance Officer) from each of the 20 DMBs in Nigeria. A standard questionnaire was adapted from 2017/2018 Internal Control Questionnaire and Assessment, Bureau of Financial Monitoring and Accountability Florida Department of Economic Opportunity. It was modified to serve the purpose for which it was meant to serve. It was self-administered to gather data from the 80 respondents at the respective headquarters of the sampled banks at their respective locations across Nigeria. Two likert-scales was used in achieving the stated objectives. A logit regression was used in analysing the stated hypotheses. It was found that effect of monitoring activities using the construct of conduct of ongoing or separate evaluation (COSE), evaluation and communication of deficiencies (ECD) revealed that monitoring activities is significant and positively related to fraud’s detection and prevention in Nigerian DMBS. So also, it was found that control activities using selection and development of control activities (SDCA), selection and development of general controls over technology to prevent financial fraud (SDGCTF), development of control activities that gives room for transparency through procedures that put policies into actions (DCATPPA) contributed to influence fraud detection and prevention in the Nigerian DMBs. In addition, it was found that transparency, accountability, reliability, independence and value relevance have significant effect on fraud detection and prevention ibn Nigerian DMBs. The study concluded that the board of directors demonstrated independence from management and exercises oversight of the development and performance of internal control. Part of the conclusion was that there was accountability on the part of the owners and preparers of the financial reports and that the system gives room for the members of staff to account for their responsibilities. Among the recommendations was that the management of Nigerian DMBs should create and establish a standard Internal Control System strong enough to deter fraud in order to encourage continuity of operations by ensuring liquidity, solvency and going concern of the banks. It was also recommended that the banks create a structure that encourages whistleblowing to complement the internal control system.

Keywords: internal control, whistleblowing, deposit money banks, fraud prevention, fraud detection

Procedia PDF Downloads 53
659 Integrated Geophysical Surveys for Sinkhole and Subsidence Vulnerability Assessment, in the West Rand Area of Johannesburg

Authors: Ramoshweu Melvin Sethobya, Emmanuel Chirenje, Mihlali Hobo, Simon Sebothoma

Abstract:

The recent surge in residential infrastructure development around the metropolitan areas of South Africa has necessitated conditions for thorough geotechnical assessments to be conducted prior to site developments to ensure human and infrastructure safety. This paper appraises the success in the application of multi-method geophysical techniques for the delineation of sinkhole vulnerability in a residential landscape. Geophysical techniques ERT, MASW, VES, Magnetics and gravity surveys were conducted to assist in mapping sinkhole vulnerability, using an existing sinkhole as a constraint at Venterspost town, West of Johannesburg city. A combination of different geophysical techniques and results integration from those proved to be useful in the delineation of the lithologic succession around sinkhole locality, and determining the geotechnical characteristics of each layer for its contribution to the development of sinkholes, subsidence and cavities at the vicinity of the site. Study results have also assisted in the determination of the possible depth extension of the currently existing sinkhole and the location of sites where other similar karstic features and sinkholes could form. Results of the ERT, VES and MASW surveys have uncovered dolomitic bedrock at varying depths around the sites, which exhibits high resistivity values in the range 2500-8000ohm.m and corresponding high velocities in the range 1000-2400 m/s. The dolomite layer was found to be overlain by a weathered chert-poor dolomite layer, which has resistivities between the range 250-2400ohm.m, and velocities ranging from 500-600m/s, from which the large sinkhole has been found to collapse/ cave in. A compiled 2.5D high resolution Shear Wave Velocity (Vs) map of the study area was created using 2D profiles of MASW data, offering insights into the prevailing lithological setup conducive for formation various types of karstic features around the site. 3D magnetic models of the site highlighted the regions of possible subsurface interconnections between the currently existing large sinkhole and the other subsidence feature at the site. A number of depth slices were used to detail the conditions near the sinkhole as depth increases. Gravity surveys results mapped the possible formational pathways for development of new karstic features around the site. Combination and correlation of different geophysical techniques proved useful in delineation of the site geotechnical characteristics and mapping the possible depth extend of the currently existing sinkhole.

Keywords: resistivity, magnetics, sinkhole, gravity, karst, delineation, VES

Procedia PDF Downloads 51
658 Applying Big Data Analysis to Efficiently Exploit the Vast Unconventional Tight Oil Reserves

Authors: Shengnan Chen, Shuhua Wang

Abstract:

Successful production of hydrocarbon from unconventional tight oil reserves has changed the energy landscape in North America. The oil contained within these reservoirs typically will not flow to the wellbore at economic rates without assistance from advanced horizontal well and multi-stage hydraulic fracturing. Efficient and economic development of these reserves is a priority of society, government, and industry, especially under the current low oil prices. Meanwhile, society needs technological and process innovations to enhance oil recovery while concurrently reducing environmental impacts. Recently, big data analysis and artificial intelligence become very popular, developing data-driven insights for better designs and decisions in various engineering disciplines. However, the application of data mining in petroleum engineering is still in its infancy. The objective of this research aims to apply intelligent data analysis and data-driven models to exploit unconventional oil reserves both efficiently and economically. More specifically, a comprehensive database including the reservoir geological data, reservoir geophysical data, well completion data and production data for thousands of wells is firstly established to discover the valuable insights and knowledge related to tight oil reserves development. Several data analysis methods are introduced to analysis such a huge dataset. For example, K-means clustering is used to partition all observations into clusters; principle component analysis is applied to emphasize the variation and bring out strong patterns in the dataset, making the big data easy to explore and visualize; exploratory factor analysis (EFA) is used to identify the complex interrelationships between well completion data and well production data. Different data mining techniques, such as artificial neural network, fuzzy logic, and machine learning technique are then summarized, and appropriate ones are selected to analyze the database based on the prediction accuracy, model robustness, and reproducibility. Advanced knowledge and patterned are finally recognized and integrated into a modified self-adaptive differential evolution optimization workflow to enhance the oil recovery and maximize the net present value (NPV) of the unconventional oil resources. This research will advance the knowledge in the development of unconventional oil reserves and bridge the gap between the big data and performance optimizations in these formations. The newly developed data-driven optimization workflow is a powerful approach to guide field operation, which leads to better designs, higher oil recovery and economic return of future wells in the unconventional oil reserves.

Keywords: big data, artificial intelligence, enhance oil recovery, unconventional oil reserves

Procedia PDF Downloads 267
657 The Impact of Shifting Trading Pattern from Long-Haul to Short-Sea to the Car Carriers’ Freight Revenues

Authors: Tianyu Wang, Nikita Karandikar

Abstract:

The uncertainty around cost, safety, and feasibility of the decarbonized shipping fuels has made it increasingly complex for the shipping companies to set pricing strategies and forecast their freight revenues going forward. The increase in the green fuel surcharges will ultimately influence the automobile’s consumer prices. The auto shipping demand (ton-miles) has been gradually shifting from long-haul to short-sea trade over the past years following the relocation of the original equipment manufacturer (OEM) manufacturing to regions such as South America and Southeast Asia. The objective of this paper is twofold: 1) to investigate the car-carriers freight revenue development over the years when the trade pattern is gradually shifting towards short-sea exports 2) to empirically identify the quantitative impact of such trade pattern shifting to mainly freight rate, but also vessel size, fleet size as well as Green House Gas (GHG) emission in Roll on-Roll Off (Ro-Ro) shipping. In this paper, a model of analyzing and forecasting ton-miles and freight revenues for the trade routes of AS-NA (Asia to North America), EU-NA (Europe to North America), and SA-NA (South America to North America) is established by deploying Automatic Identification System (AIS) data and the financial results of a selected car carrier company. More specifically, Wallenius Wilhelmsen Logistics (WALWIL), the Norwegian Ro-Ro carrier listed on Oslo Stock Exchange, is selected as the case study company in this paper. AIS-based ton-mile datasets of WALWIL vessels that are sailing into North America region from three different origins (Asia, Europe, and South America), together with WALWIL’s quarterly freight revenues as reported in trade segments, will be investigated and compared for the past five years (2018-2022). Furthermore, ordinary‐least‐square (OLS) regression is utilized to construct the ton-mile demand and freight revenue forecasting. The determinants of trade pattern shifting, such as import tariffs following the China-US trade war and fuel prices following the 0.1% Emission Control Areas (ECA) zone requirement after IMO2020 will be set as key variable inputs to the machine learning model. The model will be tested on another newly listed Norwegian Car Carrier, Hoegh Autoliner, to forecast its 2022 financial results and to validate the accuracy based on its actual results. GHG emissions on the three routes will be compared and discussed based on a constant emission per mile assumption and voyage distances. Our findings will provide important insights about 1) the trade-off evaluation between revenue reduction and energy saving with the new ton-mile pattern and 2) how the trade flow shifting would influence the future need for the vessel and fleet size.

Keywords: AIS, automobile exports, maritime big data, trade flows

Procedia PDF Downloads 102
656 MBES-CARIS Data Validation for the Bathymetric Mapping of Shallow Water in the Kingdom of Bahrain on the Arabian Gulf

Authors: Abderrazak Bannari, Ghadeer Kadhem

Abstract:

The objectives of this paper are the validation and the evaluation of MBES-CARIS BASE surface data performance for bathymetric mapping of shallow water in the Kingdom of Bahrain. The latter is an archipelago with a total land area of about 765.30 km², approximately 126 km of coastline and 8,000 km² of marine area, located in the Arabian Gulf, east of Saudi Arabia and west of Qatar (26° 00’ N, 50° 33’ E). To achieve our objectives, bathymetric attributed grid files (X, Y, and depth) generated from the coverage of ship-track MBSE data with 300 x 300 m cells, processed with CARIS-HIPS, were downloaded from the General Bathymetric Chart of the Oceans (GEBCO). Then, brought into ArcGIS and converted into a raster format following five steps: Exportation of GEBCO BASE surface data to the ASCII file; conversion of ASCII file to a points shape file; extraction of the area points covering the water boundary of the Kingdom of Bahrain and multiplying the depth values by -1 to get the negative values. Then, the simple Kriging method was used in ArcMap environment to generate a new raster bathymetric grid surface of 30×30 m cells, which was the basis of the subsequent analysis. Finally, for validation purposes, 2200 bathymetric points were extracted from a medium scale nautical map (1:100 000) considering different depths over the Bahrain national water boundary. The nautical map was scanned, georeferenced and overlaid on the MBES-CARIS generated raster bathymetric grid surface (step 5 above), and then homologous depth points were selected. Statistical analysis, expressed as a linear error at the 95% confidence level, showed a strong correlation coefficient (R² = 0.96) and a low RMSE (± 0.57 m) between the nautical map and derived MBSE-CARIS depths if we consider only the shallow areas with depths of less than 10 m (about 800 validation points). When we consider only deeper areas (> 10 m) the correlation coefficient is equal to 0.73 and the RMSE is equal to ± 2.43 m while if we consider the totality of 2200 validation points including all depths, the correlation coefficient is still significant (R² = 0.81) with satisfactory RMSE (± 1.57 m). Certainly, this significant variation can be caused by the MBSE that did not completely cover the bottom in several of the deeper pockmarks because of the rapid change in depth. In addition, steep slopes and the rough seafloor probably affect the acquired MBSE raw data. In addition, the interpolation of missed area values between MBSE acquisition swaths-lines (ship-tracked sounding data) may not reflect the true depths of these missed areas. However, globally the results of the MBES-CARIS data are very appropriate for bathymetric mapping of shallow water areas.

Keywords: bathymetry mapping, multibeam echosounder systems, CARIS-HIPS, shallow water

Procedia PDF Downloads 366
655 Sustainable Harvesting, Conservation and Analysis of Genetic Diversity in Polygonatum Verticillatum Linn.

Authors: Anchal Rana

Abstract:

Indian Himalayas with their diverse climatic conditions are home to many rare and endangered medicinal flora. One such species is Polygonatum verticillatum Linn., popularly known as King Solomon’s Seal or Solomon’s Seal. Its mention as an incredible medicinal herb comes from 5000 years ago in Indian Materia Medica as a component of Ashtavarga, a poly-herbal formulation comprising of eight herbs illustrated as world’s first ever revitalizing and rejuvenating nutraceutical food, which is now commercialised in the name ‘Chaywanprash’. It is an erect tall (60 to 120 cm) perennial herb with sessile, linear leaves and white pendulous flowers. The species grows well in an altitude range of 1600 to 3600 m amsl, and propagates mostly through rhizomes. The rhizomes are potential source for significant phytochemicals like flavonoids, phenolics, lectins, terpenoids, allantoin, diosgenin, β-Sitosterol and quinine. The presence of such phytochemicals makes the species an asset for antioxidant, cardiotonic, demulcent, diuretic, energizer, emollient, aphrodisiac, appetizer, glactagogue, etc. properties. Having profound concentrations of macro and micronutrients, species has fine prospects of being used as a diet supplement. However, due to unscientific and gregarious uprooting, it has been assigned a status of ‘vulnerable’ and ‘endangered’ in the Conservation Assessment and Management Plan (CAMP) process conducted by Foundation for Revitalisation of Local Health Traditions (FRLHT) during 2010, according to IUCN Red-List Criteria. Further, destructive harvesting, land use disturbances, heavy livestock grazing, climatic changes and habitat fragmentation have substantially contributed towards anomaly of the species. It, therefore, became imperative to conserve the diversity of the species and make judicious use in future research and commercial programme and schemes. A Gene Bank was therefore established at High Altitude Herbal Garden of the Forest Research Institute, Dehradun, India situated at Chakarata (30042’52.99’’N, 77051’36.77’’E, 2205 m amsl) consisting 149 accessions collected from thirty-one geographical locations spread over three Himalayan States of Jammu and Kashmir, Himachal Pradesh, and Uttarakhand. The present investigations purport towards sampling and collection of divergent germplasm followed by planting and cultivation techniques. The ultimate aim is thereby focussed on analysing genetic diversity of the species and capturing promising genotypes for carrying out further genetic improvement programme so to contribute towards sustainable development and healthcare.

Keywords: Polygonatum verticillatum Linn., phytochemicals, genetic diversity, conservation, gene bank

Procedia PDF Downloads 145
654 Harrison’s Stolen: Addressing Aboriginal and Indigenous Islanders Human Rights

Authors: M. Shukry

Abstract:

According to the United Nations Declaration of Human Rights in 1948, every human being is entitled to rights in life that should be respected by others and protected by the state and community. Such rights are inherent regardless of colour, ethnicity, gender, religion or otherwise, and it is expected that all humans alike have the right to live without discrimination of any sort. However, that has not been the case with Aborigines in Australia. Over a long period of time, the governments of the State and the Territories and the Australian Commonwealth denied the Aboriginal and Indigenous inhabitants of the Torres Strait Islands such rights. Past Australian governments set policies and laws that enabled them to forcefully remove Indigenous children from their parents, which resulted in creating lost generations living the trauma of the loss of cultural identity, alienation and even their own selfhood. Intending to reduce that population of natives and their Aboriginal culture while, on the other hand, assimilate them into mainstream society, they gave themselves the right to remove them from their families with no hope of return. That practice has led to tragic consequences due to the trauma that has affected those children, an experience that is depicted by Jane Harrison in her play Stolen. The drama is the outcome of a six-year project on lost children and which was first performed in 1997 in Melbourne. Five actors only appear on the stage, playing the role of all the different characters, whether the main protagonists or the remaining cast, present or non-present ones as voices. The play outlines the life of five children who have been taken from their parents at an early age, entailing a disastrous negative impact that differs from one to the other. Unknown to each other, what connects between them is being put in a children’s home. The purpose of this paper is to analyse the play’s text in light of the 1948 Declaration of Human Rights, using it as a lens that reflects the atrocities practiced against the Aborigines. It highlights how such practices formed an outrageous violation of those natives’ rights as human beings. Harrison’s dramatic technique in conveying the children’s experiences is through a non-linear structure, fluctuating between past and present that are linked together within each of the five characters, reflecting their suffering and pain to create an emotional link between them and the audience. Her dramatic handling of the issue by fusing tragedy with humour as well as symbolism is a successful technique in revealing the traumatic memory of those children and their present life. The play has made a difference in commencing to address the problem of the right of all children to be with their families, which renders the real meaning of having a home and an identity as people.

Keywords: aboriginal, audience, Australia, children, culture, drama, home, human rights, identity, Indigenous, Jane Harrison, memory, scenic effects, setting, stage, stage directions, Stolen, trauma

Procedia PDF Downloads 284
653 Mature Field Rejuvenation Using Hydraulic Fracturing: A Case Study of Tight Mature Oilfield with Reveal Simulator

Authors: Amir Gharavi, Mohamed Hassan, Amjad Shah

Abstract:

The main characteristics of unconventional reservoirs include low-to ultra low permeability and low-to-moderate porosity. As a result, hydrocarbon production from these reservoirs requires different extraction technologies than from conventional resources. An unconventional reservoir must be stimulated to produce hydrocarbons at an acceptable flow rate to recover commercial quantities of hydrocarbons. Permeability for unconventional reservoirs is mostly below 0.1 mD, and reservoirs with permeability above 0.1 mD are generally considered to be conventional. The hydrocarbon held in these formations naturally will not move towards producing wells at economic rates without aid from hydraulic fracturing which is the only technique to assess these tight reservoir productions. Horizontal well with multi-stage fracking is the key technique to maximize stimulated reservoir volume and achieve commercial production. The main objective of this research paper is to investigate development options for a tight mature oilfield. This includes multistage hydraulic fracturing and spacing by building of reservoir models in the Reveal simulator to model potential development options based on sidetracking the existing vertical well. To simulate potential options, reservoir models have been built in the Reveal. An existing Petrel geological model was used to build the static parts of these models. A FBHP limit of 40bars was assumed to take into account pump operating limits and to maintain the reservoir pressure above the bubble point. 300m, 600m and 900m lateral length wells were modelled, in conjunction with 4, 6 and 8 stages of fracs. Simulation results indicate that higher initial recoveries and peak oil rates are obtained with longer well lengths and also with more fracs and spacing. For a 25year forecast, the ultimate recovery ranging from 0.4% to 2.56% for 300m and 1000m laterals respectively. The 900m lateral with 8 fracs 100m spacing gave the highest peak rate of 120m3/day, with the 600m and 300m cases giving initial peak rates of 110m3/day. Similarly, recovery factor for the 900m lateral with 8 fracs and 100m spacing was the highest at 2.65% after 25 years. The corresponding values for the 300m and 600m laterals were 2.37% and 2.42%. Therefore, the study suggests that longer laterals with 8 fracs and 100m spacing provided the optimal recovery, and this design is recommended as the basis for further study.

Keywords: unconventional, resource, hydraulic, fracturing

Procedia PDF Downloads 283
652 Development of a Risk Governance Index and Examination of Its Determinants: An Empirical Study in Indian Context

Authors: M. V. Shivaani, P. K. Jain, Surendra S. Yadav

Abstract:

Risk management has been gaining extensive focus from international organizations like Committee of Sponsoring Organizations and Financial Stability Board, and, the foundation of such an effective and efficient risk management system lies in a strong risk governance structure. In view of this, an attempt (perhaps a first of its kind) has been made to develop a risk governance index, which could be used as proxy for quality of risk governance structures. The index (normative framework) is based on eleven variables, namely, size of board, board diversity in terms of gender, proportion of executive directors, executive/non-executive status of chairperson, proportion of independent directors, CEO duality, chief risk officer (CRO), risk management committee, mandatory committees, voluntary committees and existence/non-existence of whistle blower policy. These variables are scored on a scale of 1 to 5 with the exception of the variables, namely, status of chairperson and CEO duality (which have been scored on a dichotomous scale with the score of 3 or 5). In case there is a legal/statutory requirement in respect of above-mentioned variables and there is a non-compliance with such requirement a score of one has been envisaged. Though there is no legal requirement, for the larger part of study, in context of CRO, risk management committee and whistle blower policy, still a score of 1 has been assigned in the event of their non-existence. Recognizing the importance of these variables in context of risk governance structure and the fact that the study basically focuses on risk governance, the absence of these variables has been equated to non-compliance with a legal/statutory requirement. Therefore, based on this the minimum score is 15 and the maximum possible is 55. In addition, an attempt has been made to explore the determinants of this index. For this purpose, the sample consists of non-financial companies (429) that constitute S&P CNX500 index. The study covers a 10 years period from April 1, 2005 to March 31, 2015. Given the panel nature of data, Hausman test was applied, and it suggested that fixed effects regression would be appropriate. The results indicate that age and size of firms have significant positive impact on its risk governance structures. Further, post-recession period (2009-2015) has witnessed significant improvement in quality of governance structures. In contrast, profitability (positive relationship), leverage (negative relationship) and growth (negative relationship) do not have significant impact on quality of risk governance structures. The value of rho indicates that about 77.74% variation in risk governance structures is due to firm specific factors. Given the fact that each firm is unique in terms of its risk exposure, risk culture, risk appetite, and risk tolerance levels, it appears reasonable to assume that the specific conditions and circumstances that a company is beset with, could be the biggest determinants of its risk governance structures. Given the recommendations put forth in the paper (particularly for regulators and companies), the study is expected to be of immense utility in an important yet neglected aspect of risk management.

Keywords: corporate governance, ERM, risk governance, risk management

Procedia PDF Downloads 238
651 Determination of Friction and Damping Coefficients of Folded Cover Mechanism Deployed by Torsion Springs

Authors: I. Yilmaz, O. Taga, F. Kosar, O. Keles

Abstract:

In this study, friction and damping coefficients of folded cover mechanism were obtained in accordance with experimental studies and data. Friction and damping coefficients are the most important inputs to accomplish a mechanism analysis. Friction and damping are two objects that change the time of deployment of mechanisms and their dynamic behaviors. Though recommended friction coefficient values exist in literature, damping is differentiating feature according to mechanic systems. So the damping coefficient should be obtained from mechanism test outputs. In this study, the folded cover mechanism use torsion springs for deploying covers that are formerly close folded position. Torsion springs provide folded covers with desirable deploying time according to variable environmental conditions. To verify all design revisions with system tests will be so costly so that some decisions are taken in accordance with numerical methods. In this study, there are two folded covers required to deploy simultaneously. Scotch-yoke and crank-rod mechanisms were combined to deploy folded covers simultaneously. The mechanism was unlocked with a pyrotechnic bolt onto scotch-yoke disc. When pyrotechnic bolt was exploded, torsion springs provided rotational movement for mechanism. Quick motion camera was recording dynamic behaviors of system during deployment case. Dynamic model of mechanism was modeled as rigid body with Adams MBD (multi body dynamics) then torque values provided by torsion springs were used as an input. A well-advised range of friction and damping coefficients were defined in Adams DOE (design of experiment) then a large number of analyses were performed until deployment time of folded covers run in with test data observed in record of quick motion camera, thus the deployment time of mechanism and dynamic behaviors were obtained. Same mechanism was tested with different torsion springs and torque values then outputs were compared with numerical models. According to comparison, it was understood that friction and damping coefficients obtained in this study can be used safely when studying on folded objects required to deploy simultaneously. In addition to model generated with Adams as rigid body the finite element model of folded mechanism was generated with Abaqus then the outputs of rigid body model and finite element model was compared. Finally, the reasonable solutions were suggested about different outputs of these solution methods.

Keywords: damping, friction, pyro-technic, scotch-yoke

Procedia PDF Downloads 310
650 Modeling Driving Distraction Considering Psychological-Physical Constraints

Authors: Yixin Zhu, Lishengsa Yue, Jian Sun, Lanyue Tang

Abstract:

Modeling driving distraction in microscopic traffic simulation is crucial for enhancing simulation accuracy. Current driving distraction models are mainly derived from physical motion constraints under distracted states, in which distraction-related error terms are added to existing microscopic driver models. However, the model accuracy is not very satisfying, due to a lack of modeling the cognitive mechanism underlying the distraction. This study models driving distraction based on the Queueing Network Human Processor model (QN-MHP). This study utilizes the queuing structure of the model to perform task invocation and switching for distracted operation and control of the vehicle under driver distraction. Based on the assumption of the QN-MHP model about the cognitive sub-network, server F is a structural bottleneck. The latter information must wait for the previous information to leave server F before it can be processed in server F. Therefore, the waiting time for task switching needs to be calculated. Since the QN-MHP model has different information processing paths for auditory information and visual information, this study divides driving distraction into two types: auditory distraction and visual distraction. For visual distraction, both the visual distraction task and the driving task need to go through the visual perception sub-network, and the stimuli of the two are asynchronous, which is called stimulus on asynchrony (SOA), so when calculating the waiting time for switching tasks, it is necessary to consider it. In the case of auditory distraction, the auditory distraction task and the driving task do not need to compete for the server resources of the perceptual sub-network, and their stimuli can be synchronized without considering the time difference in receiving the stimuli. According to the Theory of Planned Behavior for drivers (TPB), this study uses risk entropy as the decision criterion for driver task switching. A logistic regression model is used with risk entropy as the independent variable to determine whether the driver performs a distraction task, to explain the relationship between perceived risk and distraction. Furthermore, to model a driver’s perception characteristics, a neurophysiological model of visual distraction tasks is incorporated into the QN-MHP, and executes the classical Intelligent Driver Model. The proposed driving distraction model integrates the psychological cognitive process of a driver with the physical motion characteristics, resulting in both high accuracy and interpretability. This paper uses 773 segments of distracted car-following in Shanghai Naturalistic Driving Study data (SH-NDS) to classify the patterns of distracted behavior on different road facilities and obtains three types of distraction patterns: numbness, delay, and aggressiveness. The model was calibrated and verified by simulation. The results indicate that the model can effectively simulate the distracted car-following behavior of different patterns on various roadway facilities, and its performance is better than the traditional IDM model with distraction-related error terms. The proposed model overcomes the limitations of physical-constraints-based models in replicating dangerous driving behaviors, and internal characteristics of an individual. Moreover, the model is demonstrated to effectively generate more dangerous distracted driving scenarios, which can be used to construct high-value automated driving test scenarios.

Keywords: computational cognitive model, driving distraction, microscopic traffic simulation, psychological-physical constraints

Procedia PDF Downloads 68
649 Anti-DNA Antibodies from Patients with Schizophrenia Hydrolyze DNA

Authors: Evgeny A. Ermakov, Lyudmila P. Smirnova, Valentina N. Buneva

Abstract:

Schizophrenia associated with dysregulation of neurotransmitter processes in the central nervous system and disturbances in the humoral immune system resulting in the formation of antibodies (Abs) to the various components of the nervous tissue. Abs to different neuronal receptors and DNA were detected in the blood of patients with schizophrenia. Abs hydrolyzing DNA were detected in pool of polyclonal autoantibodies in autoimmune and infectious diseases, such catalytic Abs were named abzymes. It is believed that DNA-hydrolyzing abzymes are cytotoxic, cause nuclear DNA fragmentation and induce cell death by apoptosis. Abzymes with DNAase activity are interesting because of the mechanism of formation and the possibility of use as diagnostic markers. Therefore, in our work we have set following goals: to determine the level anti-DNA Abs in the serum of patients with schizophrenia and to study DNA-hydrolyzing activity of IgG of patients with schizophrenia. Materials and methods: In our study there were included 41 patients with a verified diagnosis of paranoid or simple schizophrenia and 24 healthy donors. Electrophoretically and immunologically homogeneous IgGs were obtained by sequential affinity chromatography of the serum proteins on protein G-Sepharose and gel filtration. The levels of anti-DNA Abs were determined using ELISA. DNA-hydrolyzing activity was detected as the level of supercoiled pBluescript DNA transition in circular and linear forms, the hydrolysis products were analyzed by agarose electrophoresis followed by ethidium bromide stain. To correspond the registered catalytic activity directly to the antibodies we carried out a number of strict criteria: electrophoretic homogeneity of the antibodies, gel filtration (acid shock analysis) and in situ activity. Statistical analysis was performed in ‘Statistica 9.0’ using the non-parametric Mann-Whitney test. Results: The sera of approximately 30% of schizophrenia patients displayed a higher level of Abs interacting with single-stranded (ssDNA) and double-stranded DNA (dsDNA) compared with healthy donors. The average level of Abs interacting with ssDNA was only 1.1-fold lower than that for interacting with dsDNA. IgG of patient with schizophrenia were shown to possess DNA hydrolyzing activity. Using affinity chromatography, electrophoretic analysis of isolated IgG homogeneity, gel filtration in acid shock conditions and in situ DNAse activity analysis we proved that the observed activity is intrinsic property of studied antibodies. We have shown that the relative DNAase activity of IgG in patients with schizophrenia averaged 55.4±32.5%, IgG of healthy donors showed much lower activity (average of 9.1±6.5%). It should be noted that DNAase activity of IgG in patients with schizophrenia with a negative symptoms was significantly higher (73.3±23.8%), than in patients with positive symptoms (43.3±33.1%). Conclusion: Anti-DNA Abs of patients with schizophrenia not only bind DNA, but quite efficiently hydrolyze the substrate. The data show a correlation with the level of DNase activity and leading symptoms of patients with schizophrenia.

Keywords: anti-DNA antibodies, abzymes, DNA hydrolysis, schizophrenia

Procedia PDF Downloads 308
648 Characterization of the MOSkin Dosimeter for Accumulated Dose Assessment in Computed Tomography

Authors: Lenon M. Pereira, Helen J. Khoury, Marcos E. A. Andrade, Dean L. Cutajar, Vinicius S. M. Barros, Anatoly B. Rozenfeld

Abstract:

With the increase of beam widths and the advent of multiple-slice and helical scanners, concerns related to the current dose measurement protocols and instrumentation in computed tomography (CT) have arisen. The current methodology of dose evaluation, which is based on the measurement of the integral of a single slice dose profile using a 100 mm long cylinder ionization chamber (Ca,100 and CPPMA, 100), has been shown to be inadequate for wide beams as it does not collect enough of the scatter-tails to make an accurate measurement. In addition, a long ionization chamber does not offer a good representation of the dose profile when tube current modulation is used. An alternative approach has been suggested by translating smaller detectors through the beam plane and assessing the accumulated dose trough the integral of the dose profile, which can be done for any arbitrary length in phantoms or in the air. For this purpose, a MOSFET dosimeter of small dosimetric volume was used. One of its recently designed versions is known as the MOSkin, which is developed by the Centre for Medical Radiation Physics at the University of Wollongong, and measures the radiation dose at a water equivalent depth of 0.07 mm, allowing the evaluation of skin dose when placed at the surface, or internal point doses when placed within a phantom. Thus, the aim of this research was to characterize the response of the MOSkin dosimeter for X-ray CT beams and to evaluate its application for the accumulated dose assessment. Initially, tests using an industrial x-ray unit were carried out at the Laboratory of Ionization Radiation Metrology (LMRI) of Federal University of Pernambuco, in order to investigate the sensitivity, energy dependence, angular dependence, and reproducibility of the dose response for the device for the standard radiation qualities RQT 8, RQT 9 and RQT 10. Finally, the MOSkin was used for the accumulated dose evaluation of scans using a Philips Brilliance 6 CT unit, with comparisons made between the CPPMA,100 value assessed with a pencil ionization chamber (PTW Freiburg TW 30009). Both dosimeters were placed in the center of a PMMA head phantom (diameter of 16 cm) and exposed in the axial mode with collimation of 9 mm, 250 mAs and 120 kV. The results have shown that the MOSkin response was linear with doses in the CT range and reproducible (98.52%). The sensitivity for a single MOSkin in mV/cGy was as follows: 9.208, 7.691 and 6.723 for the RQT 8, RQT 9 and RQT 10 beams qualities respectively. The energy dependence varied up to a factor of ±1.19 among those energies and angular dependence was not greater than 7.78% within the angle range from 0 to 90 degrees. The accumulated dose and the CPMMA, 100 value were 3,97 and 3,79 cGy respectively, which were statistically equivalent within the 95% confidence level. The MOSkin was shown to be a good alternative for CT dose profile measurements and more than adequate to provide accumulated dose assessments for CT procedures.

Keywords: computed tomography dosimetry, MOSFET, MOSkin, semiconductor dosimetry

Procedia PDF Downloads 294
647 Molecular Characterization, Host Plant Resistance and Epidemiology of Bean Common Mosaic Virus Infecting Cowpea (Vigna unguiculata L. Walp)

Authors: N. Manjunatha, K. T. Rangswamy, N. Nagaraju, H. A. Prameela, P. Rudraswamy, M. Krishnareddy

Abstract:

The identification of virus in cowpea especially potyviruses is confusing. Even though there are several studies on viruses causing diseases in cowpea, difficult to distinguish based on symptoms and serological detection. The differentiation of potyviruses considering as a constraint, the present study is initiated for molecular characterization, host plant resistance and epidemiology of the BCMV infecting cowpea. The etiological agent causing cowpea mosaic was identified as Bean Common Mosaic Virus (BCMV) on the basis of RT-PCR and electron microscopy. An approximately 750bp PCR product corresponding to coat protein (CP) region of the virus and the presence of long flexuous filamentous particles measuring about 952 nm in size typical to genus potyvirus were observed under electron microscope. The characterized virus isolate genome had 10054 nucleotides, excluding the 3’ terminal poly (A) tail. Comparison of polyprotein of the virus with other potyviruses showed similar genome organization with 9 cleavage sites resulted in 10 functional proteins. The pairwise sequence comparison of individual genes, P1 showed most divergent, but CP gene was less divergent at nucleotide and amino acid level. A phylogenetic tree constructed based on multiple sequence alignments of the polyprotein nucleotide and amino acid sequences of cowpea BCMV and potyviruses showed virus is closely related to BCMV-HB. Whereas, Soybean variant of china (KJ807806) and NL1 isolate (AY112735) showed 93.8 % (5’UTR) and 94.9 % (3’UTR) homology respectively with other BCMV isolates. This virus transmitted to different leguminous plant species and produced systemic symptoms under greenhouse conditions. Out of 100 cowpea genotypes screened, three genotypes viz., IC 8966, V 5 and IC 202806 showed immune reaction in both field and greenhouse conditions. Single marker analysis (SMA) was revealed out of 4 SSR markers linked to BCMV resistance, M135 marker explains 28.2 % of phenotypic variation (R2) and Polymorphic information content (PIC) value of these markers was ranged from 0.23 to 0.37. The correlation and regression analysis showed rainfall, and minimum temperature had significant negative impact and strong relationship with aphid population, whereas weak correlation was observed with disease incidence. Path coefficient analysis revealed most of the weather parameters exerted their indirect contributions to the aphid population and disease incidence except minimum temperature. This study helps to identify specific gaps in knowledge for researchers who may wish to further analyse the science behind complex interactions between vector-virus and host in relation to the environment. The resistant genotypes identified are could be effectively used in resistance breeding programme.

Keywords: cowpea, epidemiology, genotypes, virus

Procedia PDF Downloads 213
646 Enhancing Athlete Training using Real Time Pose Estimation with Neural Networks

Authors: Jeh Patel, Chandrahas Paidi, Ahmed Hambaba

Abstract:

Traditional methods for analyzing athlete movement often lack the detail and immediacy required for optimal training. This project aims to address this limitation by developing a Real-time human pose estimation system specifically designed to enhance athlete training across various sports. This system leverages the power of convolutional neural networks (CNNs) to provide a comprehensive and immediate analysis of an athlete’s movement patterns during training sessions. The core architecture utilizes dilated convolutions to capture crucial long-range dependencies within video frames. Combining this with the robust encoder-decoder architecture to further refine pose estimation accuracy. This capability is essential for precise joint localization across the diverse range of athletic poses encountered in different sports. Furthermore, by quantifying movement efficiency, power output, and range of motion, the system provides data-driven insights that can be used to optimize training programs. Pose estimation data analysis can also be used to develop personalized training plans that target specific weaknesses identified in an athlete’s movement patterns. To overcome the limitations posed by outdoor environments, the project employs strategies such as multi-camera configurations or depth sensing techniques. These approaches can enhance pose estimation accuracy in challenging lighting and occlusion scenarios, where pose estimation accuracy in challenging lighting and occlusion scenarios. A dataset is collected From the labs of Martin Luther King at San Jose State University. The system is evaluated through a series of tests that measure its efficiency and accuracy in real-world scenarios. Results indicate a high level of precision in recognizing different poses, substantiating the potential of this technology in practical applications. Challenges such as enhancing the system’s ability to operate in varied environmental conditions and further expanding the dataset for training were identified and discussed. Future work will refine the model’s adaptability and incorporate haptic feedback to enhance the interactivity and richness of the user experience. This project demonstrates the feasibility of an advanced pose detection model and lays the groundwork for future innovations in assistive enhancement technologies.

Keywords: computer vision, deep learning, human pose estimation, U-NET, CNN

Procedia PDF Downloads 16