Search results for: in-phase component
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2578

Search results for: in-phase component

2158 The Contemporary Visual Spectacle: Critical Visual Literacy

Authors: Lai-Fen Yang

Abstract:

In this increasingly visual world, how can we best decipher and understand the many ways that our everyday lives are organized around looking practices and the many images we encounter each day? Indeed, how we interact with and interpret visual images is a basic component of human life. Today, however, we are living in one of the most artificial visual and image-saturated cultures in human history, which makes understanding the complex construction and multiple social functions of visual imagery more important than ever before. Themes regarding our experience of a visually pervasive mediated culture, here, termed visual spectacle.

Keywords: visual culture, contemporary, images, literacy

Procedia PDF Downloads 513
2157 Enhancing Project Management Performance in Prefabricated Building Construction under Uncertainty: A Comprehensive Approach

Authors: Niyongabo Elyse

Abstract:

Prefabricated building construction is a pioneering approach that combines design, production, and assembly to attain energy efficiency, environmental sustainability, and economic feasibility. Despite continuous development in the industry in China, the low technical maturity of standardized design, factory production, and construction assembly introduces uncertainties affecting prefabricated component production and on-site assembly processes. This research focuses on enhancing project management performance under uncertainty to help enterprises navigate these challenges and optimize project resources. The study introduces a perspective on how uncertain factors influence the implementation of prefabricated building construction projects. It proposes a theoretical model considering project process management ability, adaptability to uncertain environments, and collaboration ability of project participants. The impact of uncertain factors is demonstrated through case studies and quantitative analysis, revealing constraints on implementation time, cost, quality, and safety. To address uncertainties in prefabricated component production scheduling, a fuzzy model is presented, expressing processing times in interval values. The model utilizes a cooperative co-evolution evolution algorithm (CCEA) to optimize scheduling, demonstrated through a real case study showcasing reduced project duration and minimized effects of processing time disturbances. Additionally, the research addresses on-site assembly construction scheduling, considering the relationship between task processing times and assigned resources. A multi-objective model with fuzzy activity durations is proposed, employing a hybrid cooperative co-evolution evolution algorithm (HCCEA) to optimize project scheduling. Results from real case studies indicate improved project performance in terms of duration, cost, and resilience to processing time delays and resource changes. The study also introduces a multistage dynamic process control model, utilizing IoT technology for real-time monitoring during component production and construction assembly. This approach dynamically adjusts schedules when constraints arise, leading to enhanced project management performance, as demonstrated in a real prefabricated housing project. Key contributions include a fuzzy prefabricated components production scheduling model, a multi-objective multi-mode resource-constrained construction project scheduling model with fuzzy activity durations, a multi-stage dynamic process control model, and a cooperative co-evolution evolution algorithm. The integrated mathematical model addresses the complexity of prefabricated building construction project management, providing a theoretical foundation for practical decision-making in the field.

Keywords: prefabricated construction, project management performance, uncertainty, fuzzy scheduling

Procedia PDF Downloads 50
2156 Linking Enhanced Resting-State Brain Connectivity with the Benefit of Desirable Difficulty to Motor Learning: A Functional Magnetic Resonance Imaging Study

Authors: Chien-Ho Lin, Ho-Ching Yang, Barbara Knowlton, Shin-Leh Huang, Ming-Chang Chiang

Abstract:

Practicing motor tasks arranged in an interleaved order (interleaved practice, or IP) generally leads to better learning than practicing tasks in a repetitive order (repetitive practice, or RP), an example of how desirable difficulty during practice benefits learning. Greater difficulty during practice, e.g. IP, is associated with greater brain activity measured by higher blood-oxygen-level dependent (BOLD) signal in functional magnetic resonance imaging (fMRI) in the sensorimotor areas of the brain. In this study resting-state fMRI was applied to investigate whether increase in resting-state brain connectivity immediately after practice predicts the benefit of desirable difficulty to motor learning. 26 healthy adults (11M/15F, age = 23.3±1.3 years) practiced two sets of three sequences arranged in a repetitive or an interleaved order over 2 days, followed by a retention test on Day 5 to evaluate learning. On each practice day, fMRI data were acquired in a resting state after practice. The resting-state fMRI data was decomposed using a group-level spatial independent component analysis (ICA), yielding 9 independent components (IC) matched to the precuneus network, primary visual networks (two ICs, denoted by I and II respectively), sensorimotor networks (two ICs, denoted by I and II respectively), the right and the left frontoparietal networks, occipito-temporal network, and the frontal network. A weighted resting-state functional connectivity (wRSFC) was then defined to incorporate information from within- and between-network brain connectivity. The within-network functional connectivity between a voxel and an IC was gauged by a z-score derived from the Fisher transformation of the IC map. The between-network connectivity was derived from the cross-correlation of time courses across all possible pairs of ICs, leading to a symmetric nc x nc matrix of cross-correlation coefficients, denoted by C = (pᵢⱼ). Here pᵢⱼ is the extremum of cross-correlation between ICs i and j; nc = 9 is the number of ICs. This component-wise cross-correlation matrix C was then projected to the voxel space, with the weights for each voxel set to the z-score that represents the above within-network functional connectivity. The wRSFC map incorporates the global characteristics of brain networks measured by the between-network connectivity, and the spatial information contained in the IC maps measured by the within-network connectivity. Pearson correlation analysis revealed that greater IP-minus-RP difference in wRSFC was positively correlated with the RP-minus-IP difference in the response time on Day 5, particularly in brain regions crucial for motor learning, such as the right dorsolateral prefrontal cortex (DLPFC), and the right premotor and supplementary motor cortices. This indicates that enhanced resting brain connectivity during the early phase of memory consolidation is associated with enhanced learning following interleaved practice, and as such wRSFC could be applied as a biomarker that measures the beneficial effects of desirable difficulty on motor sequence learning.

Keywords: desirable difficulty, functional magnetic resonance imaging, independent component analysis, resting-state networks

Procedia PDF Downloads 203
2155 Liquefaction Potential Assessment Using Screw Driving Testing and Microtremor Data: A Case Study in the Philippines

Authors: Arturo Daag

Abstract:

The Philippine Institute of Volcanology and Seismology (PHIVOLCS) is enhancing its liquefaction hazard map towards a detailed probabilistic approach using SDS and geophysical data. Target sites for liquefaction assessment are public schools in Metro Manila. Since target sites are in highly urbanized-setting, the objective of the project is to conduct both non-destructive geotechnical studies using Screw Driving Testing (SDFS) combined with geophysical data such as refraction microtremor array (ReMi), 3 component microtremor Horizontal to Vertical Spectral Ratio (HVSR), and ground penetrating RADAR (GPR). Initial test data was conducted in liquefaction impacted areas from the Mw 6.1 earthquake in Central Luzon last April 22, 2019 Province of Pampanga. Numerous accounts of liquefaction events were documented areas underlain by quaternary alluvium and mostly covered by recent lahar deposits. SDS estimated values showed a good correlation to actual SPT values obtained from available borehole data. Thus, confirming that SDS can be an alternative tool for liquefaction assessment and more efficient in terms of cost and time compared to SPT and CPT. Conducting borehole may limit its access in highly urbanized areas. In order to extend or extrapolate the SPT borehole data, non-destructive geophysical equipment was used. A 3-component microtremor obtains a subsurface velocity model in 1-D seismic shear wave velocity of the upper 30 meters of the profile (Vs30). For the ReMi, 12 geophone array with 6 to 8-meter spacing surveys were conducted. Microtremor data were computed through the Factor of Safety, which is the quotient of Cyclic Resistance Ratio (CRR) and Cyclic Stress Ratio (CSR). Complementary GPR was used to study the subsurface structure and used to inferred subsurface structures and groundwater conditions.

Keywords: screw drive testing, microtremor, ground penetrating RADAR, liquefaction

Procedia PDF Downloads 201
2154 Sulfanilamide/Epoxy Resin and Its Application as Tackifier in Epoxy Adhesives

Authors: Oiane Ruiz de Azua, Salvador Borros, Nuria Agullo, Jordi Arbusa

Abstract:

Tackiness is described as the ability to spontaneously form a bond to another material under light pressures within a short application time. During the first few minutes of the adhesive's curing, it is necessary to have enough tack to keep the substrates together while cohesion is increasing within the adhesive. This property plays a key role in the manufacturing process of pieces. Epoxy adhesives, unlike other adhesives, usually present low tackiness before curing; however, there is very little literature about the use of tackifiers in epoxy adhesives, except for the high molecular weight epoxy additives. In the present work, a tetrafunctional epoxy resin based on Bisphenol-A and Sulfanilamide has been synthesized in order to be used as a tackifier. This additive offers improved specific adhesion to two-component (2K) epoxy adhesives. The dosage of the tackifier has to be done carefully not to alter the mechanical and rheological properties of the adhesive. The synthetized product has been analyzed by FTIR and ¹H-NMR analysis, and the effect of the addition of 1 wt % of the tackifier on rheological properties, viscoelastic behavior, and mechanical properties has been studied. On one hand, the addition of the product in the epoxy resin part showed a significant increase in tackiness regarding the neat epoxy resin. On the other hand, tackiness of the whole formulation was also increased. Curing time of the adhesive has not undergone any relevant changes with the tackifier addition. Regarding viscoelastic properties, Storage Modulus (G') and Loss Modulus (G'') remain also unchanged at ambient temperature. Probably, in case higher tackifier concentration would be added, differences in viscoelastic properties would be observed. The study of mechanical properties shows that hardness and tensile strength also keep their values unchanged regarding neat two component adhesive. In conclusion, the addition of 1 wt % of sulfanilamide/epoxy enhanced the tackiness of the epoxy resin part, improves tack without modifying significantly either the rheological, the mechanical, or the viscoelastic properties of the product. Thus, the sulfanilamide presented could be a good candidate to be used as an additive to the 2k epoxy formulation for the manufacturing process of pieces.

Keywords: epoxy adhesive, manufacturing process of pieces, sulfanilamide, tackifiers

Procedia PDF Downloads 183
2153 Resting-State Functional Connectivity Analysis Using an Independent Component Approach

Authors: Eric Jacob Bacon, Chaoyang Jin, Dianning He, Shuaishuai Hu, Lanbo Wang, Han Li, Shouliang Qi

Abstract:

Objective: Refractory epilepsy is a complicated type of epilepsy that can be difficult to diagnose. Recent technological advancements have made resting-state functional magnetic resonance (rsfMRI) a vital technique for studying brain activity. However, there is still much to learn about rsfMRI. Investigating rsfMRI connectivity may aid in the detection of abnormal activities. In this paper, we propose studying the functional connectivity of rsfMRI candidates to diagnose epilepsy. Methods: 45 rsfMRI candidates, comprising 26 with refractory epilepsy and 19 healthy controls, were enrolled in this study. A data-driven approach known as independent component analysis (ICA) was used to achieve our goal. First, rsfMRI data from both patients and healthy controls were analyzed using group ICA. The components that were obtained were then spatially sorted to find and select meaningful ones. A two-sample t-test was also used to identify abnormal networks in patients and healthy controls. Finally, based on the fractional amplitude of low-frequency fluctuations (fALFF), a chi-square statistic test was used to distinguish the network properties of the patient and healthy control groups. Results: The two-sample t-test analysis yielded abnormal in the default mode network, including the left superior temporal lobe and the left supramarginal. The right precuneus was found to be abnormal in the dorsal attention network. In addition, the frontal cortex showed an abnormal cluster in the medial temporal gyrus. In contrast, the temporal cortex showed an abnormal cluster in the right middle temporal gyrus and the right fronto-operculum gyrus. Finally, the chi-square statistic test was significant, producing a p-value of 0.001 for the analysis. Conclusion: This study offers evidence that investigating rsfMRI connectivity provides an excellent diagnosis option for refractory epilepsy.

Keywords: ICA, RSN, refractory epilepsy, rsfMRI

Procedia PDF Downloads 76
2152 Reliability and Maintainability Optimization for Aircraft’s Repairable Components Based on Cost Modeling Approach

Authors: Adel A. Ghobbar

Abstract:

The airline industry is continuously challenging how to safely increase the service life of the aircraft with limited maintenance budgets. Operators are looking for the most qualified maintenance providers of aircraft components, offering the finest customer service. Component owner and maintenance provider is offering an Abacus agreement (Aircraft Component Leasing) to increase the efficiency and productivity of the customer service. To increase the customer service, the current focus on No Fault Found (NFF) units must change into the focus on Early Failure (EF) units. Since the effect of EF units has a significant impact on customer satisfaction, this needs to increase the reliability of EF units at minimal cost, which leads to the goal of this paper. By identifying the reliability of early failure (EF) units with regards to No Fault Found (NFF) units, in particular, the root cause analysis with an integrated cost analysis of EF units with the use of a failure mode analysis tool and a cost model, there will be a set of EF maintenance improvements. The data used for the investigation of the EF units will be obtained from the Pentagon system, an Enterprise Resource Planning (ERP) system used by Fokker Services. The Pentagon system monitors components, which needs to be repaired from Fokker aircraft owners, Abacus exchange pool, and commercial customers. The data will be selected on several criteria’s: time span, failure rate, and cost driver. When the selected data has been acquired, the failure mode and root cause analysis of EF units are initiated. The failure analysis approach tool was implemented, resulting in the proposed failure solution of EF. This will lead to specific EF maintenance improvements, which can be set-up to decrease the EF units and, as a result of this, increasing the reliability. The investigated EFs, between the time period over ten years, showed to have a significant reliability impact of 32% on the total of 23339 unscheduled failures. Since the EFs encloses almost one-third of the entire population.

Keywords: supportability, no fault found, FMEA, early failure, availability, operational reliability, predictive model

Procedia PDF Downloads 127
2151 Determination of the Volatile Organic Compounds, Antioxidant and Antimicrobial Properties of Microwave-Assisted Green Extracted Ficus Carica Linn Leaves

Authors: Pelin Yilmaz, Gizemnur Yildiz Uysal, Elcin Demirhan, Belma Ozbek

Abstract:

The edible fig plant, Ficus carica Linn, belongs to the Moraceae family, and the leaves are mainly considered agricultural waste after harvesting. It has been demonstrated in the literature that fig leaves contain appealing properties such as high vitamins, fiber, amino acids, organic acids, and phenolic or flavonoid content. The extraction of these valuable products has gained importance. Microwave-assisted extraction (MAE) is a method using microwave energy to heat the solvents, thereby transferring the bioactive compounds from the sample to the solvent. The main advantage of the MAE is the rapid extraction of bioactive compounds. In the present study, the MAE was applied to extract the bioactive compounds from Ficus carica L. leaves, and the effect of microwave power (180-900 W), extraction time (60-180 s), and solvent to sample amount (mL/g) (10-30) on the antioxidant property of the leaves. Then, the volatile organic component profile was determined at the specified extraction point. Additionally, antimicrobial studies were carried out to determine the minimum inhibitory concentration of the microwave-extracted leaves. As a result, according to the data obtained from the experimental studies, the highest antimicrobial properties were obtained under the process parameters such as 540 W, 180 s, and 20 mL/g concentration. The volatile organic compound profile showed that isobergapten, which belongs to the furanocoumarins family exhibiting anticancer, antioxidant, and antimicrobial activity besides promoting bone health, was the main compound. Acknowledgments: This work has been supported by Yildiz Technical University Scientific Research Projects Coordination Unit under project number FBA-2021-4409. The authors would like to acknowledge the financial support from Tubitak 1515 - Frontier R&D Laboratory Support Programme.

Keywords: Ficus carica Linn leaves, volatile organic component, GC-MS, microwave extraction, isobergapten, antimicrobial

Procedia PDF Downloads 80
2150 Kernel-Based Double Nearest Proportion Feature Extraction for Hyperspectral Image Classification

Authors: Hung-Sheng Lin, Cheng-Hsuan Li

Abstract:

Over the past few years, kernel-based algorithms have been widely used to extend some linear feature extraction methods such as principal component analysis (PCA), linear discriminate analysis (LDA), and nonparametric weighted feature extraction (NWFE) to their nonlinear versions, kernel principal component analysis (KPCA), generalized discriminate analysis (GDA), and kernel nonparametric weighted feature extraction (KNWFE), respectively. These nonlinear feature extraction methods can detect nonlinear directions with the largest nonlinear variance or the largest class separability based on the given kernel function. Moreover, they have been applied to improve the target detection or the image classification of hyperspectral images. The double nearest proportion feature extraction (DNP) can effectively reduce the overlap effect and have good performance in hyperspectral image classification. The DNP structure is an extension of the k-nearest neighbor technique. For each sample, there are two corresponding nearest proportions of samples, the self-class nearest proportion and the other-class nearest proportion. The term “nearest proportion” used here consider both the local information and other more global information. With these settings, the effect of the overlap between the sample distributions can be reduced. Usually, the maximum likelihood estimator and the related unbiased estimator are not ideal estimators in high dimensional inference problems, particularly in small data-size situation. Hence, an improved estimator by shrinkage estimation (regularization) is proposed. Based on the DNP structure, LDA is included as a special case. In this paper, the kernel method is applied to extend DNP to kernel-based DNP (KDNP). In addition to the advantages of DNP, KDNP surpasses DNP in the experimental results. According to the experiments on the real hyperspectral image data sets, the classification performance of KDNP is better than that of PCA, LDA, NWFE, and their kernel versions, KPCA, GDA, and KNWFE.

Keywords: feature extraction, kernel method, double nearest proportion feature extraction, kernel double nearest feature extraction

Procedia PDF Downloads 344
2149 Verification and Validation of Simulated Process Models of KALBR-SIM Training Simulator

Authors: T. Jayanthi, K. Velusamy, H. Seetha, S. A. V. Satya Murty

Abstract:

Verification and Validation of Simulated Process Model is the most important phase of the simulator life cycle. Evaluation of simulated process models based on Verification and Validation techniques checks the closeness of each component model (in a simulated network) with the real system/process with respect to dynamic behaviour under steady state and transient conditions. The process of Verification and validation helps in qualifying the process simulator for the intended purpose whether it is for providing comprehensive training or design verification. In general, model verification is carried out by comparison of simulated component characteristics with the original requirement to ensure that each step in the model development process completely incorporates all the design requirements. Validation testing is performed by comparing the simulated process parameters to the actual plant process parameters either in standalone mode or integrated mode. A Full Scope Replica Operator Training Simulator for PFBR - Prototype Fast Breeder Reactor has been developed at IGCAR, Kalpakkam, INDIA named KALBR-SIM (Kalpakkam Breeder Reactor Simulator) wherein the main participants are engineers/experts belonging to Modeling Team, Process Design and Instrumentation and Control design team. This paper discusses the Verification and Validation process in general, the evaluation procedure adopted for PFBR operator training Simulator, the methodology followed for verifying the models, the reference documents and standards used etc. It details out the importance of internal validation by design experts, subsequent validation by external agency consisting of experts from various fields, model improvement by tuning based on expert’s comments, final qualification of the simulator for the intended purpose and the difficulties faced while co-coordinating various activities.

Keywords: Verification and Validation (V&V), Prototype Fast Breeder Reactor (PFBR), Kalpakkam Breeder Reactor Simulator (KALBR-SIM), steady state, transient state

Procedia PDF Downloads 266
2148 Urban Flood Risk Mapping–a Review

Authors: Sherly M. A., Subhankar Karmakar, Terence Chan, Christian Rau

Abstract:

Floods are one of the most frequent natural disasters, causing widespread devastation, economic damage and threat to human lives. Hydrologic impacts of climate change and intensification of urbanization are two root causes of increased flood occurrences, and recent research trends are oriented towards understanding these aspects. Due to rapid urbanization, population of cities across the world has increased exponentially leading to improperly planned developments. Climate change due to natural and anthropogenic activities on our environment has resulted in spatiotemporal changes in rainfall patterns. The combined effect of both aggravates the vulnerability of urban populations to floods. In this context, an efficient and effective flood risk management with its core component as flood risk mapping is essential in prevention and mitigation of flood disasters. Urban flood risk mapping involves zoning of an urban region based on its flood risk, which depicts the spatiotemporal pattern of frequency and severity of hazards, exposure to hazards, and degree of vulnerability of the population in terms of socio-economic, environmental and infrastructural aspects. Although vulnerability is a key component of risk, its assessment and mapping is often less advanced than hazard mapping and quantification. A synergic effort from technical experts and social scientists is vital for the effectiveness of flood risk management programs. Despite an increasing volume of quality research conducted on urban flood risk, a comprehensive multidisciplinary approach towards flood risk mapping still remains neglected due to which many of the input parameters and definitions of flood risk concepts are imprecise. Thus, the objectives of this review are to introduce and precisely define the relevant input parameters, concepts and terms in urban flood risk mapping, along with its methodology, current status and limitations. The review also aims at providing thought-provoking insights to potential future researchers and flood management professionals.

Keywords: flood risk, flood hazard, flood vulnerability, flood modeling, urban flooding, urban flood risk mapping

Procedia PDF Downloads 590
2147 Analysis of Biomarkers Intractable Epileptogenic Brain Networks with Independent Component Analysis and Deep Learning Algorithms: A Comprehensive Framework for Scalable Seizure Prediction with Unimodal Neuroimaging Data in Pediatric Patients

Authors: Bliss Singhal

Abstract:

Epilepsy is a prevalent neurological disorder affecting approximately 50 million individuals worldwide and 1.2 million Americans. There exist millions of pediatric patients with intractable epilepsy, a condition in which seizures fail to come under control. The occurrence of seizures can result in physical injury, disorientation, unconsciousness, and additional symptoms that could impede children's ability to participate in everyday tasks. Predicting seizures can help parents and healthcare providers take precautions, prevent risky situations, and mentally prepare children to minimize anxiety and nervousness associated with the uncertainty of a seizure. This research proposes a comprehensive framework to predict seizures in pediatric patients by evaluating machine learning algorithms on unimodal neuroimaging data consisting of electroencephalogram signals. The bandpass filtering and independent component analysis proved to be effective in reducing the noise and artifacts from the dataset. Various machine learning algorithms’ performance is evaluated on important metrics such as accuracy, precision, specificity, sensitivity, F1 score and MCC. The results show that the deep learning algorithms are more successful in predicting seizures than logistic Regression, and k nearest neighbors. The recurrent neural network (RNN) gave the highest precision and F1 Score, long short-term memory (LSTM) outperformed RNN in accuracy and convolutional neural network (CNN) resulted in the highest Specificity. This research has significant implications for healthcare providers in proactively managing seizure occurrence in pediatric patients, potentially transforming clinical practices, and improving pediatric care.

Keywords: intractable epilepsy, seizure, deep learning, prediction, electroencephalogram channels

Procedia PDF Downloads 84
2146 VaR Estimation Using the Informational Content of Futures Traded Volume

Authors: Amel Oueslati, Olfa Benouda

Abstract:

New Value at Risk (VaR) estimation is proposed and investigated. The well-known two stages Garch-EVT approach uses conditional volatility to generate one step ahead forecasts of VaR. With daily data for twelve stocks that decompose the Dow Jones Industrial Average (DJIA) index, this paper incorporates the volume in the first stage volatility estimation. Afterwards, the forecasting ability of this conditional volatility concerning the VaR estimation is compared to that of a basic volatility model without considering any trading component. The results are significant and bring out the importance of the trading volume in the VaR measure.

Keywords: Garch-EVT, value at risk, volume, volatility

Procedia PDF Downloads 285
2145 New Roles of Telomerase and Telomere-Associated Proteins in the Regulation of Telomere Length

Authors: Qin Yang, Fan Zhang, Juan Du, Chongkui Sun, Krishna Kota, Yun-Ling Zheng

Abstract:

Telomeres are specialized structures at chromosome ends consisting of tandem repetitive DNA sequences [(TTAGGG)n in humans] and associated proteins, which are necessary for telomere function. Telomere lengths are tightly regulated within a narrow range in normal human somatic cells, the basis of cellular senescence and aging. Previous studies have extensively focused on how short telomeres are extended and have demonstrated that telomerase plays a central role in telomere maintenance through elongating the short telomeres. However, the molecular mechanisms of regulating excessively long telomeres are unknown. Here, we found that telomerase enzymatic component hTERT plays a dual role in the regulation of telomeres length. We analyzed single telomere alterations at each chromosomal end led to the discoveries that hTERT shortens excessively long telomeres and elongates short telomeres simultaneously, thus maintaining the optimal telomere length at each chromosomal end for an efficient protection. The hTERT-mediated telomere shortening removes large segments of telomere DNA rapidly without inducing telomere dysfunction foci or affecting cell proliferation, thus it is mechanistically distinct from rapid telomere deletion. We found that expression of hTERT generates telomeric circular DNA, suggesting that telomere homologous recombination may be involved in this telomere shortening process. Moreover, the hTERT-mediated telomere shortening is required its enzymatic activity, but telomerase RNA component hTR is not involved in it. Furthermore, shelterin protein TPP1 interacts with hTERT and recruits it on telomeres to mediate telomere shortening. In addition, telomere-associated proteins, DKC1 and TCAB1 also play roles in this process. This novel hTERT-mediated telomere shortening mechanism not only exists in cancer cells, but also in primary human cells. Thus, the hTERT-mediated telomere shortening is expected to shift the paradigm on current molecular models of telomere length maintenance, with wide-reaching consequences in cancer and aging fields.

Keywords: aging, hTERT, telomerase, telomeres, human cells

Procedia PDF Downloads 427
2144 Modal Approach for Decoupling Damage Cost Dependencies in Building Stories

Authors: Haj Najafi Leila, Tehranizadeh Mohsen

Abstract:

Dependencies between diverse factors involved in probabilistic seismic loss evaluation are recognized to be an imperative issue in acquiring accurate loss estimates. Dependencies among component damage costs could be taken into account considering two partial distinct states of independent or perfectly-dependent for component damage states; however, in our best knowledge, there is no available procedure to take account of loss dependencies in story level. This paper attempts to present a method called "modal cost superposition method" for decoupling story damage costs subjected to earthquake ground motions dealt with closed form differential equations between damage cost and engineering demand parameters which should be solved in complex system considering all stories' cost equations by the means of the introduced "substituted matrixes of mass and stiffness". Costs are treated as probabilistic variables with definite statistic factors of median and standard deviation amounts and a presumed probability distribution. To supplement the proposed procedure and also to display straightforwardness of its application, one benchmark study has been conducted. Acceptable compatibility has been proven for the estimated damage costs evaluated by the new proposed modal and also frequently used stochastic approaches for entire building; however, in story level, insufficiency of employing modification factor for incorporating occurrence probability dependencies between stories has been revealed due to discrepant amounts of dependency between damage costs of different stories. Also, more dependency contribution in occurrence probability of loss could be concluded regarding more compatibility of loss results in higher stories than the lower ones, whereas reduction in incorporation portion of cost modes provides acceptable level of accuracy and gets away from time consuming calculations including some limited number of cost modes in high mode situation.

Keywords: dependency, story-cost, cost modes, engineering demand parameter

Procedia PDF Downloads 180
2143 Multidimensional Poverty and Its Correlates among Rural Households in Limpopo Province, South Africa

Authors: Tamunotonye Mayowa Braide, Isaac Oluwatayo

Abstract:

This study investigates multidimensional poverty, and its correlates among rural households in Sekhukhune and Capricorn District municipalities (SDM & CDM) in Limpopo Province, South Africa. Primary data were collected from 407 rural households selected through purposive and simple random sampling techniques. Analytical techniques employed include descriptive statistics, principal component analysis (PCA), and the Alkire Foster (A-F) methodology. The results of the descriptive statistics showed there are more females (66%) than males (34%) in rural areas of Limpopo Province, with about 45% of them having secondary school education as the highest educational level attained and only about 3% do not have formal education. In the analysis of deprivation, eight dimensions of deprivation, constructed from 21 variables, were identified using the PCA. These dimensions include type and condition of dwelling water and sanitation, educational attainment and income, type of fuel for cooking and heating, access to clothing and cell phone, assets and fuel for light, health condition, crowding, and child health. In identifying the poor with poverty cut-off (0.13) of all indicators, about 75.9% of the rural households are deprived in 25% of the total dimensions, with the adjusted headcount ratio (M0) being 0.19. Multidimensional poverty estimates showed higher estimates of poor rural households with 71%, compared to 29%, which fall below the income poverty line. The study conducted poverty decomposition, using sub-groups within the area by examining regions and household characteristics. In SDM, there are more multidimensionally poor households than in CDM. The water and sanitation dimension is the largest contributor to the multidimensional poverty index (MPI) in rural areas of Limpopo Province. The findings can, therefore, assist in better design of welfare policy and target poverty alleviation programs and as well help in efficient resource allocation at the provincial and local municipality levels.

Keywords: Alkire-Foster methodology, Limpopo province, multidimensional poverty, principal component analysis, South Africa

Procedia PDF Downloads 164
2142 Network Analysis and Sex Prediction based on a full Human Brain Connectome

Authors: Oleg Vlasovets, Fabian Schaipp, Christian L. Mueller

Abstract:

we conduct a network analysis and predict the sex of 1000 participants based on ”connectome” - pairwise Pearson’s correlation across 436 brain parcels. We solve the non-smooth convex optimization problem, known under the name of Graphical Lasso, where the solution includes a low-rank component. With this solution and machine learning model for a sex prediction, we explain the brain parcels-sex connectivity patterns.

Keywords: network analysis, neuroscience, machine learning, optimization

Procedia PDF Downloads 147
2141 Cost Effective and Efficient Feeding: A Way Forward for Sustainable and Profitable Aquaculture

Authors: Pawan Kumar Sharma, J. Stephan Sampath Kumar, S. Anand, Chandana B. L.

Abstract:

Protein is the major component for the success in culture of shrimp and fishes. Apparently, excess dietary protein is undesirable, as it not only enhances the production cost but also leads to water quality deterioration. A field survey was conducted with aqua farmers of Kerala, India, a leading state in coastal aquaculture, to assess the role of protein component in feed that can be efficiently and effectively managed for sustainable aquaculture. The study showed an average feed amount of 13.55 ± 2.16 tonnes per hectare was being used by the farmers of Kerala. The average feed cost percentage of Rs. 57.76 ± 13.46 /kg was invested for an average protein level of 36.26 % ± 0.082 in the feed and Rs.78.95 ± 3.086 per kilogram of feed was being paid by the farmers. Study revealed that replacement of fish meal and fish oil within shrimp aquafeeds with alternative protein, and lipid sources can only be achieved if changes are made in the basic shrimp culturing practices, such as closed farming system through water recycling or zero-water exchange, and by maximizing in-situ, floc and natural food production within the culture system. The upshot of such production systems is that imports of high-quality feed ingredients and aqua feeds can eventually be eliminated, and the utilization of locally available feed ingredients from agricultural by-products can be greatly improved and maximized. The promotion of closed shrimp production systems would also greatly reduce water use and increase shrimp production per unit area but would necessitate the continuous provision of electricity for aeration during production. Alternative energy sources such as solar power might be used, and resource poor farming communities should also explore wind energy for use. The study concluded that farm made feed and closed farming systems are essential for the sustainability and profitability of the aquaculture industry.

Keywords: aqua feeds, floc, fish meal, protein, zero-water exchange

Procedia PDF Downloads 145
2140 Contribution to the Understanding of the Hydrodynamic Behaviour of Aquifers of the Taoudéni Sedimentary Basin (South-eastern Part, Burkina Faso)

Authors: Kutangila Malundama Succes, Koita Mahamadou

Abstract:

In the context of climate change and demographic pressure, groundwater has emerged as an essential and strategic resource whose sustainability relies on good management. The accuracy and relevance of decisions made in managing these resources depend on the availability and quality of scientific information they must rely on. It is, therefore, more urgent to improve the state of knowledge on groundwater to ensure sustainable management. This study is conducted for the particular case of the aquifers of the transboundary sedimentary basin of Taoudéni in its Burkinabe part. Indeed, Burkina Faso (and the Sahel region in general), marked by low rainfall, has experienced episodes of severe drought, which have justified the use of groundwater as the primary source of water supply. This study aims to improve knowledge of the hydrogeology of this area to achieve sustainable management of transboundary groundwater resources. The methodological approach first described lithological units regarding the extension and succession of different layers. Secondly, the hydrodynamic behavior of these units was studied through the analysis of spatio-temporal variations of piezometric. The data consists of 692 static level measurement points and 8 observation wells located in the usual manner in the area and capturing five of the identified geological formations. Monthly piezometric level chronicles are available for each observation and cover the period from 1989 to 2020. The temporal analysis of piezometric, carried out in comparison with rainfall chronicles, revealed a general upward trend in piezometric levels throughout the basin. The reaction of the groundwater generally occurs with a delay of 1 to 2 months relative to the flow of the rainy season. Indeed, the peaks of the piezometric level generally occur between September and October in reaction to the rainfall peaks between July and August. Low groundwater levels are observed between May and July. This relatively slow reaction of the aquifer is observed in all wells. The influence of the geological nature through the structure and hydrodynamic properties of the layers was deduced. The spatial analysis reveals that piezometric contours vary between 166 and 633 m with a trend indicating flow that generally goes from southwest to northeast, with the feeding areas located towards the southwest and northwest. There is a quasi-concordance between the hydrogeological basins and the overlying hydrological basins, as well as a bimodal flow with a component following the topography and another significant component deeper, controlled by the regional gradient SW-NE. This latter component may present flows directed from the high reliefs towards the sources of Nasso. In the source area (Kou basin), the maximum average stock variation, calculated by the Water Table Fluctuation (WTF) method, varies between 35 and 48.70 mm per year for 2012-2014.

Keywords: hydrodynamic behaviour, taoudeni basin, piezometry, water table fluctuation

Procedia PDF Downloads 65
2139 Investigating Software Engineering Challenges in Game Development

Authors: Fawad Zaidi

Abstract:

This paper discusses a variety of challenges and solutions involved with creating computer games and the issues faced by the software engineers working in this field. This review further investigates the articles coverage of project scope and the problem of feature creep that appears to be inherent with game development. The paper tries to answer the following question: Is this a problem caused by a shortage, or bad software engineering practices, or is this outside the control of the software engineering component of the game production process?

Keywords: software engineering, computer games, software applications, development

Procedia PDF Downloads 475
2138 Directivity in the Dramatherapeutic Process for People with Addictive Behaviour

Authors: Jakub Vávra, Milan Valenta, Petr Kosek

Abstract:

This article presents a perspective on the conduct of the dramatherapy process with persons with addictive behaviours with regard to the directiveness of the process. Although drama therapy as one of the creative arts approaches is rather non-directive in nature, depending on the clientele, there may be a need to structure the process more and, depending on the needs of the clients, to guide the process more directive. The specificity for people with addictive behaviours is discussed through the prism of the dramatherapeutic perspective, where we can find both a psychotherapeutic component as well as a component touching on expression and art, which is rather non-directive in nature. Within the context of practice with clients, this theme has repeatedly emerged and dramatherapists themselves have sought to find ways of coping with clients' demands and needs for structure and guidance within the dramatherapy process. Some of the outcomes from the supervision work also guided the research. Based on this insight, the research questions were approached. The first research question asks: in what ways is directive in dramatherapy manifested and manifested in the process? The second research question then complements the first and asks: to which phenomena are directivity in dramatherapy linked? In relation to the research questions, data were collected using focus groups and field notes. The qualitative approach of Content analysis and Relational analysis was chosen as the methodology. For analyzing qualitative research, we chose an Inductive coding scheme: Open coding, Axial coding, Pattern matching, Member checking, and Creating a coding scheme. In the presented partial research results, we find recurrent schemes related to directive coding in drama therapy. As an important element, directive leadership emerges in connection with safety for the client group, then in connection with the clients' order and also the department of the facility, and last but not least, to the personality of the drama therapist. By careful analysis and looking for patterns in the research results, we can see connections that are impossible to interpret at this stage but already provide clues to our understanding of the topic and open up further avenues for research in this area.

Keywords: dramatherapy, directivity, personal approach, aims of dramatherapy process, safetyness

Procedia PDF Downloads 69
2137 Multimodal Integration of EEG, fMRI and Positron Emission Tomography Data Using Principal Component Analysis for Prognosis in Coma Patients

Authors: Denis Jordan, Daniel Golkowski, Mathias Lukas, Katharina Merz, Caroline Mlynarcik, Max Maurer, Valentin Riedl, Stefan Foerster, Eberhard F. Kochs, Andreas Bender, Ruediger Ilg

Abstract:

Introduction: So far, clinical assessments that rely on behavioral responses to differentiate coma states or even predict outcome in coma patients are unreliable, e.g. because of some patients’ motor disabilities. The present study was aimed to provide prognosis in coma patients using markers from electroencephalogram (EEG), blood oxygen level dependent (BOLD) functional magnetic resonance imaging (fMRI) and [18F]-fluorodeoxyglucose (FDG) positron emission tomography (PET). Unsuperwised principal component analysis (PCA) was used for multimodal integration of markers. Methods: Approved by the local ethics committee of the Technical University of Munich (Germany) 20 patients (aged 18-89) with severe brain damage were acquired through intensive care units at the Klinikum rechts der Isar in Munich and at the Therapiezentrum Burgau (Germany). At the day of EEG/fMRI/PET measurement (date I) patients (<3.5 month in coma) were grouped in the minimal conscious state (MCS) or vegetative state (VS) on the basis of their clinical presentation (coma recovery scale-revised, CRS-R). Follow-up assessment (date II) was also based on CRS-R in a period of 8 to 24 month after date I. At date I, 63 channel EEG (Brain Products, Gilching, Germany) was recorded outside the scanner, and subsequently simultaneous FDG-PET/fMRI was acquired on an integrated Siemens Biograph mMR 3T scanner (Siemens Healthineers, Erlangen Germany). Power spectral densities, permutation entropy (PE) and symbolic transfer entropy (STE) were calculated in/between frontal, temporal, parietal and occipital EEG channels. PE and STE are based on symbolic time series analysis and were already introduced as robust markers separating wakefulness from unconsciousness in EEG during general anesthesia. While PE quantifies the regularity structure of the neighboring order of signal values (a surrogate of cortical information processing), STE reflects information transfer between two signals (a surrogate of directed connectivity in cortical networks). fMRI was carried out using SPM12 (Wellcome Trust Center for Neuroimaging, University of London, UK). Functional images were realigned, segmented, normalized and smoothed. PET was acquired for 45 minutes in list-mode. For absolute quantification of brain’s glucose consumption rate in FDG-PET, kinetic modelling was performed with Patlak’s plot method. BOLD signal intensity in fMRI and glucose uptake in PET was calculated in 8 distinct cortical areas. PCA was performed over all markers from EEG/fMRI/PET. Prognosis (persistent VS and deceased patients vs. recovery to MCS/awake from date I to date II) was evaluated using the area under the curve (AUC) including bootstrap confidence intervals (CI, *: p<0.05). Results: Prognosis was reliably indicated by the first component of PCA (AUC=0.99*, CI=0.92-1.00) showing a higher AUC when compared to the best single markers (EEG: AUC<0.96*, fMRI: AUC<0.86*, PET: AUC<0.60). CRS-R did not show prediction (AUC=0.51, CI=0.29-0.78). Conclusion: In a multimodal analysis of EEG/fMRI/PET in coma patients, PCA lead to a reliable prognosis. The impact of this result is evident, as clinical estimates of prognosis are inapt at time and could be supported by quantitative biomarkers from EEG, fMRI and PET. Due to the small sample size, further investigations are required, in particular allowing superwised learning instead of the basic approach of unsuperwised PCA.

Keywords: coma states and prognosis, electroencephalogram, entropy, functional magnetic resonance imaging, machine learning, positron emission tomography, principal component analysis

Procedia PDF Downloads 339
2136 Modelling Retirement Outcomes: An Australian Case Study

Authors: Colin O’Hare, Zili Zho, Thomas Sneddon

Abstract:

The Australian superannuation system has received high praise for its participation rates and level of funding in retirement yet it is only 25 years old. In recent years, with increasing longevity and persistent lower rates of investment return, how adequate will the funds accumulated through a superannuation system be? In this paper we take Australia as a case study and build a stochastic model of accumulation and decummulation of funds and determine the expected number of years a fund may last an individual in retirement.

Keywords: component, mortality, stochastic models, superannuation

Procedia PDF Downloads 245
2135 Perceptual Image Coding by Exploiting Internal Generative Mechanism

Authors: Kuo-Cheng Liu

Abstract:

In the perceptual image coding, the objective is to shape the coding distortion such that the amplitude of distortion does not exceed the error visibility threshold, or to remove perceptually redundant signals from the image. While most researches focus on color image coding, the perceptual-based quantizer developed for luminance signals are always directly applied to chrominance signals such that the color image compression methods are inefficient. In this paper, the internal generative mechanism is integrated into the design of a color image compression method. The internal generative mechanism working model based on the structure-based spatial masking is used to assess the subjective distortion visibility thresholds that are visually consistent to human eyes better. The estimation method of structure-based distortion visibility thresholds for color components is further presented in a locally adaptive way to design quantization process in the wavelet color image compression scheme. Since the lowest subband coefficient matrix of images in the wavelet domain preserves the local property of images in the spatial domain, the error visibility threshold inherent in each coefficient of the lowest subband for each color component is estimated by using the proposed spatial error visibility threshold assessment. The threshold inherent in each coefficient of other subbands for each color component is then estimated in a local adaptive fashion based on the distortion energy allocation. By considering that the error visibility thresholds are estimated using predicting and reconstructed signals of the color image, the coding scheme incorporated with locally adaptive perceptual color quantizer does not require side information. Experimental results show that the entropies of three color components obtained by using proposed IGM-based color image compression scheme are lower than that obtained by using the existing color image compression method at perceptually lossless visual quality.

Keywords: internal generative mechanism, structure-based spatial masking, visibility threshold, wavelet domain

Procedia PDF Downloads 248
2134 Alumina Nanoparticles in One-Pot Synthesis of Pyrazolopyranopyrimidinones

Authors: Saeed Khodabakhshi, Alimorad Rashidi, Ziba Tavakoli, Sajad Kiani, Sadegh Dastkhoon

Abstract:

Alumina nanoparticles (γ-Al2O3 NPs) were prepared via a new and simple synthetic route and characterized by field emission scanning electron microscope, X-ray diffraction, and Fourier transform infrared spectroscopy. The catalytic activity of prepared γ-Al2O3 NPs was investigated for the one-pot, four-component synthesis of fused tri-heterocyclic compounds containing pyrazole, pyran, and pyrimidine. This procedure has some advantages such as high efficiency, simplicity, high rate and environmental safety.

Keywords: alumina nanoparticles, one-pot, fused tri-heterocyclic compounds, pyran

Procedia PDF Downloads 332
2133 The Modulation of Health and Inflammatory Status in Young Pigs by Grape Waste Enriched in Polyphenols

Authors: Gina Cecilia Pistol, Loredana Calin, Mariana Stancu, Veronica Chedea, Ionelia Taranu

Abstract:

Inflammatory-associated diseases have an increased trend in the past decades. The pharmacological strategies aimed to treat these inflammatory diseases are very expensive and with non-beneficial results. The current trend is to find alternative strategies to counteract or to control inflammatory component of diseases. The grape by-products either seeds or pomace are rich in bioactive compounds (e.g. polyphenols) which may be beneficial in prevention of inflammation associated with cancer progression and other pathologies with inflammatory component. The in vivo models are very useful for studying the immune and inflammatory status. The domestic pig (Sus scrofa domesticus) is related to human from anatomic and physiologic point of view, representing a feasible model for studying the human inflammatory pathologies. Starting from these data, we evaluated the effect of a diet containing 5% grape seed cakes (GS) on piglets blood biochemical parameters and immune pro- and anti-inflammatory biomarkers (IL-1 beta, IL-8, TNF-alpha, IL-6, IFN-gamma, IL-10, IL-4) in spleen and lymph nodes. 12 weaned piglets were fed for 30 days with a control diet or an experimental diet containing 5% GS. At the end of trial, plasma and tissue samples (spleen and lymph nodes) were collected and the biochemical and inflammatory markers were analysed by using biochemistry analyser and ELISA techniques. Our results showed that diet included 5% GS did not influence the health status determined by plasma biochemical parameters. Only a tendency for a slight increase of the biochemical parameters associated with energetic profile (glucose, cholesterol, triglycerides) was observed. Also, GS diet had no effect on pro- and anti-inflammatory cytokines content in spleen and lymph nodes tissue. Further experiments are needed in order to investigate other rate of dietary inclusion which could provide more evidence about the effect of grape bioactive compounds on pigs used as animal model.

Keywords: animal model, inflammation, grape seed by-product, immune organs

Procedia PDF Downloads 290
2132 Analysis of the Evolution of Techniques and Review in Cleft Surgery

Authors: Tomaz Oliveira, Rui Medeiros, André Lacerda

Abstract:

Introduction: Cleft lip and/or palate are the most frequent forms of congenital craniofacial anomalies, affecting mainly the middle third of the face and manifesting by functional and aesthetic changes. Bilateral cleft lip represents a reconstructive surgical challenge, not only for the labial component but also for the associated nasal deformation. Recently, the paradigm of the approach to this pathology has changed, placing the focus on muscle reconstruction and anatomical repositioning of the nasal cartilages in order to obtain the best aesthetic and functional results. The aim of this study is to carry out a systematic review of the surgical approach to bilateral cleft lip, retrospectively analyzing the case series of Plastic Surgery Service at Hospital Santa Maria (Lisbon, Portugal) regarding this pathology, the global assessment of the characteristics of the operated patients and the study of the different surgical approaches and their complications in the last 20 years. Methods: The present work demonstrates a retrospective and descriptive study of patients who underwent at least one reconstructive surgery for cleft lip and/or palate, in the CPRE service of the HSM, in the period between January 1 of 1997 and December 31 of 2017, in which the data relating to 361 individuals were analyzed who, after applying the exclusion criteria, constituted a sample of 212 participants. The variables analyzed were the year of the first surgery, gender, age, type of orofacial cleft, surgical approach, and its complications. Results: There was a higher overall prevalence in males, with cleft lip and cleft palate occurring in greater proportion in males, with the cleft palate being more common in females. The most frequently recorded malformation was cleft lip and palate, which is complete in most cases. Regarding laterality, alterations with a unilateral labial component were the most commonly observed, with the left lip being described as the most affected. It was found that the vast majority of patients underwent primary intervention up to 12 months of age. The surgical techniques used in the approach to this pathology showed an important chronological variation over the years. Discussion: Cleft lip and/or palate is a medical condition associated with high aesthetic and functional morbidity, which requires early treatment in order to optimize the long-term outcome. The existence of a nasolabial component and its respective surgical correction plays a central role in the treatment of this pathology. The high rates of post-surgical complications and unconvincing aesthetic results have motivated an evolution of the surgical technique, increasingly evident in recent years, allowing today to achieve satisfactory aesthetic results, even in bilateral cleft lip with high deformation complexity. The introduction of techniques that favor nasolabial reconstruction based on anatomical principles has been producing increasingly convincing results. The analyzed sample shows that most of the results obtained in this study are, in general, compatible with the results published in the literature. Conclusion: This work showed that the existence of small variations in the surgical technique can bring significant improvements in the functional and aesthetic results in the treatment of bilateral cleft lip.

Keywords: cleft lip, palate lip, congenital abnormalities, cranofacial malformations

Procedia PDF Downloads 110
2131 Evaluation of Different Anticoagulant Effects on Flow Properties of Human Blood Using Falling Needle Rheometer

Authors: Hiroki Tsuneda, Takamasa Suzuki, Hideki Yamamoto, Kimito Kawamura, Eiji Tamura, Katharina Wochner, Roberto Plasenzotti

Abstract:

Flow property of human blood is one of the important factors on the prevention of the circulatory condition such as a high blood pressure, a diabetes mellitus, and a cardiac infarction. However, the measurement of flow property of human blood, especially blood viscosity, is not so easy, because of their coagulation or aggregation behaviors after taking a sample from blood vessel. In the experiment, some kinds of anticoagulant were added into the human blood to avoid its solidification. Anticoagulant used in the blood test has been chosen for each purpose of blood test, for anticoagulant effect on blood is different mechanism for each. So that, there is a problem that the evaluation of measured blood property with different anticoagulant is so difficult. Therefore, it is so important to make clear the difference of anticoagulant effect on the blood property. In the previous work, a compact-size falling needle rheometer (FNR) has been developed in order to measure the flow property of human blood such as a flow curve, an apparent viscosity. It was found that FNR system can apply to a rheometer or a viscometry for various experimental conditions for not only human blood but also mammalians blood. In this study, the measurements of human blood viscosity with different anticoagulant (EDTA and Heparin) were carried out using newly developed FNR system. The effect of anticoagulant on blood viscosity was also tested by using the standard liquid for each. The accuracy on the viscometry was also tested by using the standard liquid for calibrating materials (JS-10, JS-20) and observed data have satisfactory agreement with reference data around 1.0% at 310K. The flow curve of six males and females with different anticoagulant were measured using FNR. In this experiment, EDTA and Heparin were chosen as anticoagulant for blood. Heparin can inhibit the coagulation of human blood by activating the body of anti-thrombin. To examine the effect of human blood viscosity on anticoagulant, flow curve was measured at high shear rate (>350s-1), and apparent viscosity of each person were determined with different anticoagulant. The apparent viscosity of human blood with heparin was 2%-9% higher than that with EDTA. However, the difference of blood viscosity for two anticoagulants for same blood was different for each. Further discussion, we need the consideration of effect on other physical property, such as cellular component and plasma component.

Keywords: falling-needle rheometer, human blood, viscosity, anticoagulant

Procedia PDF Downloads 442
2130 Determinants of Corporate Social Responsibility Adoption: Evidence from China

Authors: Jing (Claire) LI

Abstract:

More than two decades from 2000 to 2020 of economic reforms have brought China unprecedented economic growth. There is an urgent call of research towards corporate social responsibility (CSR) in the context of China because while China continues to develop into a global trading market, it suffers from various serious problems relating to CSR. This study analyses the factors affecting the adoption of CSR practices by Chinese listed companies. The author proposes a new framework of factors of CSR adoption. Following common organisational factors and external factors in the literature (including organisational support, company size, shareholder pressures, and government support), this study introduces two additional factors, dynamic capability and regional culture. A survey questionnaire was conducted on the CSR adoption of Chinese listed companies in Shen Zhen and Shang Hai index from December 2019 to March 2020. The survey was conducted to collect data on the factors that affect the adoption of CSR. After collection of data, this study performed factor analysis to reduce the number of measurement items to several main factors. This procedure is to confirm the proposed framework and ensure the significant factors. Through analysis, this study identifies four grouped factors as determinants of the CSR adoption. The first factor loading includes dynamic capability and organisational support. The study finds that they are positively related to the first factor, so the first factor mainly reflects the capabilities of companies, which is one component in internal factors. In the second factor, measurement items of stakeholder pressures mainly are from regulatory bodies, customer and supplier, employees and community, and shareholders. In sum, they are positively related to the second factor and they reflect stakeholder pressures, which is one component of external factors. The third factor reflects organisational characteristics. Variables include company size and cultural score. Among these variables, company size is negatively related to the third factor. The resulted factor loading of the third factor implies that organisational factor is an important determinant of CSR adoption. Cultural consistency, the variable in the fourth factor, is positively related to the factor. It represents the difference between perception of managers and actual culture of the organisations in terms of cultural dimensions, which is one component in internal factors. It implies that regional culture is an important factor of CSR adoption. Overall, the results are consistent with previous literature. This study is of significance from both theoretical and empirical perspectives. First, from the significance of theoretical perspective, this research combines stakeholder theory, dynamic capability view of a firm, and neo-institutional theory in CSR research. Based on association of these three theories, this study introduces two new factors (dynamic capability and regional culture) to have a better framework for CSR adoption. Second, this study contributes to empirical literature of CSR in the context of China. Extant Chinese companies lack recognition of the importance of CSR practices adoption. This study built a framework and may help companies to design resource allocation strategies and evaluate future CSR and management practices in an early stage.

Keywords: China, corporate social responsibility, CSR adoption, dynamic capability, regional culture

Procedia PDF Downloads 134
2129 The Formation of Motivational Sphere for Learning Activity under Conditions of Change of One of Its Leading Components

Authors: M. Rodionov, Z. Dedovets

Abstract:

This article discusses ways to implement a differentiated approach to developing academic motivation for mathematical studies which relies on defining the primary structural characteristics of motivation. The following characteristics are considered: features of realization of cognitive activity, meaning-making characteristics, level of generalization and consistency of knowledge acquired by personal experience. The assessment of the present level of individual student understanding of each component of academic motivation is the basis for defining the relevant educational strategy for its further development.

Keywords: learning activity, mathematics, motivation, student

Procedia PDF Downloads 417