Search results for: hierarchical graph neuron
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1090

Search results for: hierarchical graph neuron

100 MicroRNA-1246 Expression Associated with Resistance to Oncogenic BRAF Inhibitors in Mutant BRAF Melanoma Cells

Authors: Jae-Hyeon Kim, Michael Lee

Abstract:

Intrinsic and acquired resistance limits the therapeutic benefits of oncogenic BRAF inhibitors in melanoma. MicroRNAs (miRNA) regulate the expression of target mRNAs by repressing their translation. Thus, we investigated miRNA expression patterns in melanoma cell lines to identify candidate biomarkers for acquired resistance to BRAF inhibitor. Here, we used Affymetrix miRNA V3.0 microarray profiling platform to compare miRNA expression levels in three cell lines containing BRAF inhibitor-sensitive A375P BRAF V600E cells, their BRAF inhibitor-resistant counterparts (A375P/Mdr), and SK-MEL-2 BRAF-WT cells with intrinsic resistance to BRAF inhibitor. The miRNAs with at least a two-fold change in expression between BRAF inhibitor-sensitive and –resistant cell lines, were identified as differentially expressed. Averaged intensity measurements identified 138 and 217 miRNAs that were differentially expressed by 2 fold or more between: 1) A375P and A375P/Mdr; 2) A375P and SK-MEL-2, respectively. The hierarchical clustering revealed differences in miRNA expression profiles between BRAF inhibitor-sensitive and –resistant cell lines for miRNAs involved in intrinsic and acquired resistance to BRAF inhibitor. In particular, 43 miRNAs were identified whose expression was consistently altered in two BRAF inhibitor-resistant cell lines, regardless of intrinsic and acquired resistance. Twenty five miRNAs were consistently upregulated and 18 downregulated more than 2-fold. Although some discrepancies were detected when miRNA microarray data were compared with qPCR-measured expression levels, qRT-PCR for five miRNAs (miR-3617, miR-92a1, miR-1246, miR-1936-3p, and miR-17-3p) results showed excellent agreement with microarray experiments. To further investigate cellular functions of miRNAs, we examined effects on cell proliferation. Synthetic oligonucleotide miRNA mimics were transfected into three cell lines, and proliferation was quantified using a colorimetric assay. Of the 5 miRNAs tested, only miR-1246 altered cell proliferation of A375P/Mdr cells. The transfection of miR-1246 mimic strongly conferred PLX-4720 resistance to A375P/Mdr cells, implying that miR-1246 upregulation confers acquired resistance to BRAF inhibition. We also found that PLX-4720 caused much greater G2/M arrest in A375P/Mdr cells transfected with miR-1246mimic than that seen in scrambled RNA-transfected cells. Additionally, miR-1246 mimic partially caused a resistance to autophagy induction by PLX-4720. These results indicate that autophagy does play an essential death-promoting role inPLX-4720-induced cell death. Taken together, these results suggest that miRNA expression profiling in melanoma cells can provide valuable information for a network of BRAF inhibitor resistance-associated miRNAs.

Keywords: microRNA, BRAF inhibitor, drug resistance, autophagy

Procedia PDF Downloads 305
99 The Link Between Collaboration Interactions and Team Creativity Among Nursing Student Teams in Taiwan: A Moderated Mediation Model

Authors: Hsing Yuan Liu

Abstract:

Background: Considerable theoretical and empirical work has identified a relationship between collaboration interactions and creativity in an organizational context. The mechanisms underlying this link, however, are not well understood in healthcare education. Objectives: The aims of this study were to explore the impact of collaboration interactions on team creativity and its underlying mechanism and to verify a moderated mediation model. Design, setting, and participants: This study utilized a cross-sectional, quantitative, descriptive design. The survey data were collected from 177 nursing students who enrolled in 18-week capstone courses of small interdisciplinary groups collaborating to design healthcare products in Taiwan during 2018 and 2019. Methods: Questionnaires assessed the nursing students' perceptions about their teams' swift trust (of cognition- and affect-based), conflicts (of task, process, and relationship), interaction behaviors (constructive controversy, helping behaviors, and spontaneous communication), and creativity. This study used descriptive statistics to compare demographics, swift trust scores, conflict scores, interaction behavior scores, and creativity scores for interdisciplinary teams. Data were analyzed using Pearson’s correlation coefficient and simple and hierarchical multiple regression models. Results: Pearson’s correlation analysis showed the cognition-based team swift trust was positively correlated with team creativity. The mediation model indicated constructive controversy fully mediated the effect of cognition-based team swift trust on student teams’ creativity. The moderated mediation model indicated that task conflict negatively moderates the mediating effect of the constructive controversy on the link between cognition-based team swift trust and team creativity. Conclusion: Our findings suggest nursing student teams’ interaction behaviors and task conflict are crucial mediating and moderated mediation variables on the relationship between collaboration interactions and team creativity, respectively. The empirical data confirms the validity of our proposed moderated mediation models of team creativity. Therefore, this study's validated moderated mediation model could provide guidance for nursing educators to improve collaboration interaction outcomes and creativity on nursing student teams.

Keywords: team swift trust, team conflict, team interaction behavior, moderated mediating effects, interdisciplinary education, nursing students

Procedia PDF Downloads 169
98 Automated Building Internal Layout Design Incorporating Post-Earthquake Evacuation Considerations

Authors: Sajjad Hassanpour, Vicente A. González, Yang Zou, Jiamou Liu

Abstract:

Earthquakes pose a significant threat to both structural and non-structural elements in buildings, putting human lives at risk. Effective post-earthquake evacuation is critical for ensuring the safety of building occupants. However, current design practices often neglect the integration of post-earthquake evacuation considerations into the early-stage architectural design process. To address this gap, this paper presents a novel automated internal architectural layout generation tool that optimizes post-earthquake evacuation performance. The tool takes an initial plain floor plan as input, along with specific requirements from the user/architect, such as minimum room dimensions, corridor width, and exit lengths. Based on these inputs, firstly, the tool randomly generates different architectural layouts. Secondly, the human post-earthquake evacuation behaviour will be thoroughly assessed for each generated layout using the advanced Agent-Based Building Earthquake Evacuation Simulation (AB2E2S) model. The AB2E2S prototype is a post-earthquake evacuation simulation tool that incorporates variables related to earthquake intensity, architectural layout, and human factors. It leverages a hierarchical agent-based simulation approach, incorporating reinforcement learning to mimic human behaviour during evacuation. The model evaluates different layout options and provides feedback on evacuation flow, time, and possible casualties due to earthquake non-structural damage. By integrating the AB2E2S model into the automated layout generation tool, architects and designers can obtain optimized architectural layouts that prioritize post-earthquake evacuation performance. Through the use of the tool, architects and designers can explore various design alternatives, considering different minimum room requirements, corridor widths, and exit lengths. This approach ensures that evacuation considerations are embedded in the early stages of the design process. In conclusion, this research presents an innovative automated internal architectural layout generation tool that integrates post-earthquake evacuation simulation. By incorporating evacuation considerations into the early-stage design process, architects and designers can optimize building layouts for improved post-earthquake evacuation performance. This tool empowers professionals to create resilient designs that prioritize the safety of building occupants in the face of seismic events.

Keywords: agent-based simulation, automation in design, architectural layout, post-earthquake evacuation behavior

Procedia PDF Downloads 82
97 Integrating System-Level Infrastructure Resilience and Sustainability Based on Fractal: Perspectives and Review

Authors: Qiyao Han, Xianhai Meng

Abstract:

Urban infrastructures refer to the fundamental facilities and systems that serve cities. Due to the global climate change and human activities in recent years, many urban areas around the world are facing enormous challenges from natural and man-made disasters, like flood, earthquake and terrorist attack. For this reason, urban resilience to disasters has attracted increasing attention from researchers and practitioners. Given the complexity of infrastructure systems and the uncertainty of disasters, this paper suggests that studies of resilience could focus on urban functional sustainability (in social, economic and environmental dimensions) supported by infrastructure systems under disturbance. It is supposed that urban infrastructure systems with high resilience should be able to reconfigure themselves without significant declines in critical functions (services), such as primary productivity, hydrological cycles, social relations and economic prosperity. Despite that some methods have been developed to integrate the resilience and sustainability of individual infrastructure components, more work is needed to enable system-level integration. This research presents a conceptual analysis framework for integrating resilience and sustainability based on fractal theory. It is believed that the ability of an ecological system to maintain structure and function in face of disturbance and to reorganize following disturbance-driven change is largely dependent on its self-similar and hierarchical fractal structure, in which cross-scale resilience is produced by the replication of ecosystem processes dominating at different levels. Urban infrastructure systems are analogous to ecological systems because they are interconnected, complex and adaptive, are comprised of interconnected components, and exhibit characteristic scaling properties. Therefore, analyzing resilience of ecological system provides a better understanding about the dynamics and interactions of infrastructure systems. This paper discusses fractal characteristics of ecosystem resilience, reviews literature related to system-level infrastructure resilience, identifies resilience criteria associated with sustainability dimensions, and develops a conceptual analysis framework. Exploration of the relevance of identified criteria to fractal characteristics reveals that there is a great potential to analyze infrastructure systems based on fractal. In the conceptual analysis framework, it is proposed that in order to be resilient, urban infrastructure system needs to be capable of “maintaining” and “reorganizing” multi-scale critical functions under disasters. Finally, the paper identifies areas where further research efforts are needed.

Keywords: fractal, urban infrastructure, sustainability, system-level resilience

Procedia PDF Downloads 257
96 The Dynamics of Planktonic Crustacean Populations in an Open Access Lagoon, Bordered by Heavy Industry, Southwest, Nigeria

Authors: E. O. Clarke, O. J. Aderinola, O. A. Adeboyejo, M. A. Anetekhai

Abstract:

Aims: The study is aimed at establishing the influence of some physical and chemical parameters on the abundance, distribution pattern and seasonal variations of the planktonic crustacean populations. Place and Duration of Study: A premier investigation into the dynamics of planktonic crustacean populations in Ologe lagoon was carried out from January 2011 to December 2012. Study Design: The study covered identification, temporal abundance, spatial distribution and diversity of the planktonic crustacea. Methodology: Standard techniques were used to collect samples from eleven stations covering five proximal satellite towns (Idoluwo, Oto, Ibiye, Obele, and Gbanko) bordering the lagoon. Data obtained were statistically analyzed using linear regression and hierarchical clustering. Results:Thirteen (13) planktonic crustacean populations were identified. Total percentage abundance was highest for Bosmina species (20%) and lowest for Polyphemus species (0.8%). The Pearson’s correlation coefficient (“r” values) between total planktonic crustacean population and some physical and chemical parameters showed that positive correlations having low level of significance occurred with salinity (r = 0.042) (sig = 0.184) and with surface water dissolved oxygen (r = 0.299) (sig = 0.155). Linear regression plots indicated that, the total population of planktonic crustacea were mainly influenced and only increased with an increase in value of surface water temperature (Rsq = 0.791) and conductivity (Rsq = 0.589). The total population of planktonic crustacea had a near neutral (zero correlation) with the surface water dissolved oxygen and thus, does not significantly change with the level of the surface water dissolved oxygen. The correlations were positive with NO3-N (midstream) at Ibiye (Rsq =0.022) and (downstream) Gbanko (Rsq =0.013), PO4-P at Ibiye (Rsq =0.258), K at Idoluwo (Rsq =0.295) and SO4-S at Oto (Rsq = 0.094) and Gbanko (Rsq = 0.457). The Berger-Parker Dominance Index (BPDI) showed that the most dominant species was Bosmina species (BPDI = 1.000), followed by Calanus species (BPDI = 1.254). Clusters by squared Euclidan distances using average linkage between groups showed proximities, transcending the borders of genera. Conclusion: The results revealed that planktonic crustacean population in Ologe lagoon undergo seasonal perturbations, were highly influenced by nutrient, metal and organic matter inputs from river Owoh, Agbara industrial estate and surrounding farmlands and were patchy in spatial distribution.

Keywords: diversity, dominance, perturbations, richness, crustacea, lagoon

Procedia PDF Downloads 704
95 Identification of Peroxisome Proliferator-Activated Receptors α/γ Dual Agonists for Treatment of Metabolic Disorders, Insilico Screening, and Molecular Dynamics Simulation

Authors: Virendra Nath, Vipin Kumar

Abstract:

Background: TypeII Diabetes mellitus is a foremost health problem worldwide, predisposing to increased mortality and morbidity. Undesirable effects of the current medications have prompted the researcher to develop more potential drug(s) against the disease. The peroxisome proliferator-activated receptors (PPARs) are members of the nuclear receptors family and take part in a vital role in the regulation of metabolic equilibrium. They can induce or repress genes associated with adipogenesis, lipid, and glucose metabolism. Aims: Investigation of PPARα/γ agonistic hits were screened by hierarchical virtual screening followed by molecular dynamics simulation and knowledge-based structure-activity relation (SAR) analysis using approved PPAR α/γ dual agonist. Methods: The PPARα/γ agonistic activity of compounds was searched by using Maestro through structure-based virtual screening and molecular dynamics (MD) simulation application. Virtual screening of nuclear-receptor ligands was done, and the binding modes with protein-ligand interactions of newer entity(s) were investigated. Further, binding energy prediction, Stability studies using molecular dynamics (MD) simulation of PPARα and γ complex was performed with the most promising hit along with the structural comparative analysis of approved PPARα/γ agonists with screened hit was done for knowledge-based SAR. Results and Discussion: The silicone chip-based approach recognized the most capable nine hits and had better predictive binding energy as compared to the reference drug compound (Tesaglitazar). In this study, the key amino acid residues of binding pockets of both targets PPARα/γ were acknowledged as essential and were found to be associated in the key interactions with the most potential dual hit (ChemDiv-3269-0443). Stability studies using molecular dynamics (MD) simulation of PPARα and γ complex was performed with the most promising hit and found root mean square deviation (RMSD) stabile around 2Å and 2.1Å, respectively. Frequency distribution data also revealed that the key residues of both proteins showed maximum contacts with a potent hit during the MD simulation of 20 nanoseconds (ns). The knowledge-based SAR studies of PPARα/γ agonists were studied using 2D structures of approved drugs like aleglitazar, tesaglitazar, etc. for successful designing and synthesis of compounds PPARγ agonistic candidates with anti-hyperlipidimic potential.

Keywords: computational, diabetes, PPAR, simulation

Procedia PDF Downloads 81
94 Simscape Library for Large-Signal Physical Network Modeling of Inertial Microelectromechanical Devices

Authors: S. Srinivasan, E. Cretu

Abstract:

The information flow (e.g. block-diagram or signal flow graph) paradigm for the design and simulation of Microelectromechanical (MEMS)-based systems allows to model MEMS devices using causal transfer functions easily, and interface them with electronic subsystems for fast system-level explorations of design alternatives and optimization. Nevertheless, the physical bi-directional coupling between different energy domains is not easily captured in causal signal flow modeling. Moreover, models of fundamental components acting as building blocks (e.g. gap-varying MEMS capacitor structures) depend not only on the component, but also on the specific excitation mode (e.g. voltage or charge-actuation). In contrast, the energy flow modeling paradigm in terms of generalized across-through variables offers an acausal perspective, separating clearly the physical model from the boundary conditions. This promotes reusability and the use of primitive physical models for assembling MEMS devices from primitive structures, based on the interconnection topology in generalized circuits. The physical modeling capabilities of Simscape have been used in the present work in order to develop a MEMS library containing parameterized fundamental building blocks (area and gap-varying MEMS capacitors, nonlinear springs, displacement stoppers, etc.) for the design, simulation and optimization of MEMS inertial sensors. The models capture both the nonlinear electromechanical interactions and geometrical nonlinearities and can be used for both small and large signal analyses, including the numerical computation of pull-in voltages (stability loss). Simscape behavioral modeling language was used for the implementation of reduced-order macro models, that present the advantage of a seamless interface with Simulink blocks, for creating hybrid information/energy flow system models. Test bench simulations of the library models compare favorably with both analytical results and with more in-depth finite element simulations performed in ANSYS. Separate MEMS-electronic integration tests were done on closed-loop MEMS accelerometers, where Simscape was used for modeling the MEMS device and Simulink for the electronic subsystem.

Keywords: across-through variables, electromechanical coupling, energy flow, information flow, Matlab/Simulink, MEMS, nonlinear, pull-in instability, reduced order macro models, Simscape

Procedia PDF Downloads 118
93 Fuzzy Optimization for Identifying Anticancer Targets in Genome-Scale Metabolic Models of Colon Cancer

Authors: Feng-Sheng Wang, Chao-Ting Cheng

Abstract:

Developing a drug from conception to launch is costly and time-consuming. Computer-aided methods can reduce research costs and accelerate the development process during the early drug discovery and development stages. This study developed a fuzzy multi-objective hierarchical optimization framework for identifying potential anticancer targets in a metabolic model. First, RNA-seq expression data of colorectal cancer samples and their healthy counterparts were used to reconstruct tissue-specific genome-scale metabolic models. The aim of the optimization framework was to identify anticancer targets that lead to cancer cell death and evaluate metabolic flux perturbations in normal cells that have been caused by cancer treatment. Four objectives were established in the optimization framework to evaluate the mortality of cancer cells for treatment and to minimize side effects causing toxicity-induced tumorigenesis on normal cells and smaller metabolic perturbations. Through fuzzy set theory, a multiobjective optimization problem was converted into a trilevel maximizing decision-making (MDM) problem. The applied nested hybrid differential evolution was applied to solve the trilevel MDM problem using two nutrient media to identify anticancer targets in the genome-scale metabolic model of colorectal cancer, respectively. Using Dulbecco’s Modified Eagle Medium (DMEM), the computational results reveal that the identified anticancer targets were mostly involved in cholesterol biosynthesis, pyrimidine and purine metabolisms, glycerophospholipid biosynthetic pathway and sphingolipid pathway. However, using Ham’s medium, the genes involved in cholesterol biosynthesis were unidentifiable. A comparison of the uptake reactions for the DMEM and Ham’s medium revealed that no cholesterol uptake reaction was included in DMEM. Two additional media, i.e., a cholesterol uptake reaction was included in DMEM and excluded in HAM, were respectively used to investigate the relationship of tumor cell growth with nutrient components and anticancer target genes. The genes involved in the cholesterol biosynthesis were also revealed to be determinable if a cholesterol uptake reaction was not induced when the cells were in the culture medium. However, the genes involved in cholesterol biosynthesis became unidentifiable if such a reaction was induced.

Keywords: Cancer metabolism, genome-scale metabolic model, constraint-based model, multilevel optimization, fuzzy optimization, hybrid differential evolution

Procedia PDF Downloads 63
92 Layer-By-Layer Deposition of Poly (Amidoamine) and Poly (Acrylic Acid) on Grafted-Polylactide Nonwoven with Different Surface Charge

Authors: Sima Shakoorjavan, Mahdieh Eskafi, Dawid Stawski, Somaye Akbari

Abstract:

In this study, poly (amidoamine) dendritic material (PAMAM) and poly (acrylic acid) (PAA) as polycation and polyanion were deposited on surface charged polylactide (PLA) nonwoven to study the relationship of dye absorption capacity of layered-PLA with the number of deposited layers. To produce negatively charged-PLA, acrylic acid (AA) was grafted on the PLA surface (PLA-g-AA) through a chemical redox reaction with the strong oxidizing agent. Spectroscopy analysis, water contact measurement, and FTIR-ATR analysis confirm the successful grafting of AA on the PLA surface through the chemical redox reaction method. In detail, an increase in dye absorption percentage by 19% and immediate absorption of water droplets ensured hydrophilicity of PLA-g-AA surface; and the presence of new carbonyl bond at 1530 cm-¹ and a wide peak of hydroxyl between 3680-3130 cm-¹ confirm AA grafting. In addition, PLA as linear polyester can undergo aminolysis, which is the cleavage of ester bonds and replacement with amid bonds when exposed to an aminolysis agent. Therefore, to produce positively charged PLA, PAMAM as amine-terminated dendritic material was introduced to PLA molecular chains at different conditions; (1) at 60 C for 0.5, 1, 1.5, 2 hours of aminolysis and (2) at room temperature (RT) for 1, 2, 3, and 4 hours of aminolysis. Weight changes and spectrophotometer measurements showed a maximum in weight gain graph and K/S value curve indicating the highest PAMAM attachment at 60 C for 1 hour and RT for 2 hours which is considered as an optimum condition. Also, the emerging new peak around 1650 cm-1 corresponding to N-H bending vibration and double wide peak at around 3670-3170 cm-1 corresponding to N-H stretching vibration confirm PAMAM attachment in selected optimum condition. In the following, regarding the initial surface charge of grafted-PLA, lbl deposition was performed and started with PAA or PAMAM. FTIR-ATR results confirm chemical changes in samples due to deposition of the first layer (PAA or PAMAM). Generally, spectroscopy analysis indicated that an increase in layer number costed dye absorption capacity. It can be due to the partial deposition of a new layer on the previously deposited layer; therefore, the available PAMAM at the first layer is more than the third layer. In detail, in the case of layer-PLA starting lbl with negatively charged, having PAMAM as the first top layer (PLA-g-AA/PAMAM) showed the highest dye absorption of both cationic and anionic model dye.

Keywords: surface modification, layer-by-layer technique, dendritic materials, PAMAM, dye absorption capacity, PLA nonwoven

Procedia PDF Downloads 66
91 Cluster-Based Exploration of System Readiness Levels: Mathematical Properties of Interfaces

Authors: Justin Fu, Thomas Mazzuchi, Shahram Sarkani

Abstract:

A key factor in technological immaturity in defense weapons acquisition is lack of understanding critical integrations at the subsystem and component level. To address this shortfall, recent research in integration readiness level (IRL) combines with technology readiness level (TRL) to form a system readiness level (SRL). SRL can be enriched with more robust quantitative methods to provide the program manager a useful tool prior to committing to major weapons acquisition programs. This research harnesses previous mathematical models based on graph theory, Petri nets, and tropical algebra and proposes a modification of the desirable SRL mathematical properties such that a tightly integrated (multitude of interfaces) subsystem can display a lower SRL than an inherently less coupled subsystem. The synthesis of these methods informs an improved decision tool for the program manager to commit to expensive technology development. This research ties the separately developed manufacturing readiness level (MRL) into the network representation of the system and addresses shortfalls in previous frameworks, including the lack of integration weighting and the over-importance of a single extremely immature component. Tropical algebra (based on the minimum of a set of TRLs or IRLs) allows one low IRL or TRL value to diminish the SRL of the entire system, which may not be reflective of actuality if that component is not critical or tightly coupled. Integration connections can be weighted according to importance and readiness levels are modified to be a cardinal scale (based on an analytic hierarchy process). Integration arcs’ importance are dependent on the connected nodes and the additional integrations arcs connected to those nodes. Lack of integration is not represented by zero, but by a perfect integration maturity value. Naturally, the importance (or weight) of such an arc would be zero. To further explore the impact of grouping subsystems, a multi-objective genetic algorithm is then used to find various clusters or communities that can be optimized for the most representative subsystem SRL. This novel calculation is then benchmarked through simulation and using past defense acquisition program data, focusing on the newly introduced Middle Tier of Acquisition (rapidly field prototypes). The model remains a relatively simple, accessible tool, but at higher fidelity and validated with past data for the program manager to decide major defense acquisition program milestones.

Keywords: readiness, maturity, system, integration

Procedia PDF Downloads 72
90 Evaluating the Factors Controlling the Hydrochemistry of Gaza Coastal Aquifer Using Hydrochemical and Multivariate Statistical Analysis

Authors: Madhat Abu Al-Naeem, Ismail Yusoff, Ng Tham Fatt, Yatimah Alias

Abstract:

Groundwater in Gaza strip is increasingly being exposed to anthropic and natural factors that seriously impacted the groundwater quality. Physiochemical data of groundwater can offer important information on changes in groundwater quality that can be useful in improving water management tactics. An integrative hydrochemical and statistical techniques (Hierarchical cluster analysis (HCA) and factor analysis (FA)) have been applied on the existence ten physiochemical data of 84 samples collected in (2000/2001) using STATA, AquaChem, and Surfer softwares to: 1) Provide valuable insight into the salinization sources and the hydrochemical processes controlling the chemistry of groundwater. 2) Differentiate the influence of natural processes and man-made activities. The recorded large diversity in water facies with dominance Na-Cl type that reveals a highly saline aquifer impacted by multiple complex hydrochemical processes. Based on WHO standards, only (15.5%) of the wells were suitable for drinking. HCA yielded three clusters. Cluster 1 is the highest in salinity, mainly due to the impact of Eocene saline water invasion mixed with human inputs. Cluster 2 is the lowest in salinity also due to Eocene saline water invasion but mixed with recent rainfall recharge and limited carbonate dissolution and nitrate pollution. Cluster 3 is similar in salinity to Cluster 2, but with a high diversity of facies due to the impact of many sources of salinity as sea water invasion, carbonate dissolution and human inputs. Factor analysis yielded two factors accounting for 88% of the total variance. Factor 1 (59%) is a salinization factor demonstrating the mixing contribution of natural saline water with human inputs. Factor 2 measure the hardness and pollution which explained 29% of the total variance. The negative relationship between the NO3- and pH may reveal a denitrification process in a heavy polluted aquifer recharged by a limited oxygenated rainfall. Multivariate statistical analysis combined with hydrochemical analysis indicate that the main factors controlling groundwater chemistry were Eocene saline invasion, seawater invasion, sewage invasion and rainfall recharge and the main hydrochemical processes were base ion and reverse ion exchange processes with clay minerals (water rock interactions), nitrification, carbonate dissolution and a limited denitrification process.

Keywords: dendrogram and cluster analysis, water facies, Eocene saline invasion and sea water invasion, nitrification and denitrification

Procedia PDF Downloads 344
89 Winter – Not Spring - Climate Drives Annual Adult Survival in Common Passerines: A Country-Wide, Multi-Species Modeling Exercise

Authors: Manon Ghislain, Timothée Bonnet, Olivier Gimenez, Olivier Dehorter, Pierre-Yves Henry

Abstract:

Climatic fluctuations affect the demography of animal populations, generating changes in population size, phenology, distribution and community assemblages. However, very few studies have identified the underlying demographic processes. For short-lived species, like common passerine birds, are these changes generated by changes in adult survival or in fecundity and recruitment? This study tests for an effect of annual climatic conditions (spring and winter) on annual, local adult survival at very large spatial (a country, 252 sites), temporal (25 years) and biological (25 species) scales. The Constant Effort Site ringing has allowed the collection of capture - mark - recapture data for 100 000 adult individuals since 1989, over metropolitan France, thus documenting annual, local survival rates of the most common passerine birds. We specifically developed a set of multi-year, multi-species, multi-site Bayesian models describing variations in local survival and recapture probabilities. This method allows for a statistically powerful hierarchical assessment (global versus species-specific) of the effects of climate variables on survival. A major part of between-year variations in survival rate was common to all species (74% of between-year variance), whereas only 26% of temporal variation was species-specific. Although changing spring climate is commonly invoked as a cause of population size fluctuations, spring climatic anomalies (mean precipitation or temperature for March-August) do not impact adult survival: only 1% of between-year variation of species survival is explained by spring climatic anomalies. However, for sedentary birds, winter climatic anomalies (North Atlantic Oscillation) had a significant, quadratic effect on adult survival, birds surviving less during intermediate years than during more extreme years. For migratory birds, we do not detect an effect of winter climatic anomalies (Sahel Rainfall). We will analyze the life history traits (migration, habitat, thermal range) that could explain a different sensitivity of species to winter climate anomalies. Overall, we conclude that changes in population sizes for passerine birds are unlikely to be the consequences of climate-driven mortality (or emigration) in spring but could be induced by other demographic parameters, like fecundity.

Keywords: Bayesian approach, capture-recapture, climate anomaly, constant effort sites scheme, passerine, seasons, survival

Procedia PDF Downloads 282
88 Evaluation of Electrophoretic and Electrospray Deposition Methods for Preparing Graphene and Activated Carbon Modified Nano-Fibre Electrodes for Hydrogen/Vanadium Flow Batteries and Supercapacitors

Authors: Barun Chakrabarti, Evangelos Kalamaras, Vladimir Yufit, Xinhua Liu, Billy Wu, Nigel Brandon, C. T. John Low

Abstract:

In this work, we perform electrophoretic deposition of activated carbon on a number of substrates to prepare symmetrical coin cells for supercapacitor applications. From several recipes that involve the evaluation of a few solvents such as isopropyl alcohol, N-Methyl-2-pyrrolidone (NMP), or acetone to binders such as polyvinylidene fluoride (PVDF) and charging agents such as magnesium chloride, we display a working means for achieving supercapacitors that can achieve 100 F/g in a consistent manner. We then adapt this EPD method to deposit reduced graphene oxide on SGL 10AA carbon paper to achieve cathodic materials for testing in a hydrogen/vanadium flow battery. In addition, a self-supported hierarchical carbon nano-fibre is prepared by means of electrospray deposition of an iron phthalocyanine solution onto a temporary substrate followed by carbonisation to remove heteroatoms. This process also induces a degree of nitrogen doping on the carbon nano-fibres (CNFs), which allows its catalytic performance to improve significantly as detailed in other publications. The CNFs are then used as catalysts by attaching them to graphite felt electrodes facing the membrane inside an all-vanadium flow battery (Scribner cell using serpentine flow distribution channels) and efficiencies as high as 60% is noted at high current densities of 150 mA/cm². About 20 charge and discharge cycling show that the CNF catalysts consistently perform better than pristine graphite felt electrodes. Following this, we also test the CNF as an electro-catalyst in the hydrogen/vanadium flow battery (cathodic side as mentioned briefly in the first paragraph) facing the membrane, based upon past studies from our group. Once again, we note consistently good efficiencies of 85% and above for CNF modified graphite felt electrodes in comparison to 60% for pristine felts at low current density of 50 mA/cm² (this reports 20 charge and discharge cycles of the battery). From this preliminary investigation, we conclude that the CNFs may be used as catalysts for other systems such as vanadium/manganese, manganese/manganese and manganese/hydrogen flow batteries in the future. We are generating data for such systems at present, and further publications are expected.

Keywords: electrospinning, carbon nano-fibres, all-vanadium redox flow battery, hydrogen-vanadium fuel cell, electrocatalysis

Procedia PDF Downloads 280
87 Frequency Interpretation of a Wave Function, and a Vertical Waveform Treated as A 'Quantum Leap'

Authors: Anthony Coogan

Abstract:

Born’s probability interpretation of wave functions would have led to nearly identical results had he chosen a frequency interpretation instead. Logically, Born may have assumed that only one electron was under consideration, making it nonsensical to propose a frequency wave. Author’s suggestion: the actual experimental results were not of a single electron; rather, they were groups of reflected x-ray photons. The vertical waveform used by Scrhödinger in his Particle in the Box Theory makes sense if it was intended to represent a quantum leap. The author extended the single vertical panel to form a bar chart: separate panels would represent different energy levels. The proposed bar chart would be populated by reflected photons. Expansion of basic ideas: Part of Scrhödinger’s ‘Particle in the Box’ theory may be valid despite negative criticism. The waveform used in the diagram is vertical, which may seem absurd because real waves decay at a measurable rate, rather than instantaneously. However, there may be one notable exception. Supposedly, following from the theory, the Uncertainty Principle was derived – may a Quantum Leap not be represented as an instantaneous waveform? The great Scrhödinger must have had some reason to suggest a vertical waveform if the prevalent belief was that they did not exist. Complex wave forms representing a particle are usually assumed to be continuous. The actual observations made were x-ray photons, some of which had struck an electron, been reflected, and then moved toward a detector. From Born’s perspective, doing similar work the years in question 1926-7, he would also have considered a single electron – leading him to choose a probability distribution. Probability Distributions appear very similar to Frequency Distributions, but the former are considered to represent the likelihood of future events. Born’s interpretation of the results of quantum experiments led (or perhaps misled) many researchers into claiming that humans can influence events just by looking at them, e.g. collapsing complex wave functions by 'looking at the electron to see which slit it emerged from', while in reality light reflected from the electron moved in the observer’s direction after the electron had moved away. Astronomers may say that they 'look out into the universe' but are actually using logic opposed to the views of Newton and Hooke and many observers such as Romer, in that light carries information from a source or reflector to an observer, rather the reverse. Conclusion: Due to the controversial nature of these ideas, especially its implications about the nature of complex numbers used in applications in science and engineering, some time may pass before any consensus is reached.

Keywords: complex wave functions not necessary, frequency distributions instead of wave functions, information carried by light, sketch graph of uncertainty principle

Procedia PDF Downloads 188
86 Predictive Modelling of Curcuminoid Bioaccessibility as a Function of Food Formulation and Associated Properties

Authors: Kevin De Castro Cogle, Mirian Kubo, Maria Anastasiadi, Fady Mohareb, Claire Rossi

Abstract:

Background: The bioaccessibility of bioactive compounds is a critical determinant of the nutritional quality of various food products. Despite its importance, there is a limited number of comprehensive studies aimed at assessing how the composition of a food matrix influences the bioaccessibility of a compound of interest. This knowledge gap has prompted a growing need to investigate the intricate relationship between food matrix formulations and the bioaccessibility of bioactive compounds. One such class of bioactive compounds that has attracted considerable attention is curcuminoids. These naturally occurring phytochemicals, extracted from the roots of Curcuma longa, have gained popularity owing to their purported health benefits and also well known for their poor bioaccessibility Project aim: The primary objective of this research project is to systematically assess the influence of matrix composition on the bioaccessibility of curcuminoids. Additionally, this study aimed to develop a series of predictive models for bioaccessibility, providing valuable insights for optimising the formula for functional foods and provide more descriptive nutritional information to potential consumers. Methods: Food formulations enriched with curcuminoids were subjected to in vitro digestion simulation, and their bioaccessibility was characterized with chromatographic and spectrophotometric techniques. The resulting data served as the foundation for the development of predictive models capable of estimating bioaccessibility based on specific physicochemical properties of the food matrices. Results: One striking finding of this study was the strong correlation observed between the concentration of macronutrients within the food formulations and the bioaccessibility of curcuminoids. In fact, macronutrient content emerged as a very informative explanatory variable of bioaccessibility and was used, alongside other variables, as predictors in a Bayesian hierarchical model that predicted curcuminoid bioaccessibility accurately (optimisation performance of 0.97 R2) for the majority of cross-validated test formulations (LOOCV of 0.92 R2). These preliminary results open the door to further exploration, enabling researchers to investigate a broader spectrum of food matrix types and additional properties that may influence bioaccessibility. Conclusions: This research sheds light on the intricate interplay between food matrix composition and the bioaccessibility of curcuminoids. This study lays a foundation for future investigations, offering a promising avenue for advancing our understanding of bioactive compound bioaccessibility and its implications for the food industry and informed consumer choices.

Keywords: bioactive bioaccessibility, food formulation, food matrix, machine learning, probabilistic modelling

Procedia PDF Downloads 56
85 Magnetic Navigation in Underwater Networks

Authors: Kumar Divyendra

Abstract:

Underwater Sensor Networks (UWSNs) have wide applications in areas such as water quality monitoring, marine wildlife management etc. A typical UWSN system consists of a set of sensors deployed randomly underwater which communicate with each other using acoustic links. RF communication doesn't work underwater, and GPS too isn't available underwater. Additionally Automated Underwater Vehicles (AUVs) are deployed to collect data from some special nodes called Cluster Heads (CHs). These CHs aggregate data from their neighboring nodes and forward them to the AUVs using optical links when an AUV is in range. This helps reduce the number of hops covered by data packets and helps conserve energy. We consider the three-dimensional model of the UWSN. Nodes are initially deployed randomly underwater. They attach themselves to the surface using a rod and can only move upwards or downwards using a pump and bladder mechanism. We use graph theory concepts to maximize the coverage volume while every node maintaining connectivity with at least one surface node. We treat the surface nodes as landmarks and each node finds out its hop distance from every surface node. We treat these hop-distances as coordinates and use them for AUV navigation. An AUV intending to move closer to a node with given coordinates moves hop by hop through nodes that are closest to it in terms of these coordinates. In absence of GPS, multiple different approaches like Inertial Navigation System (INS), Doppler Velocity Log (DVL), computer vision-based navigation, etc., have been proposed. These systems have their own drawbacks. INS accumulates error with time, vision techniques require prior information about the environment. We propose a method that makes use of the earth's magnetic field values for navigation and combines it with other methods that simultaneously increase the coverage volume under the UWSN. The AUVs are fitted with magnetometers that measure the magnetic intensity (I), horizontal inclination (H), and Declination (D). The International Geomagnetic Reference Field (IGRF) is a mathematical model of the earth's magnetic field, which provides the field values for the geographical coordinateson earth. Researchers have developed an inverse deep learning model that takes the magnetic field values and predicts the location coordinates. We make use of this model within our work. We combine this with with the hop-by-hop movement described earlier so that the AUVs move in such a sequence that the deep learning predictor gets trained as quickly and precisely as possible We run simulations in MATLAB to prove the effectiveness of our model with respect to other methods described in the literature.

Keywords: clustering, deep learning, network backbone, parallel computing

Procedia PDF Downloads 81
84 Modelling the Antecedents of Supply Chain Enablers in Online Groceries Using Interpretive Structural Modelling and MICMAC Analysis

Authors: Rose Antony, Vivekanand B. Khanapuri, Karuna Jain

Abstract:

Online groceries have transformed the way the supply chains are managed. These are facing numerous challenges in terms of product wastages, low margins, long breakeven to achieve and low market penetration to mention a few. The e-grocery chains need to overcome these challenges in order to survive the competition. The purpose of this paper is to carry out a structural analysis of the enablers in e-grocery chains by applying Interpretive Structural Modeling (ISM) and MICMAC analysis in the Indian context. The research design is descriptive-explanatory in nature. The enablers have been identified from the literature and through semi-structured interviews conducted among the managers having relevant experience in e-grocery supply chains. The experts have been contacted through professional/social networks by adopting a purposive snowball sampling technique. The interviews have been transcribed, and manual coding is carried using open and axial coding method. The key enablers are categorized into themes, and the contextual relationship between these and the performance measures is sought from the Industry veterans. Using ISM, the hierarchical model of the enablers is developed and MICMAC analysis identifies the driver and dependence powers. Based on the driver-dependence power the enablers are categorized into four clusters namely independent, autonomous, dependent and linkage. The analysis found that information technology (IT) and manpower training acts as key enablers towards reducing the lead time and enhancing the online service quality. Many of the enablers fall under the linkage cluster viz., frequent software updating, branding, the number of delivery boys, order processing, benchmarking, product freshness and customized applications for different stakeholders, depicting these as critical in online food/grocery supply chains. Considering the perishability nature of the product being handled, the impact of the enablers on the product quality is also identified. Hence, study aids as a tool to identify and prioritize the vital enablers in the e-grocery supply chain. The work is perhaps unique, which identifies the complex relationships among the supply chain enablers in fresh food for e-groceries and linking them to the performance measures. It contributes to the knowledge of supply chain management in general and e-retailing in particular. The approach focus on the fresh food supply chains in the Indian context and hence will be applicable in developing economies context, where supply chains are evolving.

Keywords: interpretive structural modelling (ISM), India, online grocery, retail operations, supply chain management

Procedia PDF Downloads 190
83 The Role of Goal Orientation on the Structural-Psychological Empowerment Link in the Public Sector

Authors: Beatriz Garcia-Juan, Ana B. Escrig-Tena, Vicente Roca-Puig

Abstract:

The aim of this article is to conduct a theoretical and empirical study in order to examine how the goal orientation (GO) of public employees affects the relationship between the structural and psychological empowerment that they experience at their workplaces. In doing so, we follow structural empowerment (SE) and psychological empowerment (PE) conceptualizations, and relate them to the public administration framework. Moreover, we review arguments from GO theories, and previous related contributions. Empowerment has emerged as an important issue in the public sector organization setting in the wake of mainstream New Public Management (NPM), the new orientation in the public sector that aims to provide a better service for citizens. It is closely linked to the drive to improve organizational effectiveness through the wise use of human resources. Nevertheless, it is necessary to combine structural (managerial) and psychological (individual) approaches in an integrative study of empowerment. SE refers to a set of initiatives that aim the transference of power from managerial positions to the rest of employees. PE is defined as psychological state of competence, self-determination, impact, and meaning that an employee feels at work. Linking these two perspectives will lead to arrive at a broader understanding of the empowerment process. Specifically in the public sector, empirical contributions on this relationship are therefore important, particularly as empowerment is a very useful tool with which to face the challenges of the new public context. There is also a need to examine the moderating variables involved in this relationship, as well as to extend research on work motivation in public management. It is proposed the study of the effect of individual orientations, such as GO. GO concept refers to the individual disposition toward developing or confirming one’s capacity in achievement situations. Employees’ GO may be a key factor at work and in workforce selection processes, since it explains the differences in personal work interests, and in receptiveness to and interpretations of professional development activities. SE practices could affect PE feelings in different ways, depending on employees’ GO, since they perceive and respond differently to such practices, which is likely to yield distinct PE results. The model is tested on a sample of 521 Spanish local authority employees. Hierarchical regression analysis was conducted to test the research hypotheses using SPSS 22 computer software. The results do not confirm the direct link between SE and PE, but show that learning goal orientation has considerable moderating power in this relationship, and its interaction with SE affects employees’ PE levels. Therefore, the combination of SE practices and employees’ high levels of LGO are important factors for creating psychologically empowered staff in public organizations.

Keywords: goal orientation, moderating effect, psychological empowerment, structural empowerment

Procedia PDF Downloads 265
82 Predicting Reading Comprehension in Spanish: The Evidence for the Simple View Model

Authors: Gabriela Silva-Maceda, Silvia Romero-Contreras

Abstract:

Spanish is a more transparent language than English given that it has more direct correspondences between sounds and letters. It has become important to understand how decoding and linguistic comprehension contribute to reading comprehension in the framework of the widely known Simple View Model. This study aimed to identify the level of prediction by these two components in a sample of 1st to 4th grade children attending two schools in central Mexico (one public and one private). Within each school, ten children were randomly selected in each grade level, and their parents were asked about reading habits and socioeconomic information. In total, 79 children completed three standardized tests measuring decoding (pseudo-word reading), linguistic comprehension (understanding of paragraphs) and reading comprehension using subtests from the Clinical Evaluation of Language Fundamentals-Spanish, Fourth Edition, and the Test de Lectura y Escritura en Español (LEE). The data were analyzed using hierarchical regression, with decoding as a first step and linguistic comprehension as a second step. Results showed that decoding accounted for 19.2% of the variance in reading comprehension, while linguistic comprehension accounted for an additional 10%, adding up to 29.2% of variance explained: F (2, 75)= 15.45, p <.001. Socioeconomic status derived from parental questionnaires showed a statistically significant association with the type of school attended, X2 (3, N= 79) = 14.33, p =.002. Nonetheless when analyzing the Simple View components, only decoding differences were statistically significant (t = -6.92, df = 76.81, p < .001, two-tailed); reading comprehension differences were also significant (t = -3.44, df = 76, p = .001, two-tailed). When socioeconomic status was included in the model, it predicted a 5.9% unique variance, even when already accounting for Simple View components, adding to a 35.1% total variance explained. This three-predictor model was also significant: F (3, 72)= 12.99, p <.001. In addition, socioeconomic status was significantly correlated with the amount of non-textbook books parents reported to have at home for both adults (rho = .61, p<.001) and children (rho= .47, p<.001). Results converge with a large body of literature finding socioeconomic differences in reading comprehension; in addition this study suggests that these differences were also present in decoding skills. Although linguistic comprehension differences between schools were expected, it is argued that the test used to collect this variable was not sensitive to linguistic differences, since it came from a test to diagnose clinical language disabilities. Even with this caveat, results show that the components of the Simple View Model can predict less than a third of the variance in reading comprehension in Spanish. However, the results also suggest that a fuller model of reading comprehension is obtained when considering the family’s socioeconomic status, given the potential differences shown by the socioeconomic status association with books at home, factors that are particularly important in countries where inequality gaps are relatively large.

Keywords: decoding, linguistic comprehension, reading comprehension, simple view model, socioeconomic status, Spanish

Procedia PDF Downloads 312
81 Effect of Ease of Doing Business to Economic Growth among Selected Countries in Asia

Authors: Teodorica G. Ani

Abstract:

Economic activity requires an encouraging regulatory environment and effective rules that are transparent and accessible to all. The World Bank has been publishing the annual Doing Business reports since 2004 to investigate the scope and manner of regulations that enhance business activity and those that constrain it. A streamlined business environment supporting the development of competitive small and medium enterprises (SMEs) may expand employment opportunities and improve the living conditions of low income households. Asia has emerged as one of the most attractive markets in the world. Economies in East Asia and the Pacific were among the most active in making it easier for local firms to do business. The study aimed to describe the ease of doing business and its effect to economic growth among selected economies in Asia for the year 2014. The study covered 29 economies in East Asia, Southeast Asia, South Asia and Middle Asia. Ease of doing business is measured by the Doing Business indicators (DBI) of the World Bank. The indicators cover ten aspects of the ease of doing business such as starting a business, dealing with construction permits, getting electricity, registering property, getting credit, protecting investors, paying taxes, trading across borders, enforcing contracts and resolving insolvency. In the study, Gross Domestic Product (GDP) was used as the proxy variable for economic growth. Descriptive research was the research design used. Graphical analysis was used to describe the income and doing business among selected economies. In addition, multiple regression was used to determine the effect of doing business to economic growth. The study presented the income among selected economies. The graph showed China has the highest income while Maldives produces the lowest and that observation were supported by gathered literatures. The study also presented the status of the ten indicators of doing business among selected economies. The graphs showed varying trends on how easy to start a business, deal with construction permits and to register property. Starting a business is easiest in Singapore followed by Hong Kong. The study found out that the variations in ease of doing business is explained by starting a business, dealing with construction permits and registering property. Moreover, an explanation of the regression result implies that a day increase in the average number of days it takes to complete a procedure will decrease the value of GDP in general. The research proposed inputs to policy which may increase the awareness of local government units of different economies on the simplification of the policies of the different components used in measuring doing business.

Keywords: doing business, economic growth, gross domestic product, Asia

Procedia PDF Downloads 362
80 Family Income and Parental Behavior: Maternal Personality as a Moderator

Authors: Robert H. Bradley, Robert F. Corwyn

Abstract:

There is abundant research showing that socio-economic status is implicated in parenting. However, additional factors such as family context, parent personality, parenting history and child behavior also help determine how parents enact the role of caregiver. Each of these factors not only helps determine how a parent will act in a given situation, but each can serve to moderate the influence of the other factors. Personality has long been studied as a factor that influences parental behavior, but it has almost never been considered as a moderator of family contextual factors. For this study, relations between three maternal personality characteristics (agreeableness, extraversion, neuroticism) and four aspects of parenting (harshness, sensitivity, stimulation, learning materials) were examined when children were 6 months, 36 months, and 54 months old and again at 5th grade. Relations between these three aspects of personality and the overall home environment were also examined. A key concern was whether maternal personality characteristics moderated relations between household income and the four aspects of parenting and between household income and the overall home environment. The data for this study were taken from the NICHD Study of Early Child Care and Youth Development (NICHD SECCYD). The total sample consisted of 1364 families living in ten different sites in the United States. However, the samples analyzed included only those with complete data on all four parenting outcomes (i.e., sensitivity, harshness, stimulation, and provision of learning materials), income, maternal education and all three measures of personality (i.e., agreeableness, neuroticism, extraversion) at each age examined. Results from hierarchical regression analysis showed that mothers high in agreeableness were more likely to demonstrate sensitivity and stimulation as well as provide more learning materials to their children but were less likely to manifest harshness. Maternal agreeableness also consistently moderated the effects of low income on parental behavior. Mothers high in extraversion were more likely to provide stimulation and learning materials, with extraversion serving as a moderator of low income on both. By contrast, mothers high in neuroticism were less likely to demonstrate positive aspects of parenting and more likely to manifest negative aspects (e.g., harshness). Neuroticism also served to moderate the influence of low income on parenting, especially for stimulation and learning materials. The most consistent effects of parent personality were on the overall home environment, with significant main and interaction effects observed in 11 of the 12 models tested. These findings suggest that it may behoove professional who work with parents living in adverse circumstances to consider parental personality in helping to better target prevention or intervention efforts aimed at supporting parental efforts to act in ways that benefit children.

Keywords: home environment, household income, learning materials, personality, sensitivity, stimulation

Procedia PDF Downloads 196
79 Psychopathic Manager Behavior and the Employee Workplace Deviance: The Mediating Role of Revenge Motive, the Moderating Roles of Core Self-Evaluations and Attitude Importance

Authors: Sinem Bulkan

Abstract:

This study introduces the construct of psychopathic manager behaviour and aims for the development of psychopathic manager behaviour (Psycho-Man B) measure. The study also aims for the understanding of the relationship between psychopathic manager behaviour and workplace deviance while investigating the mediating role of a revenge motive and the moderating roles of the core self-evaluations and the attitude importance. Data were collected from 519 employees from a wide variety of jobs and industries who currently work for or previously worked for a manager in a collectivist culture, Turkey. Psycho-Man B Measure was developed resulting in five dimensions as opposed to the proposed ten dimensions. Simple linear and hierarchical regression analyses were conducted to test the hypotheses. The results of simple linear regression analyses showed that psychopathic manager behaviour was positively significantly related to supervisor-directed and organisation-directed deviance. Revenge motive towards the manager partially mediated the relationship between psychopathic manager behaviour and supervisor-directed deviance. Similarly, revenge motive towards the organisation partially mediated the relationship between psychopathic manager behaviour and organisation-directed deviance. Furthermore, no support was found for the expected moderating role of core self-evaluations in the revenge motive towards the manager-supervisor-directed deviance and revenge motive towards the organisation-organisation-directed deviance relationships. Attitude importance moderated the relationship between revenge motive towards the manager and supervisor-directed deviance; revenge motive towards the organisation and organisation-directed deviance. Moderated-mediation hypotheses were not supported for core self-evaluations but were supported for the attitude importance. Additional analyses for sub-dimensions were conducted to further examine the hypotheses. Demographic variables were examined through independent samples t-tests, and one way ANOVA. Finally, findings are discussed; limitations, suggestions and implications are presented. The major contribution of this study is that ‘psychopathic manager behaviour’ construct was introduced to the literature and a scale for the reliable identification of psychopathic manager behaviour was developed in to evaluate managers’ level of sub-clinical psychopathy in the workforce. The study introduced that employees engage in different forms of supervisor-directed deviance and organisation-directed deviance depending on the level of the emotions and personal goals. Supervisor-directed deviant behaviours and organisation-directed deviant behaviours became distinct in a way as impulsive and premeditated, active or passive, direct and indirect actions. Accordingly, it is important for organisations to notice that employees’ level of affective state and attitude importance for psychopathic manager behaviours predetermine the certain type of employee deviant behaviours.

Keywords: attitude importance, core self evaluations, psychopathic manager behaviour, revenge motive, workplace deviance

Procedia PDF Downloads 250
78 Analysis of Epileptic Electroencephalogram Using Detrended Fluctuation and Recurrence Plots

Authors: Mrinalini Ranjan, Sudheesh Chethil

Abstract:

Epilepsy is a common neurological disorder characterised by the recurrence of seizures. Electroencephalogram (EEG) signals are complex biomedical signals which exhibit nonlinear and nonstationary behavior. We use two methods 1) Detrended Fluctuation Analysis (DFA) and 2) Recurrence Plots (RP) to capture this complex behavior of EEG signals. DFA considers fluctuation from local linear trends. Scale invariance of these signals is well captured in the multifractal characterisation using detrended fluctuation analysis (DFA). Analysis of long-range correlations is vital for understanding the dynamics of EEG signals. Correlation properties in the EEG signal are quantified by the calculation of a scaling exponent. We report the existence of two scaling behaviours in the epileptic EEG signals which quantify short and long-range correlations. To illustrate this, we perform DFA on extant ictal (seizure) and interictal (seizure free) datasets of different patients in different channels. We compute the short term and long scaling exponents and report a decrease in short range scaling exponent during seizure as compared to pre-seizure and a subsequent increase during post-seizure period, while the long-term scaling exponent shows an increase during seizure activity. Our calculation of long-term scaling exponent yields a value between 0.5 and 1, thus pointing to power law behaviour of long-range temporal correlations (LRTC). We perform this analysis for multiple channels and report similar behaviour. We find an increase in the long-term scaling exponent during seizure in all channels, which we attribute to an increase in persistent LRTC during seizure. The magnitude of the scaling exponent and its distribution in different channels can help in better identification of areas in brain most affected during seizure activity. The nature of epileptic seizures varies from patient-to-patient. To illustrate this, we report an increase in long-term scaling exponent for some patients which is also complemented by the recurrence plots (RP). RP is a graph that shows the time index of recurrence of a dynamical state. We perform Recurrence Quantitative analysis (RQA) and calculate RQA parameters like diagonal length, entropy, recurrence, determinism, etc. for ictal and interictal datasets. We find that the RQA parameters increase during seizure activity, indicating a transition. We observe that RQA parameters are higher during seizure period as compared to post seizure values, whereas for some patients post seizure values exceeded those during seizure. We attribute this to varying nature of seizure in different patients indicating a different route or mechanism during the transition. Our results can help in better understanding of the characterisation of epileptic EEG signals from a nonlinear analysis.

Keywords: detrended fluctuation, epilepsy, long range correlations, recurrence plots

Procedia PDF Downloads 162
77 Assessment of Urban Environmental Noise in Urban Habitat: A Spatial Temporal Study

Authors: Neha Pranav Kolhe, Harithapriya Vijaye, Arushi Kamle

Abstract:

The economic growth engines are urban regions. As the economy expands, so does the need for peace and quiet, and noise pollution is one of the important social and environmental issue. Health and wellbeing are at risk from environmental noise pollution. Because of urbanisation, population growth, and the consequent rise in the usage of increasingly potent, diverse, and highly mobile sources of noise, it is now more severe and pervasive than ever before, and it will only become worse. Additionally, it will expand as long as there is an increase in air, train, and highway traffic, which continue to be the main contributors of noise pollution. The current study will be conducted in two zones of class I city of central India (population range: 1 million–4 million). Total 56 measuring points were chosen to assess noise pollution. The first objective evaluates the noise pollution in various urban habitats determined as formal and informal settlement. It identifies the comparison of noise pollution within the settlements using T- Test analysis. The second objective assess the noise pollution in silent zones (as stated in Central Pollution Control Board) in a hierarchical way. It also assesses the noise pollution in the settlements and compares with prescribed permissible limits using class I sound level equipment. As appropriate indices, equivalent noise level on the (A) frequency weighting network, minimum sound pressure level and maximum sound pressure level were computed. The survey is conducted for a period of 1 week. Arc GIS is used to plot and map the temporal and spatial variability in urban settings. It is discovered that noise levels at most stations, particularly at heavily trafficked crossroads and subway stations, were significantly different and higher than acceptable limits and squares. The study highlights the vulnerable areas that should be considered while city planning. The study demands area level planning while preparing a development plan. It also demands attention to noise pollution from the perspective of residential and silent zones. The city planning in urban areas neglects the noise pollution assessment at city level. This contributes to that, irrespective of noise pollution guidelines, the ground reality is far away from its applicability. The result produces incompatible land use on a neighbourhood scale with respect to noise pollution. The study's final results will be useful to policymakers, architects and administrators in developing countries. This will be useful for noise pollution in urban habitat governance by efficient decision making and policy formulation to increase the profitability of these systems.

Keywords: noise pollution, formal settlements, informal settlements, built environment, silent zone, residential area

Procedia PDF Downloads 106
76 High Performance Computing Enhancement of Agent-Based Economic Models

Authors: Amit Gill, Lalith Wijerathne, Sebastian Poledna

Abstract:

This research presents the details of the implementation of high performance computing (HPC) extension of agent-based economic models (ABEMs) to simulate hundreds of millions of heterogeneous agents. ABEMs offer an alternative approach to study the economy as a dynamic system of interacting heterogeneous agents, and are gaining popularity as an alternative to standard economic models. Over the last decade, ABEMs have been increasingly applied to study various problems related to monetary policy, bank regulations, etc. When it comes to predicting the effects of local economic disruptions, like major disasters, changes in policies, exogenous shocks, etc., on the economy of the country or the region, it is pertinent to study how the disruptions cascade through every single economic entity affecting its decisions and interactions, and eventually affect the economic macro parameters. However, such simulations with hundreds of millions of agents are hindered by the lack of HPC enhanced ABEMs. In order to address this, a scalable Distributed Memory Parallel (DMP) implementation of ABEMs has been developed using message passing interface (MPI). A balanced distribution of computational load among MPI-processes (i.e. CPU cores) of computer clusters while taking all the interactions among agents into account is a major challenge for scalable DMP implementations. Economic agents interact on several random graphs, some of which are centralized (e.g. credit networks, etc.) whereas others are dense with random links (e.g. consumption markets, etc.). The agents are partitioned into mutually-exclusive subsets based on a representative employer-employee interaction graph, while the remaining graphs are made available at a minimum communication cost. To minimize the number of communications among MPI processes, real-life solutions like the introduction of recruitment agencies, sales outlets, local banks, and local branches of government in each MPI-process, are adopted. Efficient communication among MPI-processes is achieved by combining MPI derived data types with the new features of the latest MPI functions. Most of the communications are overlapped with computations, thereby significantly reducing the communication overhead. The current implementation is capable of simulating a small open economy. As an example, a single time step of a 1:1 scale model of Austria (i.e. about 9 million inhabitants and 600,000 businesses) can be simulated in 15 seconds. The implementation is further being enhanced to simulate 1:1 model of Euro-zone (i.e. 322 million agents).

Keywords: agent-based economic model, high performance computing, MPI-communication, MPI-process

Procedia PDF Downloads 113
75 Insights into Child Malnutrition Dynamics with the Lens of Women’s Empowerment in India

Authors: Bharti Singh, Shri K. Singh

Abstract:

Child malnutrition is a multifaceted issue that transcends geographical boundaries. Malnutrition not only stunts physical growth but also leads to a spectrum of morbidities and child mortality. It is one of the leading causes of death (~50 %) among children under age five. Despite economic progress and advancements in healthcare, child malnutrition remains a formidable challenge for India. The objective is to investigate the impact of women's empowerment on child nutrition outcomes in India from 2006 to 2021. A composite index of women's empowerment was constructed using Confirmatory Factor Analysis (CFA), a rigorous technique that validates the measurement model by assessing how well-observed variables represent latent constructs. This approach ensures the reliability and validity of the empowerment index. Secondly, kernel density plots were utilised to visualise the distribution of key nutritional indicators, such as stunting, wasting, and overweight. These plots offer insights into the shape and spread of data distributions, aiding in understanding the prevalence and severity of malnutrition. Thirdly, linear polynomial graphs were employed to analyse how nutritional parameters evolved with the child's age. This technique enables the visualisation of trends and patterns over time, allowing for a deeper understanding of nutritional dynamics during different stages of childhood. Lastly, multilevel analysis was conducted to identify vulnerable levels, including State-level, PSU-level, and household-level factors impacting undernutrition. This approach accounts for hierarchical data structures and allows for the examination of factors at multiple levels, providing a comprehensive understanding of the determinants of child malnutrition. Overall, the utilisation of these statistical methodologies enhances the transparency and replicability of the study by providing clear and robust analytical frameworks for data analysis and interpretation. Our study reveals that NFHS-4 and NFHS-5 exhibit an equal density of severely stunted cases. NFHS-5 indicates a limited decline in wasting among children aged five, while the density of severely wasted children remains consistent across NFHS-3, 4, and 5. In 2019-21, women with higher empowerment had a lower risk of their children being undernourished (Regression coefficient= -0.10***; Confidence Interval [-0.18, -0.04]). Gender dynamics also play a significant role, with male children exhibiting a higher susceptibility to undernourishment. Multilevel analysis suggests household-level vulnerability (intra-class correlation=0.21), highlighting the need to address child undernutrition at the household level.

Keywords: child nutrition, India, NFHS, women’s empowerment

Procedia PDF Downloads 23
74 Social Perspective of Gender Biasness Among Rural Children in Haryna State of India

Authors: Kamaljeet Kaur, Vinod Kumari, Jatesh Kathpalia, Bas Kaur

Abstract:

A gender bias towards girl child is pervasive across the world. It is seen in all the strata of the society and manifests in various forms. However nature and extent of these inequalities are not uniform. Generally these inequalities are more prevalent in patriarchal society. Despite emerging and increasing opportunities for women, there are still inequalities between men and women in each and every sphere like education, health, economy, polity and social sphere. Patriarchal ideology as a cultural norm enforces gender construction which is oriented toward hierarchical relations between the sexes and neglect of women in Indian society. Discrimination to girls may also vary by their age and be restricted to the birth order and sex composition of her elder surviving siblings. The present study was conducted to know the gender discrimination among rural children in India. The respondents were selected from three generations as per AICRP age group viz, 18-30 years (3rd generation), 31-60 years (2nd generation) and above 60 years (1st generation). A total sample size was 600 respondents from different villages of two districts of Haryana state comprising of half males and half females. Data were collected using personal interview schedule and analysed by SPSS software. Among the total births 46.35 per cent were girl child and 53.64 % were male child. Dropout rate was more in female children as compared to male children i.e. near about one third (31.09%) female children dropped school followed by 21.17 % male children. It was quite surprising that near about two-third (61.16%) female children and more than half (59.22%) of the male children dropped school. Cooking was mainly performed by adult female with overall mean scores 2.0 and ranked first which was followed by female child (1.7 mean scores) clearly indicating that cooking was the activity performed mainly by females while activity related to purchase of fruits and vegetable, cereals and pulses was mainly done by adult male. First preference was given to male child for serving of costly and special food. Regarding professional aspiration of children of the respondents’ families, it was observed that 20.10% of the male children wanted to become engineer, whereas only 3.89 % female children wanted to become engineer. Ratio of male children was high in both generations irrespective of the districts. School dropouts were more in case of female in both the 1st and 2 nd generations. The main reasons of school dropout were lack of interest, lack of resources and early marriage in both the generations. Female enrolment was more in faculty of arts, whereas in case of male percentage it was more in faculty of non-medical and medical which showed that female children were getting traditional type of education. It is suggested to provide equal opportunities to girls and boys in home as well as outside the home for smooth functioning of society.

Keywords: gender biasness, male child, female child, education, home

Procedia PDF Downloads 75
73 Reconstruction of Alveolar Bone Defects Using Bone Morphogenetic Protein 2 Mediated Rabbit Dental Pulp Stem Cells Seeded on Nano-Hydroxyapatite/Collagen/Poly(L-Lactide)

Authors: Ling-Ling E., Hong-Chen Liu, Dong-Sheng Wang, Fang Su, Xia Wu, Zhan-Ping Shi, Yan Lv, Jia-Zhu Wang

Abstract:

Objective: The objective of the present study is to evaluate the capacity of a tissue-engineered bone complex of recombinant human bone morphogenetic protein 2 (rhBMP-2) mediated dental pulp stem cells (DPSCs) and nano-hydroxyapatite/collagen/poly(L-lactide)(nHAC/PLA) to reconstruct critical-size alveolar bone defects in New Zealand rabbit. Methods: Autologous DPSCs were isolated from rabbit dental pulp tissue and expanded ex vivo to enrich DPSCs numbers, and then their attachment and differentiation capability were evaluated when cultured on the culture plate or nHAC/PLA. The alveolar bone defects were treated with nHAC/PLA, nHAC/PLA+rhBMP-2, nHAC/PLA+DPSCs, nHAC/PLA+DPSCs+rhBMP-2, and autogenous bone (AB) obtained from iliac bone or were left untreated as a control. X-ray and a polychrome sequential fluorescent labeling were performed post-operatively and the animals were sacrificed 12 weeks after operation for histological observation and histomorphometric analysis. Results: Our results showed that DPSCs expressed STRO-1 and vementin, and favoured osteogenesis and adipogenesis in conditioned media. DPSCs attached and spread well, and retained their osteogenic phenotypes on nHAC/PLA. The rhBMP-2 could significantly increase protein content, alkaline phosphatase (ALP) activity/protein, osteocalcin (OCN) content, and mineral formation of DPSCs cultured on nHAC/PLA. The X-ray graph, the fluorescent, histological observation and histomorphometric analysis showed that the nHAC/PLA+DPSCs+rhBMP-2 tissue-engineered bone complex had an earlier mineralization and more bone formation inside the scaffold than nHAC/PLA, nHAC/PLA+rhBMP-2 and nHAC/PLA+DPSCs, or even autologous bone. Implanted DPSCs contribution to new bone were detected through transfected eGFP genes. Conclutions: Our findings indicated that stem cells existed in adult rabbit dental pulp tissue. The rhBMP-2 promoted osteogenic capability of DPSCs as a potential cell source for periodontal bone regeneration. The nHAC/PLA could serve as a good scaffold for autologous DPSCs seeding, proliferation and differentiation. The tissue-engineered bone complex with nHAC/PLA, rhBMP-2, and autologous DPSCs might be a better alternative to autologous bone for the clinical reconstruction of periodontal bone defects.

Keywords: nano-hydroxyapatite/collagen/poly (L-lactide), dental pulp stem cell, recombinant human bone morphogenetic protein, bone tissue engineering, alveolar bone

Procedia PDF Downloads 383
72 Novel Framework for MIMO-Enhanced Robust Selection of Critical Control Factors in Auto Plastic Injection Moulding Quality Optimization

Authors: Seyed Esmail Seyedi Bariran, Khairul Salleh Mohamed Sahari

Abstract:

Apparent quality defects such as warpage, shrinkage, weld line, etc. are such an irresistible phenomenon in mass production of auto plastic appearance parts. These frequently occurred manufacturing defects should be satisfied concurrently so as to achieve a final product with acceptable quality standards. Determining the significant control factors that simultaneously affect multiple quality characteristics can significantly improve the optimization results by eliminating the deviating effect of the so-called ineffective outliers. Hence, a robust quantitative approach needs to be developed upon which major control factors and their level can be effectively determined to help improve the reliability of the optimal processing parameter design. Hence, the primary objective of current study was to develop a systematic methodology for selection of significant control factors (SCF) relevant to multiple quality optimization of auto plastic appearance part. Auto bumper was used as a specimen with the most identical quality and production characteristics to APAP group. A preliminary failure modes and effect analysis (FMEA) was conducted to nominate a database of pseudo significant significant control factors prior to the optimization phase. Later, CAE simulation Moldflow analysis was implemented to manipulate four rampant plastic injection quality defects concerned with APAP group including warpage deflection, volumetric shrinkage, sink mark and weld line. Furthermore, a step-backward elimination searching method (SESME) has been developed for systematic pre-optimization selection of SCF based on hierarchical orthogonal array design and priority-based one-way analysis of variance (ANOVA). The development of robust parameter design in the second phase was based on DOE module powered by Minitab v.16 statistical software. Based on the F-test (F 0.05, 2, 14) one-way ANOVA results, it was concluded that for warpage deflection, material mixture percentage was the most significant control factor yielding a 58.34% of contribution while for the other three quality defects, melt temperature was the most significant control factor with a 25.32%, 84.25%, and 34.57% contribution for sin mark, shrinkage and weld line strength control. Also, the results on the he least significant control factors meaningfully revealed injection fill time as the least significant factor for both warpage and sink mark with respective 1.69% and 6.12% contribution. On the other hand, for shrinkage and weld line defects, the least significant control factors were holding pressure and mold temperature with a 0.23% and 4.05% overall contribution accordingly.

Keywords: plastic injection moulding, quality optimization, FMEA, ANOVA, SESME, APAP

Procedia PDF Downloads 333
71 Metacognitive Processing in Early Readers: The Role of Metacognition in Monitoring Linguistic and Non-Linguistic Performance and Regulating Students' Learning

Authors: Ioanna Taouki, Marie Lallier, David Soto

Abstract:

Metacognition refers to the capacity to reflect upon our own cognitive processes. Although there is an ongoing discussion in the literature on the role of metacognition in learning and academic achievement, little is known about its neurodevelopmental trajectories in early childhood, when children begin to receive formal education in reading. Here, we evaluate the metacognitive ability, estimated under a recently developed Signal Detection Theory model, of a cohort of children aged between 6 and 7 (N=60), who performed three two-alternative-forced-choice tasks (two linguistic: lexical decision task, visual attention span task, and one non-linguistic: emotion recognition task) including trial-by-trial confidence judgements. Our study has three aims. First, we investigated how metacognitive ability (i.e., how confidence ratings track accuracy in the task) relates to performance in general standardized tasks related to students' reading and general cognitive abilities using Spearman's and Bayesian correlation analysis. Second, we assessed whether or not young children recruit common mechanisms supporting metacognition across the different task domains or whether there is evidence for domain-specific metacognition at this early stage of development. This was done by examining correlations in metacognitive measures across different task domains and evaluating cross-task covariance by applying a hierarchical Bayesian model. Third, using robust linear regression and Bayesian regression models, we assessed whether metacognitive ability in this early stage is related to the longitudinal learning of children in a linguistic and a non-linguistic task. Notably, we did not observe any association between students’ reading skills and metacognitive processing in this early stage of reading acquisition. Some evidence consistent with domain-general metacognition was found, with significant positive correlations between metacognitive efficiency between lexical and emotion recognition tasks and substantial covariance indicated by the Bayesian model. However, no reliable correlations were found between metacognitive performance in the visual attention span and the remaining tasks. Remarkably, metacognitive ability significantly predicted children's learning in linguistic and non-linguistic domains a year later. These results suggest that metacognitive skill may be dissociated to some extent from general (i.e., language and attention) abilities and further stress the importance of creating educational programs that foster students’ metacognitive ability as a tool for long term learning. More research is crucial to understand whether these programs can enhance metacognitive ability as a transferable skill across distinct domains or whether unique domains should be targeted separately.

Keywords: confidence ratings, development, metacognitive efficiency, reading acquisition

Procedia PDF Downloads 135