Search results for: analytical validation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3537

Search results for: analytical validation

297 New Derivatives 7-(diethylamino)quinolin-2-(1H)-one Based Chalcone Colorimetric Probes for Detection of Bisulfite Anion in Cationic Micellar Media

Authors: Guillermo E. Quintero, Edwin G. Perez, Oriel Sanchez, Christian Espinosa-Bustos, Denis Fuentealba, Margarita E. Aliaga

Abstract:

Bisulfite ion (HSO3-) has been used as a preservative in food, drinks, and medication. However, it is well-known that HSO3- can cause health problems like asthma and allergic reactions in people. Due to the above, the development of analytical methods for detecting this ion has gained great interest. In line with the above, the current use of colorimetric and/or fluorescent probes as a detection technique has acquired great relevance due to their high sensitivity and accuracy. In this context, 2-quinolinone derivatives have been found to possess promising activity as antiviral agents, sensitizers in solar cells, antifungals, antioxidants, and sensors. In particular, 7-(diethylamino)-2-quinolinone derivatives have attracted attention in recent years since their suitable photophysical properties become promising fluorescent probes. In Addition, there is evidence that photophysical properties and reactivity can be affected by the study medium, such as micellar media. Based on the above background, 7-(diethylamino)-2-quinolinone derivatives based chalcone will be able to be incorporated into a cationic micellar environment (Cetyltrimethylammonium bromide, CTAB). Furthermore, the supramolecular control induced by the micellar environment will increase the reactivity of these derivatives towards nucleophilic analytes such as HSO3- (Michael-type addition reaction), leading to the generation of new colorimetric and/or fluorescent probes. In the present study, two derivatives of 7-(diethylamino)-2-quinolinone based chalcone DQD1-2 were synthesized according to the method reported by the literature. These derivatives were structurally characterized by 1H, 13C NMR, and HRMS-ESI. In addition, UV-VIS and fluorescence studies determined absorption bands near 450 nm, emission bands near 600 nm, fluorescence quantum yields near 0.01, and fluorescence lifetimes of 5 ps. In line with the foregoing, these photophysical properties aforementioned were improved in the presence of a cationic micellar medium using CTAB thanks to the formation of adducts presenting association constants of the order of 2,5x105 M-1, increasing the quantum yields to 0.12 and the fluorescence lifetimes corresponding to two lifetimes near to 120 and 400 ps for DQD1 and DQD2. Besides, thanks to the presence of the micellar medium, the reactivity of these derivatives with nucleophilic analytes, such as HSO3-, was increased. This was achieved through kinetic studies, which demonstrated an increase in the bimolecular rate constants in the presence of a micellar medium. Finally, probe DQD1 was chosen as the best sensor since it was assessed to detect HSO3- with excellent results.

Keywords: bisulfite detection, cationic micelle, colorimetric probes, quinolinone derivatives

Procedia PDF Downloads 64
296 An Interpretative Historical Analysis of Asylum and Refugee Policies and Attitudes to Australian Immigration Laws

Authors: Kamal Kithsiri Karunadasa Hewawasam Revulge

Abstract:

This paper is an interpretative historical analysis of Australian migration laws that examines asylum and refugee policies and attitudes in Australia. It looks at major turning points in Australian migration history, and in doing so, the researcher reviewed relevant literature on the aspects crucial to highlighting the current trend of Australian migration policies. The data was collected using secondary data from official government sources, including annual reports, media releases on immigration, inquiry reports, statistical information, and other available literature to identify critical historical events that significantly affected the systematic developments of asylum seekers and refugee policies in Australia and to look at the historical trends of official thinking. A reliance on using these official sources is justified as those are the most convincing sources to analyse the historical events in Australia. Additional literature provides us with critical analyses of the behaviour and culture of the Australian immigration administration. The analytical framework reviewed key Australian Government immigration policies since British colonization and the settlement era of 1787–the 1850s and to the present. The fundamental basis for doing so is that past events and incidents offer us clues and lessons relevant to the present day. Therefore, providing a perspective on migration history in Australia helps analyse how current policymakers' strategies developed and changed over time. Attention is also explicitly focused on Australian asylum and refugee policy internationally, as it helped to broaden the analysis. The finding proved a link between past events and adverse current Australian government policies towards asylum seekers and refugees. It highlighted that Australia's current migration policies are part of a carefully and deliberately planned pattern that arose from the occupation of Australia by early British settlers. In this context, the remarkable point is that the historical events of taking away children from their Australian indigenous parents, widely known as the 'stolen generation' reflected a model of assimilation, or a desire to absorb other cultures into Australian society by fully adopting the settlers' language, their culture, and losing indigenous people's traditions. Current Australian policies towards migrants reflect the same attitude. Hence, it could be argued that policies and attitudes towards asylum seekers and refugees, particularly so-called 'boat people' to some extent, still reflect Australia's earlier colonial and 'white Australia' history.

Keywords: migration law, refugee law, international law, administrative law

Procedia PDF Downloads 54
295 Molecular Detection of E. coli in Treated Wastewater and Well Water Samples Collected from Al Riyadh Governorate, Saudi Arabia

Authors: Hanouf A. S. Al Nuwaysir, Nadine Moubayed, Abir Ben Bacha, Islem Abid

Abstract:

Consumption of waste water continues to cause significant problems for human health in both developed and developing countries. Many regulations have been implied by different world authorities controlling water quality for the presence of coliforms used as standard indicators of water quality deterioration and historically leading health protection concept. In this study, the European directive for the detection of Escherichia coli, ISO 9308-1, was applied to examine and monitor coliforms in water samples collected from Wadi Hanifa and neighboring wells, Riyadh governorate, kingdom of Saudi Arabia, which is used for irrigation and industrial purposes. Samples were taken from different locations for 8 months consecutively, chlorine concentration ranging from 0.1- 0.4 mg/l, was determined using the DPD FREE CHLORINE HACH kit. Water samples were then analyzed following the ISO protocol which relies on the membrane filtration technique (0.45µm, pore size membrane filter) and a chromogenic medium TTC, a lactose based medium used for the detection and enumeration of total coliforms and E.coli. Data showed that the number of bacterial isolates ranged from 60 to 300 colonies/100ml for well and surface water samples respectively; where higher numbers were attributed to the surface samples. Organisms which apparently ferment lactose on TTC agar plates, appearing as orange colonies, were selected and additionally cultured on EMB and MacConkey agar for a further differentiation among E.coli and coliform bacteria. Two additional biochemical tests (Cytochrome oxidase and indole from tryptophan) were also investigated to detect and differentiate the presence of E.coli from other coliforms, E. coli was identified in an average of 5 to 7colonies among 25 selected colonies.On the other hand, a more rapid, specific and sensitive analytical molecular detection namely single colony PCR was also performed targeting hha gene to sensitively detect E.coli, giving more accurate and time consuming identification of colonies considered presumptively as E.coli. Comparative methodologies, such as ultrafiltration and direct DNA extraction from membrane filters (MoBio, Grermany) were also applied; however, results were not as accurate as the membrane filtration, making it a technique of choice for the detection and enumeration of water coliforms, followed by sufficiently specific enzymatic confirmatory stage.

Keywords: coliform, cytochrome oxidase, hha primer, membrane filtration, single colony PCR

Procedia PDF Downloads 297
294 Accomplishing Mathematical Tasks in Bilingual Primary Classrooms

Authors: Gabriela Steffen

Abstract:

Learning in a bilingual classroom not only implies learning in two languages or in an L2, it also means learning content subjects through the means of bilingual or plurilingual resources, which is of a qualitatively different nature than ‘monolingual’ learning. These resources form elements of a didactics of plurilingualism, aiming not only at the development of a plurilingual competence, but also at drawing on plurilingual resources for nonlinguistic subject learning. Applying a didactics of plurilingualism allows for taking account of the specificities of bilingual content subject learning in bilingual education classrooms. Bilingual education is used here as an umbrella term for different programs, such as bilingual education, immersion, CLIL, bilingual modules in which one or several non-linguistic subjects are taught partly or completely in an L2. This paper aims at discussing first results of a study on pupil group work in bilingual classrooms in several Swiss primary schools. For instance, it analyses two bilingual classes in two primary schools in a French-speaking region of Switzerland that follows a part of their school program through German in addition to French, the language of instruction in this region. More precisely, it analyses videotaped classroom interaction and in situ classroom practices of pupil group work in a mathematics lessons. The ethnographic observation of pupils’ group work and the analysis of their interaction (analytical tools of conversational analysis, discourse analysis and plurilingual interaction) enhance the description of whole-class interaction done in the same (and several other) classes. While the latter are teacher-student interactions, the former are student-student interactions giving more space to and insight into pupils’ talk. This study aims at the description of the linguistic and multimodal resources (in German L2 and/or French L1) pupils mobilize while carrying out a mathematical task. The analysis shows that the accomplishment of the mathematical task takes place in a bilingual mode, whether the whole-class interactions are conducted rather in a bilingual (German L2-French L1) or a monolingual mode in L2 (German). The pupils make plenty of use of German L2 in a setting that lends itself to use French L1 (peer groups with French as a dominant language, in absence of the teacher and a task with a mathematical aim). They switch from French to German and back ‘naturally’, which is regular for bilingual speakers. Their linguistic resources in German L2 are not sufficient to allow them to (inter-)act well enough to accomplish the task entirely in German L2, despite their efforts to do so. However, this does not stop them from carrying out the task in mathematics adequately, which is the main objective, by drawing on the bilingual resources at hand.

Keywords: bilingual content subject learning, bilingual primary education, bilingual pupil group work, bilingual teaching/learning resources, didactics of plurilingualism

Procedia PDF Downloads 139
293 Outcome of Dacryocystorhinostomy with Peroperative Local Use of Mitomycin-C

Authors: Chandra Shekhar Majumder, Orin Sultana Jamie

Abstract:

Background: Dacryocystorhinostomy (DCR) has been a widely accepted surgical intervention for nasolacrimal duct obstructions. Some previous studies demonstrated the potential benefits of the peroperative application of agents like Mitomycin-C (MMC) with DCR to improve surgical outcomes. Relevant studies are rare in Bangladesh, and there are controversies about the dose, duration of MMC, and outcome. Therefore, the present study aimed to investigate the comparative efficacy of DCR with and without MMC in a tertiary hospital in Bangladesh. Objective: The study aims to determine the outcome of a dacryocystorhinostomy with preoperative local use of mitomycin–C. Methods: An analytical study was conducted in the Department of Ophthalmology, Sir Salimullah Medical College & Mitford Hospital, Dhaka, from January 2023 to September 2023. Seventy patients who were admitted for DCR operation were included according to the inclusion and exclusion criteria. Patients were divided into two groups: those who underwent DCR with peroperative administration of 0.2 mg/ml Mitomycin-C for 5 minutes (Group I) and those who underwent DCR alone (Group II). All patients were subjected to detailed history taking, clinical examination, and relevant investigations. All patients underwent DCR according to standard guidelines and ensured the highest peroperative and postoperative care. Then, patients were followed up at 7th POD, 1-month POD, 3 months POD, and 6 months POD to observe the success rate between the two groups by assessing tearing condition, irrigation, height of tear meniscus, and FDDT- test. Data was recorded using a pre-structured questionnaire, and collected data were analyzed using SPSS 23. Results: The mean age of the study patients was 42.17±6.7 (SD) years and 42.29±7.1 (SD) years in Groups I and II, respectively, with no significant difference (p=0.945). At the 6th month’s follow-up, group I patients were observed with 94.3% frequency of symptom-free, 85.6% patency of lacrimal drainage system, 68.6% had tear meniscus <0.1mm and 88.6% had positive Fluorescence Dye Disappearance Test (FDDT test). In group II, 91.4% were symptom-free, 68.6% showed patency, 57.1% had a height of tear meniscus < 0.1 mm, and 85.6% had FDDT test positive. But no statistically significant difference was observed (p<.05). Conclusion: The use of Mitomycin-C preoperatively during DCR offers better postoperative outcomes, particularly in maintaining patency and achieving symptom resolution with more FDDT test positive and improvement of tear meniscus in the MMC group than the control group. However, this study didn’t demonstrate a statistically significant difference between the two groups. Further research with larger sample sizes and longer follow-up periods would be beneficial to corroborate these findings.

Keywords: dacryocystorhinostomy, mitomycin-c, dacryocystitis, nasolacrimal duct obstruction

Procedia PDF Downloads 28
292 The Position of Islamic Jurisprudence in UAE Private Law: Analytical Study

Authors: Iyad Jadalhaq, Mohammed El Hadi El Maknouzi

Abstract:

The place of Islamic law in the legal system of the UAE is best understood by introducing a differentiation between its role as a formal source of law and its influence as a material source of law. What this differentiation helps clarify is that the corpus of Islamic law constitutes a much deeper influence on adjudication, law-making and the legal profession in the UAE, than it might appear at first sight, by considering its formal position in the division of labor between courts, or legislative lists of sources of law. This paper aims to examine the role of Shariah in the UAE private law system by determining the comprehensiveness of Sharia in the legal system as a whole, and not in a limited way related to it as a source of law according to Article 1 of the Civil Transactions Law. Turning to the role of the Shariah as a formal source of law, it is useful to start from Article 1 of the UAE Civil Code. This provision lays out the formal hierarchy of sources of UAE private law, these being legislation, Islamic law, and custom. Hence, when deciding a civil dispute, a judge should first refer to positive legislation in force in the UAE. Lacking the rule to cover the case before him/her, the judge ought then to refer directly to Islamic law. If the matter lacks regulation in Islamic law, only then may the judge appeal to custom. Accordingly, in connection to civil transactions, Shariah is presented here, formally, as the second source of law. Still, Shariah law addresses many other issues beyond civil transactions, including matters of morals, worship, and belief. However, in Article 1 of the UAE Civil Code, the reference to Islamic law ought to be understood as limited to the rules it lays out for civil transactions. There are four main sets of courts in the judicial systems of the UAE, whose competence is based on whether a dispute touches upon civil and commercial transactions, criminal offenses, personal statuses, or labor relations. This sectorial and multi-tiered organization of courts as a whole constitutes an institutional development compatible with the long-standing affirmation in the Shariah of the legitimacy of the judiciary. Indeed, Islamic law authorizes the governing authorities to organize the judiciary, including by allocating specific types of cases to particular kinds of judges depending on the value of the case, or by assigning judges to a specific place in which they are to exercise their jurisdictional function. In view of this, the contemporary organization of courts in the UAE can be regarded as an organic adaptation, aligned with Shariah rules on the assignment of jurisdictional authority, to the growing complexity of modern society. Therefore, we can conclude to the comprehensive role of Shariah in the entire legal system of the United Arab Emirates, including legislation, a judicial system, institutional, and administrative work.

Keywords: Islamic jurisprudence, Shariah, UAE civil code, UAE private law

Procedia PDF Downloads 100
291 Synthesis and Analytical Characterisation of Polymer-Silica Nanoparticles Composite for the Protection and Preservation of Stone Monuments

Authors: Sayed M. Ahmed, Sawsan S. Darwish, Nagib A. Elmarzugi, Mohammad A. Al-Dosari, Mahmoud A. Adam, Nadia A. Al-Mouallimi

Abstract:

Historical stone surfaces and architectural heritage may undergo unwanted changes due to the exposure to many physical and chemical deterioration factors, the innovative properties of the nano - materials can have advantageous application in the restoration and conservation of the cultural heritage with relation to the tailoring of new products for protection and consolidation of stone. The current work evaluates the effectiveness of inorganic compatible treatments; based on nanosized particles of silica (SiO2) dispersed in silicon based product, commonly used as a water-repellent/ consolidation for the construction materials affected by different kinds of decay. The nanocomposites obtained by dispersing the silica nanoparticles in polymeric matrices SILRES® BS OH 100 (solventless mixtures of ethyl silicates), in order to obtain a new nanocomposite, with hydrophobic and consolidation properties, to improve the physical and mechanical properties of the stone material. The nanocomposites obtained and pure SILRES® BS OH 100 were applied by brush Experimental stone blocks. The efficacy of the treatments has been evaluated after consolidation and artificial Thermal aging, through capillary water absorption measurements, Ultraviolet-light exposure to evaluate photo-induced and the hydrophobic effects of the treated surface, Scanning electron microscopy (SEM) examination is performed to evaluate penetration depth, re-aggregating effects of the deposited phase and the surface morphology before and after artificialaging. Sterio microscopy investigation is performed to evaluate the resistant to the effects of the erosion, acids and salts. Improving of stone mechanical properties were evaluated by compressive strength tests, colorimetric measurements were used to evaluate the optical appearance. All the results get together with the apparent effect that, silica/polymer nanocomposite is efficient material for the consolidation of artistic and architectural sandstone monuments, completely compatible, enhanced the durability of sandstone toward thermal and UV aging. In addition, the obtained nanocomposite improved the stone mechanical properties and the resistant to the effects of the erosion, acids and salts compared to the samples treated with pure SILRES® BS OH 100 without silica nanoparticles.

Keywords: colorimetric measurements, compressive strength, nanocomposites, porous stone consolidation, silica nanoparticles, sandstone

Procedia PDF Downloads 231
290 Media, Myth and Hero: Sacred Political Narrative in Semiotic and Anthropological Analysis

Authors: Guilherme Oliveira

Abstract:

The assimilation of images and their potential symbolism into lived experiences is inherent. It is through this exercise of recognition via imagistic records that the questioning of the origins of a constant narrative stimulated by the media arises. The construction of the "Man" archetype and the reflections of active masculine imagery in the 21st century, when conveyed through media channels, could potentially have detrimental effects. Addressing this systematic behavioral chronology of virile cisgender, permeated imagistically through these means, involves exploring potential resolutions. Thus, an investigation process is initiated into the potential representation of the 'hero' in this media emulation through idols contextualized in the political sphere, with the purpose of elucidating the processes of simulation and emulation of narratives based on mythical, historical, and sacred accounts. In this process of sharing, the narratives contained in the imagistic structuring offered by information dissemination channels seek validation through a process of public acceptance. To achieve this consensus, a visual set adorned with mythological and sacred symbolisms adapted to the intended environment is promoted, thus utilizing sociocultural characteristics in favor of political marketing. Visual recognition, therefore, becomes a direct reflection of a cultural heritage acquired through lived human experience, stimulated by continuous representations throughout history. Echoes of imagery and narratives undergo a constant process of resignification of their concepts, sharpened by their premises, and adapted to the environment in which they seek to establish themselves. Political figures analyzed in this article employ the practice of taking possession of symbolisms, mythological stories, and heroisms and adapt their visual construction through a continuous praxis of emulation. Thus, they utilize iconic mythological narratives to gain credibility through belief. Utilizing iconic mythological narratives for credibility through belief, the idol becomes the very act of release of trauma, offering believers liberation from preconceived concepts and allowing for the attribution of new meanings. To dissolve this issue and highlight the subjectivities within the intention of the image, a linguistic, semiotic, and anthropological methodology is created. Linguistics uses expressions like 'Blaming the Image' to create a mechanism of expressive action in questioning why to blame a construction or visual composition and thus seek answers in the first act. Semiotics and anthropology develop an imagistic atlas of graphic analysis, seeking to make connections, comparisons, and relations between modern and sacred/mystical narratives, emphasizing the different subjective layers of embedded symbolism. Thus, it constitutes a performative act of disarming the image. It creates a disenchantment of the superficial gaze under the constant reproduction of visual content stimulated by virtual networks, enabling a discussion about the acceptance of caricatures characterized by past fables.

Keywords: image, heroic narrative, media heroism, virile politics, political, myth, sacred performance, visual mythmaking, characterization dynamics

Procedia PDF Downloads 31
289 Collaborative Governance in Dutch Flood Risk Management: An Historical Analysis

Authors: Emma Avoyan

Abstract:

The safety standards for flood protection in the Netherlands have been revised recently. It is expected that all major flood-protection structures will have to be reinforced to meet the new standards. The Dutch Flood Protection Programme aims at accomplishing this task through innovative integrated projects such as construction of multi-functional flood defenses. In these projects, flood safety purposes will be combined with spatial planning, nature development, emergency management or other sectoral objectives. Therefore, implementation of dike reinforcement projects requires early involvement and collaboration between public and private sectors, different governmental actors and agencies. The development and implementation of such integrated projects has been an issue in Dutch flood risk management since long. Therefore, this article analyses how cross-sector collaboration within flood risk governance in the Netherlands has evolved over time, and how this development can be explained. The integrative framework for collaborative governance is applied as an analytical tool to map external factors framing possibilities as well as constraints for cross-sector collaboration in Dutch flood risk domain. Supported by an extensive document and literature analysis, the paper offers insights on how the system context and different drivers changing over time either promoted or hindered cross-sector collaboration between flood protection sector, urban development, nature conservation or any other sector involved in flood risk governance. The system context refers to the multi-layered and interrelated suite of conditions that influence the formation and performance of complex governance systems, such as collaborative governance regimes, whereas the drivers initiate and enable the overall process of collaboration. In addition, by applying a method of process tracing we identify a causal and chronological chain of events shaping cross-sectoral interaction in Dutch flood risk management. Our results indicate that in order to evaluate the performance of complex governance systems, it is important to firstly study the system context that shapes it. Clear understanding of the system conditions and drivers for collaboration gives insight into the possibilities of and constraints for effective performance of complex governance systems. The performance of the governance system is affected by the system conditions, while at the same time the governance system can also change the system conditions. Our results show that the sequence of changes within the system conditions and drivers over time affect how cross-sector interaction in Dutch flood risk governance system happens now. Moreover, we have traced the potential of this governance system to shape and change the system context.

Keywords: collaborative governance, cross-sector interaction, flood risk management, the Netherlands

Procedia PDF Downloads 110
288 Dysphagia Tele Assessment Challenges Faced by Speech and Swallow Pathologists in India: Questionnaire Study

Authors: B. S. Premalatha, Mereen Rose Babu, Vaishali Prabhu

Abstract:

Background: Dysphagia must be assessed, either subjectively or objectively, in order to properly address the swallowing difficulty. Providing therapeutic care to patients with dysphagia via tele mode was one approach for providing clinical services during the COVID-19 epidemic. As a result, the teleassessment of dysphagia has increased in India. Aim: This study aimed to identify challenges faced by Indian SLPs while providing teleassessment to individuals with dysphagia during the outbreak of COVID-19 from 2020 to 2021. Method: After receiving approval from the institute's institutional review board and ethics committee, the current study was carried out. The study was cross-sectional in nature and lasted from 2020 to 2021. The study enrolled participants who met the inclusion and exclusion criteria of the study. It was decided to recruit roughly 246 people based on the sample size calculations. The research was done in three stages: questionnaire development and content validation, questionnaire administration. Five speech and hearing professionals' content verified the questionnaire for faults and clarity. Participants received questionnaires via various social media platforms such as e-mail and WhatsApp, which were written in Microsoft Word and then converted to Google Forms. SPSS software was used to examine the data. Results: In light of the obstacles that Indian SLPs encounter, the study's findings were examined. Only 135 people responded. During the COVID-19 lockdowns, 38% of participants said they did not deal with dysphagia patients. After the lockout, 70.4% of SLPs kept working with dysphagia patients, while 29.6% did not. From the beginning of the oromotor examination, the main problems in completing tele evaluation of dysphagia have been highlighted. Around 37.5% of SLPs said they don't undertake the OPME online because of difficulties doing the evaluation, such as the need for repeated instructions from patients and family members and trouble visualizing structures in various positions. The majority of SLPs' online assessments were inefficient and time-consuming. A bigger percentage of SLPs stated that they will not advocate tele evaluation in dysphagia to their colleagues. SLPs' use of dysphagia assessment has decreased as a result of the epidemic. When it came to the amount of food, the majority of people proposed a small amount. Apart from placing the patient for assessment and gaining less cooperation from the family, most SLPs found that Internet speed was a source of concern and a barrier. Hearing impairment and the presence of a tracheostomy in patients with dysphagia proved to be the most difficult conditions to treat online. For patients with NPO, the majority of SLPs did not advise tele-evaluation. In the anterior region of the oral cavity, oral meal residue was more visible. The majority of SLPs reported more anterior than posterior leakage. Even while the majority of SLPs could detect aspiration by coughing, many found it difficult to discern the gurgling tone of speech after swallowing. Conclusion: The current study sheds light on the difficulties that Indian SLPs experience when assessing dysphagia via tele mode, indicating that tele-assessment of dysphagia is still to gain importance in India.

Keywords: dysphagia, teleassessment, challenges, Indian SLP

Procedia PDF Downloads 105
287 Experimental-Numerical Inverse Approaches in the Characterization and Damage Detection of Soft Viscoelastic Layers from Vibration Test Data

Authors: Alaa Fezai, Anuj Sharma, Wolfgang Mueller-Hirsch, André Zimmermann

Abstract:

Viscoelastic materials have been widely used in the automotive industry over the last few decades with different functionalities. Besides their main application as a simple and efficient surface damping treatment, they may ensure optimal operating conditions for on-board electronics as thermal interface or sealing layers. The dynamic behavior of viscoelastic materials is generally dependent on many environmental factors, the most important being temperature and strain rate or frequency. Prior to the reliability analysis of systems including viscoelastic layers, it is, therefore, crucial to accurately predict the dynamic and lifetime behavior of these materials. This includes the identification of the dynamic material parameters under critical temperature and frequency conditions along with a precise damage localization and identification methodology. The goal of this work is twofold. The first part aims at applying an inverse viscoelastic material-characterization approach for a wide frequency range and under different temperature conditions. For this sake, dynamic measurements are carried on a single lap joint specimen using an electrodynamic shaker and an environmental chamber. The specimen consists of aluminum beams assembled to adapter plates through a viscoelastic adhesive layer. The experimental setup is reproduced in finite element (FE) simulations, and frequency response functions (FRF) are calculated. The parameters of both the generalized Maxwell model and the fractional derivatives model are identified through an optimization algorithm minimizing the difference between the simulated and the measured FRFs. The second goal of the current work is to guarantee an on-line detection of the damage, i.e., delamination in the viscoelastic bonding of the described specimen during frequency monitored end-of-life testing. For this purpose, an inverse technique, which determines the damage location and size based on the modal frequency shift and on the change of the mode shapes, is presented. This includes a preliminary FE model-based study correlating the delamination location and size to the change in the modal parameters and a subsequent experimental validation achieved through dynamic measurements of specimen with different, pre-generated crack scenarios and comparing it to the virgin specimen. The main advantage of the inverse characterization approach presented in the first part resides in the ability of adequately identifying the material damping and stiffness behavior of soft viscoelastic materials over a wide frequency range and under critical temperature conditions. Classic forward characterization techniques such as dynamic mechanical analysis are usually linked to limitations under critical temperature and frequency conditions due to the material behavior of soft viscoelastic materials. Furthermore, the inverse damage detection described in the second part guarantees an accurate prediction of not only the damage size but also its location using a simple test setup and outlines; therefore, the significance of inverse numerical-experimental approaches in predicting the dynamic behavior of soft bonding layers applied in automotive electronics.

Keywords: damage detection, dynamic characterization, inverse approaches, vibration testing, viscoelastic layers

Procedia PDF Downloads 179
286 Non-Perturbative Vacuum Polarization Effects in One- and Two-Dimensional Supercritical Dirac-Coulomb System

Authors: Andrey Davydov, Konstantin Sveshnikov, Yulia Voronina

Abstract:

There is now a lot of interest to the non-perturbative QED-effects, caused by diving of discrete levels into the negative continuum in the supercritical static or adiabatically slowly varying Coulomb fields, that are created by the localized extended sources with Z > Z_cr. Such effects have attracted a considerable amount of theoretical and experimental activity, since in 3+1 QED for Z > Z_cr,1 ≈ 170 a non-perturbative reconstruction of the vacuum state is predicted, which should be accompanied by a number of nontrivial effects, including the vacuum positron emission. Similar in essence effects should be expected also in both 2+1 D (planar graphene-based hetero-structures) and 1+1 D (one-dimensional ‘hydrogen ion’). This report is devoted to the study of such essentially non-perturbative vacuum effects for the supercritical Dirac-Coulomb systems in 1+1D and 2+1D, with the main attention drawn to the vacuum polarization energy. Although the most of works considers the vacuum charge density as the main polarization observable, vacuum energy turns out to be not less informative and in many respects complementary to the vacuum density. Moreover, the main non-perturbative effects, which appear in vacuum polarization for supercritical fields due to the levels diving into the lower continuum, show up in the behavior of vacuum energy even more clear, demonstrating explicitly their possible role in the supercritical region. Both in 1+1D and 2+1D, we explore firstly the renormalized vacuum density in the supercritical region using the Wichmann-Kroll method. Thereafter, taking into account the results for the vacuum density, we formulate the renormalization procedure for the vacuum energy. To evaluate the latter explicitly, an original technique, based on a special combination of analytical methods, computer algebra tools and numerical calculations, is applied. It is shown that, for a wide range of the external source parameters (the charge Z and size R), in the supercritical region the renormalized vacuum energy could significantly deviate from the perturbative quadratic growth up to pronouncedly decreasing behavior with jumps by (-2 x mc^2), which occur each time, when the next discrete level dives into the negative continuum. In the considered range of variation of Z and R, the vacuum energy behaves like ~ -Z^2/R in 1+1D and ~ -Z^3/R in 2+1D, exceeding deeply negative values. Such behavior confirms the assumption of the neutral vacuum transmutation into the charged one, and thereby of the spontaneous positron emission, accompanying the emergence of the next vacuum shell due to the total charge conservation. To the end, we also note that the methods, developed for the vacuum energy evaluation in 2+1 D, with minimal complements could be carried over to the three-dimensional case, where the vacuum energy is expected to be ~ -Z^4/R and so could be competitive with the classical electrostatic energy of the Coulomb source.

Keywords: non-perturbative QED-effects, one- and two-dimensional Dirac-Coulomb systems, supercritical fields, vacuum polarization

Procedia PDF Downloads 185
285 Standardized Testing of Filter Systems regarding Their Separation Efficiency in Terms of Allergenic Particles and Airborne Germs

Authors: Johannes Mertl

Abstract:

Our surrounding air contains various particles. Besides typical representatives of inorganic dust, such as soot and ash, also particles originating from animals, microorganisms or plants are floating through the air, so-called bioaerosols. The group of bioaerosols consists of a broad spectrum of particles of different size, including fungi, bacteria, viruses, spores, or tree, flower and grass pollen that are of high relevance for allergy sufferers. In dependence of the environmental climate and the actual season, these allergenic particles can be found in enormous numbers in the air and are inhaled by humans via the respiration tract, with a potential for inflammatory diseases of the airways, such as asthma or allergic rhinitis. As a consequence air filter systems of ventilation and air conditioning devices are required to meet very high standards to prevent, or at least lower the number of allergens and airborne germs entering the indoor air. Still, filter systems are merely classified for their separation rates using well-defined mineral test dust, while no appropriate sufficiently standardized test methods for bioaerosols exist. However, determined separation rates for mineral test particles of a certain size cannot simply be transferred to bioaerosols, as separation efficiency of particularly fine and respirable particles (< 10 microns) is dependent not only on their shape and particle diameter, but also defined by their density and physicochemical properties. For this reason, the OFI developed a test method, which directly enables a testing of filters and filter media for their separation rates on bioaerosols, as well as a classification of filters. Besides allergens from an intact or fractured tree or grass pollen, allergenic proteins bound to particulates, as well as allergenic fungal spores (e.g. Cladosporium cladosporioides), or bacteria can be used to classify filters regarding their separation rates. Allergens passing through the filter can then be detected by highly sensitive immunological assays (ELISA) or in the case of fungal spores by microbiological methods, which allow for the detection of even one single spore passing the filter. The test procedure, which is carried out in laboratory scale, was furthermore validated regarding its sufficiency to cover real life situations by upscaling using air conditioning devices showing great conformity in terms of separation rates. Additionally, a clinical study with allergy sufferers was performed to verify analytical results. Several different air conditioning filters from the car industry have been tested, showing significant differences in their separation rates.

Keywords: airborne germs, allergens, classification of filters, fine dust

Procedia PDF Downloads 226
284 Challenging Convections: Rethinking Literature Review Beyond Citations

Authors: Hassan Younis

Abstract:

Purpose: The objective of this study is to review influential papers in the sustainability and supply chain studies domain, leveraging insights from this review to develop a structured framework for academics and researchers. This framework aims to assist scholars in identifying the most impactful publications for their scholarly pursuits. Subsequently, the study will apply and trial the developed framework on selected scholarly articles within the sustainability and supply chain studies domain to evaluate its efficacy, practicality, and reliability. Design/Methodology/Approach: Utilizing the "Publish or Perish" tool, a search was conducted to locate papers incorporating "sustainability" and "supply chain" in their titles. After rigorous filtering steps, a panel of university professors identified five crucial criteria for evaluating research robustness: average yearly citation counts (25%), scholarly contribution (25%), alignment of findings with objectives (15%), methodological rigor (20%), and journal impact factor (15%). These five evaluation criteria are abbreviated as “ACMAJ" framework. Each paper then received a tiered score (1-3) for each criterion, normalized within its category, and summed using weighted averages to calculate a Final Normalized Score (FNS). This systematic approach allows for objective comparison and ranking of the research based on its impact, novelty, rigor, and publication venue. Findings: The study's findings highlight the lack of structured frameworks for assessing influential sustainability research in supply chain management, which often results in a dependence on citation counts. A complete model that incorporates five essential criteria has been suggested as a response. By conducting a methodical trial on specific academic articles in the field of sustainability and supply chain studies, the model demonstrated its effectiveness as a tool for identifying and selecting influential research papers that warrant additional attention. This work aims to fill a significant deficiency in existing techniques by providing a more comprehensive approach to identifying and ranking influential papers in the field. Practical Implications: The developed framework helps scholars identify the most influential sustainability and supply chain publications. Its validation serves the academic community by offering a credible tool and helping researchers, students, and practitioners find and choose influential papers. This approach aids field literature reviews and study suggestions. Analysis of major trends and topics deepens our grasp of this critical study area's changing terrain. Originality/Value: The framework stands as a unique contribution to academia, offering scholars an important and new tool to identify and validate influential publications. Its distinctive capacity to efficiently guide scholars, learners, and professionals in selecting noteworthy publications, coupled with the examination of key patterns and themes, adds depth to our understanding of the evolving landscape in this critical field of study.

Keywords: supply chain management, sustainability, framework, model

Procedia PDF Downloads 17
283 A New Second Tier Screening for Congenital Adrenal Hyperplasia Utilizing One Dried Blood Spot

Authors: Engy Shokry, Giancarlo La Marca, Maria Luisa Della Bona

Abstract:

Newborn screening for Congenital Adrenal Hyperplasia (CAH) relies on quantification of 17α-hydroxyprogesterone using enzyme immunoassays. These assays, in spite of being rapid, readily available and easy to perform, its reliability was found questionable due to lack of selectivity and specificity resulting in large number of false-positives, consequently family anxiety and associated hospitalization costs. To improve specificity of conventional 17α-hydroxyprogesterone screening which may experience false transient elevation in preterm, low birth weight or acutely ill neonates, steroid profiling by LC-MS/MS as a second-tier test was implemented. Unlike the previously applied LC-MS/MS methods, with the disadvantage of requiring a relatively high number of blood drops. Since newborn screening tests are increasing, it is necessary to minimize the sample volume requirement to make the maximum use of blood samples collected on filter paper. The proposed new method requires just one 3.2 mm dried blood spot (DBS) punch. Extraction was done using methanol: water: formic acid (90:10:0.1, v/v/v) containing deuterium labelled internal standards. Extracts were evaporated and reconstituted in 10 % acetone in water. Column switching strategy for on-line sample clean-up was applied to improve the chromatographic run. The first separative step retained the investigated steroids and passed through the majority of high molecular weight impurities. After the valve switching, the investigated steroids are back flushed from the POROS® column onto the analytical column and separated using gradient elution. Found quantitation limits were 5, 10 and 50 nmol/L for 17α-hydroxyprogesterone, androstenedione and cortisol respectively with mean recoveries of between 98.31-103.24 % and intra-/ inter-assay CV% < 10 % except at LLOQ. The method was validated using standard addition calibration and isotope dilution strategies. Reference ranges were determined by analysing samples from 896 infants of various ages at the time of sample collection. The method was also applied on patients with confirmed CAH. Our method represents an attractive combination of low sample volume requirement, minimal sample preparation time without derivatization and quick chromatography (5 min). The three steroid profile and the concentration ratios (17OHP + androstenedione/cortisol) allowed better screening outcomes of CAH reducing false positives, associated costs and anxiety.

Keywords: congenital adrenal hyperplasia (CAH), 17α-hydroxyprogesterone, androstenedione, cortisol, LC-MS/MS

Procedia PDF Downloads 413
282 Clubhouse: A Minor Rebellion against the Algorithmic Tyranny of the Majority

Authors: Vahid Asadzadeh, Amin Ataee

Abstract:

Since the advent of social media, there has been a wave of optimism among researchers and civic activists about the influence of virtual networks on the democratization process, which has gradually waned. One of the lesser-known concerns is how to increase the possibility of hearing the voices of different minorities. According to the theory of media logic, the media, using their technological capabilities, act as a structure through which events and ideas are interpreted. Social media, through the use of the learning machine and the use of algorithms, has formed a kind of structure in which the voices of minorities and less popular topics are lost among the commotion of the trends. In fact, the recommended systems and algorithms used in social media are designed to help promote trends and make popular content more popular, and content that belongs to minorities is constantly marginalized. As social networks gradually play a more active role in politics, the possibility of freely participating in the reproduction and reinterpretation of structures in general and political structures in particular (as Laclau‎ and Mouffe had in mind‎) can be considered as criteria to democracy in action. The point is that the media logic of virtual networks is shaped by the rule and even the tyranny of the majority, and this logic does not make it possible to design a self-foundation and self-revolutionary model of democracy. In other words, today's social networks, though seemingly full of variety But they are governed by the logic of homogeneity, and they do not have the possibility of multiplicity as is the case in immanent radical democracies (influenced by Gilles Deleuze). However, with the emergence and increasing popularity of Clubhouse as a new social media, there seems to be a shift in the social media space, and that is the diminishing role of algorithms and systems reconditioners as content delivery interfaces. This has led to the fact that in the Clubhouse, the voices of minorities are better heard, and the diversity of political tendencies manifests itself better. The purpose of this article is to show, first, how social networks serve the elimination of minorities in general, and second, to argue that the media logic of social networks must adapt to new interpretations of democracy that give more space to minorities and human rights. Finally, this article will show how the Clubhouse serves the new interpretations of democracy at least in a minimal way. To achieve the mentioned goals, in this article by a descriptive-analytical method, first, the relation between media logic and postmodern democracy will be inquired. The political economy popularity in social media and its conflict with democracy will be discussed. Finally, it will be explored how the Clubhouse provides a new horizon for the concepts embodied in radical democracy, a horizon that more effectively serves the rights of minorities and human rights in general.

Keywords: algorithmic tyranny, Clubhouse, minority rights, radical democracy, social media

Procedia PDF Downloads 123
281 Double Burden of Hypertension-Hyperalbuminuria in the Pregnant Women: Cross-Sectional Study of Prevalence and Risk Factors in Foumban, West Region, Cameroon

Authors: Pierre Mintom, Ebai Patricia, Merlin Dasse, Marlyse Chantal Nyangon Ndongo, Aicha Aretouyap Kouotou, Felix Essiben, Christine Fernande Nyangono Biyegue

Abstract:

Background: The death of women during and after pregnancy remains a major concern in public health policy in Cameroon. Among the causes of this mortality is eclampsia which is a consequence of the Pre-eclampsia characterized by the double burden of pregnancy-induced hypertension and albuminuria in pregnant women. Objective: To determine the various factors associated with the pre-eclampsia in pregnant women of Foumban. Methodology: A cross-sectional and analytical study was carried out during the period from July to August 2020 and supplemented by another study carried out from August 05 to September 05, 2022, at the Foumban district hospital. A questionnaire was administered to pregnant women. It focused on socio-demographic parameters, the state nutritional, health status, and maternal parameters. Blood pressure was taken using an electronic blood pressure monitor, and urinary albumin was measuring using urine dipstick. Pre-eclampsia was defined by three types of double burden: double burden systolic hypertension–hyperalbuminuria (SHH), defined for SBP≥140 mmHg and hyperalbuminuria ≥1+ on urine dipstick, double burden diastolic hypertension–hyperalbuminuria (DHH), defined for PAD≥90 mmHg and hyperalbuminuria ≥1+ on the urine dipstick, and the double burden systolodiastolic arterial hypertension– hyperalbuminuria (SDHH), defined for SBP ≥ 140mmHg, PAD≥90 mmHg and hyperalbuminuria ≥1+ on urine dipstick. IBM SPSS Software was used for statistical analysis. Results: The results of this study show that the prevalence of pre-eclampsia was 17.3% for the double burden SHH, 19.9% for the double burden DHH and 14.1% for double burden SDHH. Associated factors with pre-eclampsia according to the three types of double burden were marital status (P<0.05), religion (P<0.05), history of hypertension before pregnancy (P<0.05). Associated factors for the double burden of DHH and SDHH were the nutritional status before the pregnancy (P<0.05) and the number of prenatal consultations (P<0.05). In terms of food groups, regular consumption of spices significantly increased the risk of pre-eclampsia by 5.318, 6.277 and 11.271 times respectively for the SHH, DHH and SDHH double burdens, while regular consumption of sweets regular consumption of sweets increased by 2.42 times and 2.053 times respectively the double DHH and SDHH burdens respectively. Conclusion: Our study made it possible to redefine pre-eclampsia by considering the subtypes of hypertension. Certain socio-demographic parameters and certain dietary habits influence the occurrence of pre-eclampsia characterized by the double burden Hypertension-hyperalbuminuria in pregnant women, which may later lead to the occurrence of eclampsia. Moreover, albuminemia could be a good predicitive factor of pre-eclampsia and could be explored.

Keywords: hypertension, hyperalbuminuria, pregnant women, foumban

Procedia PDF Downloads 38
280 Calculation of Pressure-Varying Langmuir and Brunauer-Emmett-Teller Isotherm Adsorption Parameters

Authors: Trevor C. Brown, David J. Miron

Abstract:

Gas-solid physical adsorption methods are central to the characterization and optimization of the effective surface area, pore size and porosity for applications such as heterogeneous catalysis, and gas separation and storage. Properties such as adsorption uptake, capacity, equilibrium constants and Gibbs free energy are dependent on the composition and structure of both the gas and the adsorbent. However, challenges remain, in accurately calculating these properties from experimental data. Gas adsorption experiments involve measuring the amounts of gas adsorbed over a range of pressures under isothermal conditions. Various constant-parameter models, such as Langmuir and Brunauer-Emmett-Teller (BET) theories are used to provide information on adsorbate and adsorbent properties from the isotherm data. These models typically do not provide accurate interpretations across the full range of pressures and temperatures. The Langmuir adsorption isotherm is a simple approximation for modelling equilibrium adsorption data and has been effective in estimating surface areas and catalytic rate laws, particularly for high surface area solids. The Langmuir isotherm assumes the systematic filling of identical adsorption sites to a monolayer coverage. The BET model is based on the Langmuir isotherm and allows for the formation of multiple layers. These additional layers do not interact with the first layer and the energetics are equal to the adsorbate as a bulk liquid. This BET method is widely used to measure the specific surface area of materials. Both Langmuir and BET models assume that the affinity of the gas for all adsorption sites are identical and so the calculated adsorbent uptake at the monolayer and equilibrium constant are independent of coverage and pressure. Accurate representations of adsorption data have been achieved by extending the Langmuir and BET models to include pressure-varying uptake capacities and equilibrium constants. These parameters are determined using a novel regression technique called flexible least squares for time-varying linear regression. For isothermal adsorption the adsorption parameters are assumed to vary slowly and smoothly with increasing pressure. The flexible least squares for pressure-varying linear regression (FLS-PVLR) approach assumes two distinct types of discrepancy terms, dynamic and measurement for all parameters in the linear equation used to simulate the data. Dynamic terms account for pressure variation in successive parameter vectors, and measurement terms account for differences between observed and theoretically predicted outcomes via linear regression. The resultant pressure-varying parameters are optimized by minimizing both dynamic and measurement residual squared errors. Validation of this methodology has been achieved by simulating adsorption data for n-butane and isobutane on activated carbon at 298 K, 323 K and 348 K and for nitrogen on mesoporous alumina at 77 K with pressure-varying Langmuir and BET adsorption parameters (equilibrium constants and uptake capacities). This modeling provides information on the adsorbent (accessible surface area and micropore volume), adsorbate (molecular areas and volumes) and thermodynamic (Gibbs free energies) variations of the adsorption sites.

Keywords: Langmuir adsorption isotherm, BET adsorption isotherm, pressure-varying adsorption parameters, adsorbate and adsorbent properties and energetics

Procedia PDF Downloads 198
279 Economic Policy to Promote small and Medium-sized Enterprises in Georgia in the Post-Pandemic Period

Authors: Gulnaz Erkomaishvili

Abstract:

Introduction: The paper assesses the impact of the COVID-19 pandemic on the activities of small and medium-sized enterprises in Georgia, identifies their problems, and analyzes the state economic policy measures. During the pandemic, entrepreneurs named the imposition of restrictions, access to financial resources, shortage of qualified personnel, high tax rates, unhealthy competition in the market, etc. as the main challenges. The Georgian government has had to take special measures to mitigate the crisis impact caused by the pandemic. For example - in 2020, they mobilized more than 1,6 billion Gel for various eventsto support entrepreneurs. Small and medium-sized entrepreneurship development strategy is presented based on the research; Corresponding conclusions are made, and recommendations are developed. Objectives: The object of research is small and medium-sized enterprises and economic-political decisions aimed at their promotion.Methodology: This paper uses general and specific methods, in particular, analysis, synthesis, induction, deduction, scientific abstraction, comparative and statistical methods, as well as experts’ evaluation. In-depth interviews with experts were conducted to determine quantitative and qualitative indicators; Publications of the National Statistics Office of Georgia are used to determine the regularity between analytical and statistical estimations. Also, theoretical and applied research of international organizations and scientist-economists are used. Contributions: The COVID-19pandemic has had a significant impact on small and medium-sized enterprises. For them, Lockdown is a major challenge. Total sales volume decreased. At the same time, the innovative capabilities of enterprises and the volume of sales in remote channels have increased. As for the assessment of state support measures by small and medium-sizedentrepreneurs, despite the existence of support programs, a large number of entrepreneurs still do not evaluate the measures taken by the state positively. Among the desirable measures to be taken by the state, which would improve the activities of small and medium-sized entrepreneurs, who negatively or largely negatively assessed the activity of the state, named: tax incentives/exemption from certain taxes at the initial stage; Need for periodic trainings/organization of digital technologies, marketing training courses to improve the qualification of employees; Logic and adequacy of criteria when awarding grants and funding; Facilitating the finding of investors; Less bureaucracy, etc.

Keywords: small and medium enterprises, small and medium entrepreneurship, economic policy for small and medium entrepreneurship development, government regulations in Georgia, COVID-19 pandemic

Procedia PDF Downloads 132
278 Strengthening by Assessment: A Case Study of Rail Bridges

Authors: Evangelos G. Ilias, Panagiotis G. Ilias, Vasileios T. Popotas

Abstract:

The United Kingdom has one of the oldest railway networks in the world dating back to 1825 when the world’s first passenger railway was opened. The network has some 40,000 bridges of various construction types using a wide range of materials including masonry, steel, cast iron, wrought iron, concrete and timber. It is commonly accepted that the successful operation of the network is vital for the economy of the United Kingdom, consequently the cost effective maintenance of the existing infrastructure is a high priority to maintain the operability of the network, prevent deterioration and to extend the life of the assets. Every bridge on the railway network is required to be assessed every eighteen years and a structured approach to assessments is adopted with three main types of progressively more detailed assessments used. These assessment types include Level 0 (standardized spreadsheet assessment tools), Level 1 (analytical hand calculations) and Level 2 (generally finite element analyses). There is a degree of conservatism in the first two types of assessment dictated to some extent by the relevant standards which can lead to some structures not achieving the required load rating. In these situations, a Level 2 Assessment is often carried out using finite element analysis to uncover ‘latent strength’ and improve the load rating. If successful, the more sophisticated analysis can save on costly strengthening or replacement works and avoid disruption to the operational railway. This paper presents the ‘strengthening by assessment’ achieved by Level 2 analyses. The use of more accurate analysis assumptions and the implementation of non-linear modelling and functions (material, geometric and support) to better understand buckling modes and the structural behaviour of historic construction details that are not specifically covered by assessment codes are outlined. Metallic bridges which are susceptible to loss of section size through corrosion have largest scope for improvement by the Level 2 Assessment methodology. Three case studies are presented, demonstrating the effectiveness of the sophisticated Level 2 Assessment methodology using finite element analysis against the conservative approaches employed for Level 0 and Level 1 Assessments. One rail overbridge and two rail underbridges that did not achieve the required load rating by means of a Level 1 Assessment due to the inadequate restraint provided by U-Frame action are examined and the increase in assessed capacity given by the Level 2 Assessment is outlined.

Keywords: assessment, bridges, buckling, finite element analysis, non-linear modelling, strengthening

Procedia PDF Downloads 286
277 National Digital Soil Mapping Initiatives in Europe: A Review and Some Examples

Authors: Dominique Arrouays, Songchao Chen, Anne C. Richer-De-Forges

Abstract:

Soils are at the crossing of many issues such as food and water security, sustainable energy, climate change mitigation and adaptation, biodiversity protection, human health and well-being. They deliver many ecosystem services that are essential to life on Earth. Therefore, there is a growing demand for soil information on a national and global scale. Unfortunately, many countries do not have detailed soil maps, and, when existing, these maps are generally based on more or less complex and often non-harmonized soil classifications. An estimate of their uncertainty is also often missing. Thus, there are not easy to understand and often not properly used by end-users. Therefore, there is an urgent need to provide end-users with spatially exhaustive grids of essential soil properties, together with an estimate of their uncertainty. One way to achieve this is digital soil mapping (DSM). The concept of DSM relies on the hypothesis that soils and their properties are not randomly distributed, but that they depend on the main soil-forming factors that are climate, organisms, relief, parent material, time (age), and position in space. All these forming factors can be approximated using several exhaustive spatial products such as climatic grids, remote sensing products or vegetation maps, digital elevation models, geological or lithological maps, spatial coordinates of soil information, etc. Thus, DSM generally relies on models calibrated with existing observed soil data (point observations or maps) and so-called “ancillary co-variates” that come from other available spatial products. Then the model is generalized on grids where soil parameters are unknown in order to predict them, and the prediction performances are validated using various methods. With the growing demand for soil information at a national and global scale and the increase of available spatial co-variates national and continental DSM initiatives are continuously increasing. This short review illustrates the main national and continental advances in Europe, the diversity of the approaches and the databases that are used, the validation techniques and the main scientific and other issues. Examples from several countries illustrate the variety of products that were delivered during the last ten years. The scientific production on this topic is continuously increasing and new models and approaches are developed at an incredible speed. Most of the digital soil mapping (DSM) products rely mainly on machine learning (ML) prediction models and/or the use or pedotransfer functions (PTF) in which calibration data come from soil analyses performed in labs or for existing conventional maps. However, some scientific issues remain to be solved and also political and legal ones related, for instance, to data sharing and to different laws in different countries. Other issues related to communication to end-users and education, especially on the use of uncertainty. Overall, the progress is very important and the willingness of institutes and countries to join their efforts is increasing. Harmonization issues are still remaining, mainly due to differences in classifications or in laboratory standards between countries. However numerous initiatives are ongoing at the EU level and also at the global level. All these progress are scientifically stimulating and also promissing to provide tools to improve and monitor soil quality in countries, EU and at the global level.

Keywords: digital soil mapping, global soil mapping, national and European initiatives, global soil mapping products, mini-review

Procedia PDF Downloads 162
276 Resilience-Vulnerability Interaction in the Context of Disasters and Complexity: Study Case in the Coastal Plain of Gulf of Mexico

Authors: Cesar Vazquez-Gonzalez, Sophie Avila-Foucat, Leonardo Ortiz-Lozano, Patricia Moreno-Casasola, Alejandro Granados-Barba

Abstract:

In the last twenty years, academic and scientific literature has been focused on understanding the processes and factors of coastal social-ecological systems vulnerability and resilience. Some scholars argue that resilience and vulnerability are isolated concepts due to their epistemological origin, while others note the existence of a strong resilience-vulnerability relationship. Here we present an ordinal logistic regression model based on the analytical framework about dynamic resilience-vulnerability interaction along adaptive cycle of complex systems and disasters process phases (during, recovery and learning). In this way, we demonstrate that 1) during the disturbance, absorptive capacity (resilience as a core of attributes) and external response capacity explain the probability of households capitals to diminish the damage, and exposure sets the thresholds about the amount of disturbance that households can absorb, 2) at recovery, absorptive capacity and external response capacity explain the probability of households capitals to recovery faster (resilience as an outcome) from damage, and 3) at learning, adaptive capacity (resilience as a core of attributes) explains the probability of households adaptation measures based on the enhancement of physical capital. As a result, during the disturbance phase, exposure has the greatest weight in the probability of capital’s damage, and households with absorptive and external response capacity elements absorbed the impact of floods in comparison with households without these elements. At the recovery phase, households with absorptive and external response capacity showed a faster recovery on their capital; however, the damage sets the thresholds of recovery time. More importantly, diversity in financial capital increases the probability of recovering other capital, but it becomes a liability so that the probability of recovering the household finances in a longer time increases. At learning-reorganizing phase, adaptation (modifications to the house) increases the probability of having less damage on physical capital; however, it is not very relevant. As conclusion, resilience is an outcome but also core of attributes that interacts with vulnerability along the adaptive cycle and disaster process phases. Absorptive capacity can diminish the damage experienced by floods; however, when exposure overcomes thresholds, both absorptive and external response capacity are not enough. In the same way, absorptive and external response capacity diminish the recovery time of capital, but the damage sets the thresholds in where households are not capable of recovering their capital.

Keywords: absorptive capacity, adaptive capacity, capital, floods, recovery-learning, social-ecological systems

Procedia PDF Downloads 109
275 Numerical Simulation of the Production of Ceramic Pigments Using Microwave Radiation: An Energy Efficiency Study Towards the Decarbonization of the Pigment Sector

Authors: Pedro A. V. Ramos, Duarte M. S. Albuquerque, José C. F. Pereira

Abstract:

Global warming mitigation is one of the main challenges of this century, having the net balance of greenhouse gas (GHG) emissions to be null or negative in 2050. Industry electrification is one of the main paths to achieving carbon neutrality within the goals of the Paris Agreement. Microwave heating is becoming a popular industrial heating mechanism due to the absence of direct GHG emissions, but also the rapid, volumetric, and efficient heating. In the present study, a mathematical model is used to simulate the production using microwave heating of two ceramic pigments, at high temperatures (above 1200 Celsius degrees). The two pigments studied were the yellow (Pr, Zr)SiO₂ and the brown (Ti, Sb, Cr)O₂. The chemical conversion of reactants into products was included in the model by using the kinetic triplet obtained with the model-fitting method and experimental data present in the Literature. The coupling between the electromagnetic, thermal, and chemical interfaces was also included. The simulations were computed in COMSOL Multiphysics. The geometry includes a moving plunger to allow for the cavity impedance matching and thus maximize the electromagnetic efficiency. To accomplish this goal, a MATLAB controller was developed to automatically search the position of the moving plunger that guarantees the maximum efficiency. The power is automatically and permanently adjusted during the transient simulation to impose stationary regime and total conversion, the two requisites of every converged solution. Both 2D and 3D geometries were used and a parametric study regarding the axial bed velocity and the heat transfer coefficient at the boundaries was performed. Moreover, a Verification and Validation study was carried out by comparing the conversion profiles obtained numerically with the experimental data available in the Literature; the numerical uncertainty was also estimated to attest to the result's reliability. The results show that the model-fitting method employed in this work is a suitable tool to predict the chemical conversion of reactants into the pigment, showing excellent agreement between the numerical results and the experimental data. Moreover, it was demonstrated that higher velocities lead to higher thermal efficiencies and thus lower energy consumption during the process. This work concludes that the electromagnetic heating of materials having high loss tangent and low thermal conductivity, like ceramic materials, maybe a challenge due to the presence of hot spots, which may jeopardize the product quality or even the experimental apparatus. The MATLAB controller increased the electromagnetic efficiency by 25% and global efficiency of 54% was obtained for the titanate brown pigment. This work shows that electromagnetic heating will be a key technology in the decarbonization of the ceramic sector as reductions up to 98% in the specific GHG emissions were obtained when compared to the conventional process. Furthermore, numerical simulations appear as a suitable technique to be used in the design and optimization of microwave applicators, showing high agreement with experimental data.

Keywords: automatic impedance matching, ceramic pigments, efficiency maximization, high-temperature microwave heating, input power control, numerical simulation

Procedia PDF Downloads 116
274 A Cross-Sectional Study of Knowledge and Attitudes among College Students in a South Indian City about Intimate Partner Violence

Authors: Krithika Lakshmi Sathiya Moorthy

Abstract:

Introduction: Young people’s attitude towards Intimate partner violence (IPV) is likely to influence their indulgence in or abstinence from IPV in future. We aimed to assess the knowledge and attitudes of college students in a south Indian city regarding IPV, its associated factors and redressal mechanisms. Methods: A convenient sample of 247 students, pursuing medicine and engineering, participated in this analytical cross sectional study. They responded to a self-administered questionnaire developed and pretested for this study. The questionnaire comprises statements from a third person’s perspective and vignettes to reduce social desirability bias. Clearance was obtained from the Institute Ethical Committee of Velammal Medical College Hospital and Research Institute, Madurai, India. Data were entered in Epidata Entry v3.1, Odense, Denmark and analysed using SPSS v20.0. Results: Among 247 students, 116 (47%) were males and 59 (24.9%) hailed from rural areas. About 18% (43) of students believed that IPV was a problem only among females. Almost half of the students had witnessed IPV; at home between their parents (9.7%), other family members (13.4%), in their neighbourhood (13%) or public places (15%). Only 118 (47.8%) were aware that a law was in place in India to address IPV. The perceived risk factors for IPV were alcoholic spouse (78.9%), low income families (53.8%), personality traits (52.2%) and dowry system (51%). A sizeable number of students (38.4%) believed that some amount of physical violence was allowable in a marital relationship while 57.6% even considered IPV as an expression of love. Males as compared to females were more in agreement with negative gender stereotypes such as husband can– ‘threaten wife to ensure welfare of family’ (55% vs. 34%, p < 0.001), ‘spy on wife to check fidelity’ (41% vs. 27%, p < 0.001), ‘financially deprive housewife to punish’ (13% vs. 3.8%, p=0.001) and agreed with the statement that it is ‘duty of wife to comply with demands for sex from the husband’ (9.5% vs 4.6%, p=0.3). About 32% males and 25.6% females foresaw themselves as perpetrators of IPV in future. Conclusion: Knowledge about IPV and the associated risk factors among the study population was satisfactory. However, there was widespread acceptance of negative societal gender stereotypes, more so among males and some degrees of IPV were acceptable between married couples. The study advocates the need to halt the propagation of negative gender stereotypes in the impressionable young minds and the necessity to spread the awareness that no degree of IPV is acceptable. This knowledge is also required to plan the content and choose the appropriate media to effectively communicate the awareness about IPV among young persons.

Keywords: attitude, India, intimate partner violence, knowledge, students

Procedia PDF Downloads 197
273 Modeling and Energy Analysis of Limestone Decomposition with Microwave Heating

Authors: Sofia N. Gonçalves, Duarte M. S. Albuquerque, José C. F. Pereira

Abstract:

The energy transition is spurred by structural changes in energy demand, supply, and prices. Microwave technology was first proposed as a faster alternative for cooking food. It was found that food heated instantly when interacting with high-frequency electromagnetic waves. The dielectric properties account for a material’s ability to absorb electromagnetic energy and dissipate this energy in the form of heat. Many energy-intense industries could benefit from electromagnetic heating since many of the raw materials are dielectric at high temperatures. Limestone sedimentary rock is a dielectric material intensively used in the cement industry to produce unslaked lime. A numerical 3D model was implemented in COMSOL Multiphysics to study the limestone continuous processing under microwave heating. The model solves the two-way coupling between the Energy equation and Maxwell’s equations as well as the coupling between heat transfer and chemical interfaces. Complementary, a controller was implemented to optimize the overall heating efficiency and control the numerical model stability. This was done by continuously matching the cavity impedance and predicting the required energy for the system, avoiding energy inefficiencies. This controller was developed in MATLAB and successfully fulfilled all these goals. The limestone load influence on thermal decomposition and overall process efficiency was the main object of this study. The procedure considered the Verification and Validation of the chemical kinetics model separately from the coupled model. The chemical model was found to correctly describe the chosen kinetic equation, and the coupled model successfully solved the equations describing the numerical model. The interaction between flow of material and electric field Poynting vector revealed to influence limestone decomposition, as a result from the low dielectric properties of limestone. The numerical model considered this effect and took advantage from this interaction. The model was demonstrated to be highly unstable when solving non-linear temperature distributions. Limestone has a dielectric loss response that increases with temperature and has low thermal conductivity. For this reason, limestone is prone to produce thermal runaway under electromagnetic heating, as well as numerical model instabilities. Five different scenarios were tested by considering a material fill ratio of 30%, 50%, 65%, 80%, and 100%. Simulating the tube rotation for mixing enhancement was proven to be beneficial and crucial for all loads considered. When uniform temperature distribution is accomplished, the electromagnetic field and material interaction is facilitated. The results pointed out the inefficient development of the electric field within the bed for 30% fill ratio. The thermal efficiency showed the propensity to stabilize around 90%for loads higher than 50%. The process accomplished a maximum microwave efficiency of 75% for the 80% fill ratio, sustaining that the tube has an optimal fill of material. Electric field peak detachment was observed for the case with 100% fill ratio, justifying the lower efficiencies compared to 80%. Microwave technology has been demonstrated to be an important ally for the decarbonization of the cement industry.

Keywords: CFD numerical simulations, efficiency optimization, electromagnetic heating, impedance matching, limestone continuous processing

Procedia PDF Downloads 147
272 An Appraisal of Mining Sector Corporate Social Responsibility Processes in Mhondoro-Ngezi, Zimbabwe

Authors: A. T. Muruviwa

Abstract:

To-date, the discourse on corporate social responsibility (CSR) has primarily centred on the actions and inactions of corporations; hence, the dominant focus on CSR has been on impacts and outcomes. The obscuring effect of this approach has, arguably, resulted in the emergence of what may be termed a ‘Northern’ agenda on CSR theory and practice, in contrast to an emergency ‘Southern’ discourse, which appears to highlight the crucial issues of poverty reduction, infrastructure development and the broader questions of social provisioning and community empowerment. Some scholars have explicitly called for a CSR research agenda that focuses on the 'reciprocal duties' of the stakeholders in the CSR process rather than fixate on the actions and inactions of business. It is against the backdrop of these contestations that this study assesses the reciprocal relationships amongst CSR stakeholders in a Zimbabwean platinum mining town, with a view to demonstrating how such relationships – and the expectations and obligations embedded in them – impact on the success or failure of CSR initiatives. The existence of mutual relations between the corporation and its stakeholders signifies the successes of CSR processes and hence the outcomes. The company is Zimplats Mining Company; the community is Mhondoro-Ngezi, and the stakeholders are clearly identified in the study. The study utilised a triangulated design, with data collected using a mini survey, focus groups, in-depth interview and observation. The key findings are that the CSR process in the study community is dominated by the mining company. Despite the existence of a CSR framework that recognises government, local leaders and community members as legitimate stakeholders, there is little evidence of concrete contributions made by these stakeholders towards the realisation of CSR objectives. As a result, the community development process – in so far as CSR is concerned – fails to address the developmental concerns of the various stakeholders. On the basis of these findings, the study concludes that there is a crisis of reciprocity in the CSR process in Mhondoro-Ngezi, and that a situation where the conceptualisation of local development needs and the deployment of specific development tools seems to be driven by one stakeholder almost to the exclusion of all others, can only present contradictory development outcomes. The significance of this study is that it allows for the development of a more nuanced and robust CSR discourse. Rather than focusing on the corporate and stakeholder perspectives and outcomes of CSR initiatives, this study examines the CSR- development nexus by interrogating the idea of reciprocal responsibility as a sin qua non to CSR success. This analytical strategy and focus allow the researcher to gain a clear understanding of how stakeholder relationships and duties influence CSR processes and also the overall outcome. At a more practical level, the findings of the study should help to shape the policy on corporate community relationships with a view to enhancing the role of mining in development.

Keywords: community development, processes, reciprocity, stakeholders

Procedia PDF Downloads 324
271 Assessing Brain Targeting Efficiency of Ionisable Lipid Nanoparticles Encapsulating Cas9 mRNA/gGFP Following Different Routes of Administration in Mice

Authors: Meiling Yu, Nadia Rouatbi, Khuloud T. Al-Jamal

Abstract:

Background: Treatment of neurological disorders with modern medical and surgical approaches remains difficult. Gene therapy, allowing the delivery of genetic materials that encodes potential therapeutic molecules, represents an attractive option. The treatment of brain diseases with gene therapy requires the gene-editing tool to be delivered efficiently to the central nervous system. In this study, we explored the efficiency of different delivery routes, namely intravenous (i.v.), intra-cranial (i.c.), and intra-nasal (i.n.), to deliver stable nucleic acid-lipid particles (SNALPs) containing gene-editing tools namely Cas9 mRNA and sgRNA encoding for GFP as a reporter protein. We hypothesise that SNALPs can reach the brain and perform gene-editing to different extents depending on the administration route. Intranasal administration (i.n.) offers an attractive and non-invasive way to access the brain circumventing the blood–brain barrier. Successful delivery of gene-editing tools to the brain offers a great opportunity for therapeutic target validation and nucleic acids therapeutics delivery to improve treatment options for a range of neurodegenerative diseases. In this study, we utilised Rosa26-Cas9 knock-in mice, expressing GFP, to study brain distribution and gene-editing efficiency of SNALPs after i.v.; i.c. and i.n. routes of administration. Methods: Single guide RNA (sgRNA) against GFP has been designed and validated by in vitro nuclease assay. SNALPs were formulated and characterised using dynamic light scattering. The encapsulation efficiency of nucleic acids (NA) was measured by RiboGreen™ assay. SNALPs were incubated in serum to assess their ability to protect NA from degradation. Rosa26-Cas9 knock-in mice were i.v., i.n., or i.c. administered with SNALPs to test in vivo gene-editing (GFP knockout) efficiency. SNALPs were given as three doses of 0.64 mg/kg sgGFP following i.v. and i.n. or a single dose of 0.25 mg/kg sgGFP following i.c.. knockout efficiency was assessed after seven days using Sanger Sequencing and Inference of CRISPR Edits (ICE) analysis. In vivo, the biodistribution of DiR labelled SNALPs (SNALPs-DiR) was assessed at 24h post-administration using IVIS Lumina Series III. Results: Serum-stable SNALPs produced were 130-140 nm in diameter with ~90% nucleic acid loading efficiency. SNALPs could reach and stay in the brain for up to 24h following i.v.; i.n. and i.c. administration. Decreasing GFP expression (around 50% after i.v. and i.c. and 20% following i.n.) was confirmed by optical imaging. Despite the small number of mice used, ICE analysis confirmed GFP knockout in mice brains. Additional studies are currently taking place to increase mice numbers. Conclusion: Results confirmed efficient gene knockout achieved by SNALPs in Rosa26-Cas9 knock-in mice expressing GFP following different routes of administrations in the following order i.v.= i.c.> i.n. Each of the administration routes has its pros and cons. The next stages of the project involve assessing gene-editing efficiency in wild-type mice and replacing GFP as a model target with therapeutic target genes implicated in Motor Neuron Disease pathology.

Keywords: CRISPR, nanoparticles, brain diseases, administration routes

Procedia PDF Downloads 67
270 Evaluating the Accuracy of Biologically Relevant Variables Generated by ClimateAP

Authors: Jing Jiang, Wenhuan XU, Lei Zhang, Shiyi Zhang, Tongli Wang

Abstract:

Climate data quality significantly affects the reliability of ecological modeling. In the Asia Pacific (AP) region, low-quality climate data hinders ecological modeling. ClimateAP, a software developed in 2017, generates high-quality climate data for the AP region, benefiting researchers in forestry and agriculture. However, its adoption remains limited. This study aims to confirm the validity of biologically relevant variable data generated by ClimateAP during the normal climate period through comparison with the currently available gridded data. Climate data from 2,366 weather stations were used to evaluate the prediction accuracy of ClimateAP in comparison with the commonly used gridded data from WorldClim1.4. Univariate regressions were applied to 48 monthly biologically relevant variables, and the relationship between the observational data and the predictions made by ClimateAP and WorldClim was evaluated using Adjusted R-Squared and Root Mean Squared Error (RMSE). Locations were categorized into mountainous and flat landforms, considering elevation, slope, ruggedness, and Topographic Position Index. Univariate regressions were then applied to all biologically relevant variables for each landform category. Random Forest (RF) models were implemented for the climatic niche modeling of Cunninghamia lanceolata. A comparative analysis of the prediction accuracies of RF models constructed with distinct climate data sources was conducted to evaluate their relative effectiveness. Biologically relevant variables were obtained from three unpublished Chinese meteorological datasets. ClimateAPv3.0 and WorldClim predictions were obtained from weather station coordinates and WorldClim1.4 rasters, respectively, for the normal climate period of 1961-1990. Occurrence data for Cunninghamia lanceolata came from integrated biodiversity databases with 3,745 unique points. ClimateAP explains a minimum of 94.74%, 97.77%, 96.89%, and 94.40% of monthly maximum, minimum, average temperature, and precipitation variances, respectively. It outperforms WorldClim in 37 biologically relevant variables with lower RMSE values. ClimateAP achieves higher R-squared values for the 12 monthly minimum temperature variables and consistently higher Adjusted R-squared values across all landforms for precipitation. ClimateAP's temperature data yields lower Adjusted R-squared values than gridded data in high-elevation, rugged, and mountainous areas but achieves higher values in mid-slope drainages, plains, open slopes, and upper slopes. Using ClimateAP improves the prediction accuracy of tree occurrence from 77.90% to 82.77%. The biologically relevant climate data produced by ClimateAP is validated based on evaluations using observations from weather stations. The use of ClimateAP leads to an improvement in data quality, especially in non-mountainous regions. The results also suggest that using biologically relevant variables generated by ClimateAP can slightly enhance climatic niche modeling for tree species, offering a better understanding of tree species adaptation and resilience compared to using gridded data.

Keywords: climate data validation, data quality, Asia pacific climate, climatic niche modeling, random forest models, tree species

Procedia PDF Downloads 48
269 Assessment of Efficiency of Underwater Undulatory Swimming Strategies Using a Two-Dimensional CFD Method

Authors: Dorian Audot, Isobel Margaret Thompson, Dominic Hudson, Joseph Banks, Martin Warner

Abstract:

In competitive swimming, after dives and turns, athletes perform underwater undulatory swimming (UUS), copying marine mammals’ method of locomotion. The body, performing this wave-like motion, accelerates the fluid downstream in its vicinity, generating propulsion with minimal resistance. Through this technique, swimmers can maintain greater speeds than surface swimming and take advantage of the overspeed granted by the dive (or push-off). Almost all previous work has considered UUS when performed at maximum effort. Critical parameters to maximize UUS speed are frequently discussed; however, this does not apply to most races. In only 3 out of the 16 individual competitive swimming events are athletes likely to attempt to perform UUS with the greatest speed, without thinking of the cost of locomotion. In the other cases, athletes will want to control the speed of their underwater swimming, attempting to maximise speed whilst considering energy expenditure appropriate to the duration of the event. Hence, there is a need to understand how swimmers adapt their underwater strategies to optimize the speed within the allocated energetic cost. This paper develops a consistent methodology that enables different sets of UUS kinematics to be investigated. These may have different propulsive efficiencies and force generation mechanisms (e.g.: force distribution along with the body and force magnitude). The developed methodology, therefore, needs to: (i) provide an understanding of the UUS propulsive mechanisms at different speeds, (ii) investigate the key performance parameters when UUS is not performed solely for maximizing speed; (iii) consistently determine the propulsive efficiency of a UUS technique. The methodology is separated into two distinct parts: kinematic data acquisition and computational fluid dynamics (CFD) analysis. For the kinematic acquisition, the position of several joints along the body and their sequencing were either obtained by video digitization or by underwater motion capture (Qualisys system). During data acquisition, the swimmers were asked to perform UUS at a constant depth in a prone position (facing the bottom of the pool) at different speeds: maximum effort, 100m pace, 200m pace and 400m pace. The kinematic data were input to a CFD algorithm employing a two-dimensional Large Eddy Simulation (LES). The algorithm adopted was specifically developed in order to perform quick unsteady simulations of deforming bodies and is therefore suitable for swimmers performing UUS. Despite its approximations, the algorithm is applied such that simulations are performed with the inflow velocity updated at every time step. It also enables calculations of the resistive forces (total and applied to each segment) and the power input of the modeled swimmer. Validation of the methodology is achieved by comparing the data obtained from the computations with the original data (e.g.: sustained swimming speed). This method is applied to the different kinematic datasets and provides data on swimmers’ natural responses to pacing instructions. The results show how kinematics affect force generation mechanisms and hence how the propulsive efficiency of UUS varies for different race strategies.

Keywords: CFD, efficiency, human swimming, hydrodynamics, underwater undulatory swimming

Procedia PDF Downloads 191
268 Empirical Analysis of the Relationship between Voluntary Accounting Disclosures and Mongolian Stock Exchange Listed Companies’ Characteristics

Authors: Ernest Nweke

Abstract:

Mongolia has made giant strides in the development of its auditing and accounting system from Soviet-style to a market-oriented system. High levels of domestic and foreign investment desired by the Mongolian government require that better and improved quality of corporate information and disclosure consistent with international standards be made available to investors. However, the Mongolian Certified Public Accountants (CPA) profession is still developing, and the quality of services provided by accounting firms in most cases do not comply with International Financial Reporting Standards (IFRS) framework approved by the government for use in financial reporting. Against this backdrop, Accounting and audit reforms, liberalization and deregulation, establishment of an efficient and effective professional monitoring and supervision regime are policy necessities. These will further enhance the Mongolian business environment, eliminate incompetence in the system, make the economy more attractive to investors and ultimately lift reporting standards and bring about improved accounting, auditing and disclosure practices among Mongolian firms. This paper examines the fundamental issues in the accounting and auditing environment in Mongolia and investigates the relationship between selected characteristics of Mongolian Stock Exchange (MSE) listed firms (profitability, leverage, firm size, firm auditor size, firm listing age, board size and proportion of independent directors) and voluntary accounting disclosures in their annual reports and accounts. The selected sample of firms for the research purpose consists of the top 20 indexes of the MSE, representing over 95% of the market capitalization. An empirical analysis of the hypothesized relationship was carried out using multiple regression in EViews analytical software. Research results lend credence to the fact that only a few of the company attributes positively impact voluntary accounting disclosures in Mongolian Stock Exchange-listed firms. The research is motivated by the absence of empirical evidence on the correlation between the quality of voluntary accounting disclosures made by listed companies in Mongolia and company characteristics and the findings thereof significantly useful to both firms and regulatory authorities. The concluding part of the paper precisely consists of useful research-based recommendations for listed firms and regulatory agencies on measures to put in place in order to enhance the quality of corporate financial reporting and disclosures in Mongolia.

Keywords: accounting, auditing, corporate disclosure, listed firms

Procedia PDF Downloads 77