Search results for: energy performance gaps
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 19496

Search results for: energy performance gaps

9956 ChatGPT 4.0 Demonstrates Strong Performance in Standardised Medical Licensing Examinations: Insights and Implications for Medical Educators

Authors: K. O'Malley

Abstract:

Background: The emergence and rapid evolution of large language models (LLMs) (i.e., models of generative artificial intelligence, or AI) has been unprecedented. ChatGPT is one of the most widely used LLM platforms. Using natural language processing technology, it generates customized responses to user prompts, enabling it to mimic human conversation. Responses are generated using predictive modeling of vast internet text and data swathes and are further refined and reinforced through user feedback. The popularity of LLMs is increasing, with a growing number of students utilizing these platforms for study and revision purposes. Notwithstanding its many novel applications, LLM technology is inherently susceptible to bias and error. This poses a significant challenge in the educational setting, where academic integrity may be undermined. This study aims to evaluate the performance of the latest iteration of ChatGPT (ChatGPT4.0) in standardized state medical licensing examinations. Methods: A considered search strategy was used to interrogate the PubMed electronic database. The keywords ‘ChatGPT’ AND ‘medical education’ OR ‘medical school’ OR ‘medical licensing exam’ were used to identify relevant literature. The search included all peer-reviewed literature published in the past five years. The search was limited to publications in the English language only. Eligibility was ascertained based on the study title and abstract and confirmed by consulting the full-text document. Data was extracted into a Microsoft Excel document for analysis. Results: The search yielded 345 publications that were screened. 225 original articles were identified, of which 11 met the pre-determined criteria for inclusion in a narrative synthesis. These studies included performance assessments in national medical licensing examinations from the United States, United Kingdom, Saudi Arabia, Poland, Taiwan, Japan and Germany. ChatGPT 4.0 achieved scores ranging from 67.1 to 88.6 percent. The mean score across all studies was 82.49 percent (SD= 5.95). In all studies, ChatGPT exceeded the threshold for a passing grade in the corresponding exam. Conclusion: The capabilities of ChatGPT in standardized academic assessment in medicine are robust. While this technology can potentially revolutionize higher education, it also presents several challenges with which educators have not had to contend before. The overall strong performance of ChatGPT, as outlined above, may lend itself to unfair use (such as the plagiarism of deliverable coursework) and pose unforeseen ethical challenges (arising from algorithmic bias). Conversely, it highlights potential pitfalls if users assume LLM-generated content to be entirely accurate. In the aforementioned studies, ChatGPT exhibits a margin of error between 11.4 and 32.9 percent, which resonates strongly with concerns regarding the quality and veracity of LLM-generated content. It is imperative to highlight these limitations, particularly to students in the early stages of their education who are less likely to possess the requisite insight or knowledge to recognize errors, inaccuracies or false information. Educators must inform themselves of these emerging challenges to effectively address them and mitigate potential disruption in academic fora.

Keywords: artificial intelligence, ChatGPT, generative ai, large language models, licensing exam, medical education, medicine, university

Procedia PDF Downloads 11
9955 Process Safety Management Digitalization via SHEQTool based on Occupational Safety and Health Administration and Center for Chemical Process Safety, a Case Study in Petrochemical Companies

Authors: Saeed Nazari, Masoom Nazari, Ali Hejazi, Siamak Sanoobari Ghazi Jahani, Mohammad Dehghani, Javad Vakili

Abstract:

More than ever, digitization is an imperative for businesses to keep their competitive advantages, foster innovation and reduce paperwork. To design and successfully implement digital transformation initiatives within process safety management system, employees need to be equipped with the right tool, frameworks, and best practices. we developed a unique full stack application so-called SHEQTool which is entirely dynamic based on our extensive expertise, experience, and client feedback to help business processes particularly operations safety management. We use our best knowledge and scientific methodologies published by CCPS and OSHA Guidelines to streamline operations and integrated them into task management within Petrochemical Companies. We digitalize their main process safety management system elements and their sub elements such as hazard identification and risk management, training and communication, inspection and audit, critical changes management, contractor management, permit to work, pre-start-up safety review, incident reporting and investigation, emergency response plan, personal protective equipment, occupational health, and action management in a fully customizable manner with no programming needs for users. We review the feedback from main actors within petrochemical plant which highlights improving their business performance and productivity as well as keep tracking their functions’ key performance indicators (KPIs) because it; 1) saves time, resources, and costs of all paperwork on our businesses (by Digitalization); 2) reduces errors and improve performance within management system by covering most of daily software needs of the organization and reduce complexity and associated costs of numerous tools and their required training (One Tool Approach); 3) focuses on management systems and integrate functions and put them into traceable task management (RASCI and Flowcharting); 4) helps the entire enterprise be resilient to any change of your processes, technologies, assets with minimum costs (through Organizational Resilience); 5) reduces significantly incidents and errors via world class safety management programs and elements (by Simplification); 6) gives the companies a systematic, traceable, risk based, process based, and science based integrated management system (via proper Methodologies); 7) helps business processes complies with ISO 9001, ISO 14001, ISO 45001, ISO 31000, best practices as well as legal regulations by PDCA approach (Compliance).

Keywords: process, safety, digitalization, management, risk, incident, SHEQTool, OSHA, CCPS

Procedia PDF Downloads 46
9954 Seismic Retrofits – A Catalyst for Minimizing the Building Sector’s Carbon Footprint

Authors: Juliane Spaak

Abstract:

A life-cycle assessment was performed, looking at seven retrofit projects in New Zealand using LCAQuickV3.5. The study found that retrofits save up to 80% of embodied carbon emissions for the structural elements compared to a new building. In other words, it is only a 20% carbon investment to transform and extend a building’s life. In addition, the systems were evaluated by looking at environmental impacts over the design life of these buildings and resilience using FEMA P58 and PACT software. With the increasing interest in Zero Carbon targets, significant changes in the building and construction sector are required. Emissions for buildings arise from both embodied carbon and operations. Based on the significant advancements in building energy technology, the focus is moving more toward embodied carbon, a large portion of which is associated with the structure. Since older buildings make up most of the real estate stock of our cities around the world, their reuse through structural retrofit and wider refurbishment plays an important role in extending the life of a building’s embodied carbon. New Zealand’s building owners and engineers have learned a lot about seismic issues following a decade of significant earthquakes. Recent earthquakes have brought to light the necessity to move away from constructing code-minimum structures that are designed for life safety but are frequently ‘disposable’ after a moderate earthquake event, especially in relation to a structure’s ability to minimize damage. This means weaker buildings sit as ‘carbon liabilities’, with considerably more carbon likely to be expended remediating damage after a shake. Renovating and retrofitting older assets plays a big part in reducing the carbon profile of the buildings sector, as breathing new life into a building’s structure is vastly more sustainable than the highest quality ‘green’ new builds, which are inherently more carbon-intensive. The demolition of viable older buildings (often including heritage buildings) is increasingly at odds with society’s desire for a lower carbon economy. Bringing seismic resilience and carbon best practice together in decision-making can open the door to commercially attractive outcomes, with retrofits that include structural and sustainability upgrades transforming the asset’s revenue generation. Across the global real estate market, tenants are increasingly demanding the buildings they occupy be resilient and aligned with their own climate targets. The relationship between seismic performance and ‘sustainable design’ has yet to fully mature, yet in a wider context is of profound consequence. A whole-of-life carbon perspective on a building means designing for the likely natural hazards within the asset’s expected lifespan, be that earthquake, storms, damage, bushfires, fires, and so on, ¬with financial mitigation (e.g., insurance) part, but not all, of the picture.

Keywords: retrofit, sustainability, earthquake, reuse, carbon, resilient

Procedia PDF Downloads 61
9953 Parameter Fitting of the Discrete Element Method When Modeling the DISAMATIC Process

Authors: E. Hovad, J. H. Walther, P. Larsen, J. Thorborg, J. H. Hattel

Abstract:

In sand casting of metal parts for the automotive industry such as brake disks and engine blocks, the molten metal is poured into a sand mold to get its final shape. The DISAMATIC molding process is a way to construct these sand molds for casting of steel parts and in the present work numerical simulations of this process are presented. During the process green sand is blown into a chamber and subsequently squeezed to finally obtain the sand mould. The sand flow is modelled with the Discrete Element method (DEM) and obtaining the correct material parameters for the simulation is the main goal. Different tests will be used to find or calibrate the DEM parameters needed; Poisson ratio, Young modulus, rolling friction coefficient, sliding friction coefficient and coefficient of restitution (COR). The Young modulus and Poisson ratio are found from compression tests of the bulk material and subsequently used in the DEM model according to the Hertz-Mindlin model. The main focus will be on calibrating the rolling resistance and sliding friction in the DEM model with respect to the behavior of “real” sand piles. More specifically, the surface profile of the “real” sand pile will be compared to the sand pile predicted with the DEM for different values of the rolling and sliding friction coefficients. When the DEM parameters are found for the particle-particle (sand-sand) interaction, the particle-wall interaction parameter values are also found. Here the sliding coefficient will be found from experiments and the rolling resistance is investigated by comparing with observations of how the green sand interacts with the chamber wall during experiments and the DEM simulations will be calibrated accordingly. The coefficient of restitution will be tested with different values in the DEM simulations and compared to video footages of the DISAMATIC process. Energy dissipation will be investigated in these simulations for different particle sizes and coefficient of restitution, where scaling laws will be considered to relate the energy dissipation for these parameters. Finally, the found parameter values are used in the overall discrete element model and compared to the video footage of the DISAMATIC process.

Keywords: discrete element method, physical properties of materials, calibration, granular flow

Procedia PDF Downloads 471
9952 Linkages between Innovation Policies and SMEs' Innovation Activities: Empirical Evidence from 15 Transition Countries

Authors: Anita Richter

Abstract:

Innovation is one of the key foundations of competitive advantage, generating growth and welfare worldwide. Consequently, all firms should innovate to bring new ideas to the market. Innovation is a vital growth driver, particularly for transition countries to move towards knowledge-based, high-income economies. However, numerous barriers, such as financial, regulatory or infrastructural constraints prevent, in particular, new and small firms in transition countries from innovating. Thus SMEs’ innovation output may benefit substantially from government support. This research paper aims to assess the effect of government interventions on innovation activities in SMEs in emerging countries. Until now academic research related to the innovation policies focused either on single country and/or high-income countries assessments and less on cross-country and/or low and middle-income countries. Therefore the paper seeks to close the research gap by providing empirical evidence from 8,500 firms in 15 transition countries (Eastern Europe, South Caucasus, South East Europe, Middle East and North Africa). Using firm-level data from the Business Environment and Enterprise Performance Survey of the World Bank and EBRD and policy data from the SME Policy Index of the OECD, the paper investigates how government interventions affect SME’s likelihood of investing in any technological and non-technological innovation. Using the Standard Linear Regression, the impact of government interventions on SMEs’ innovation output and R&D activities is measured. The empirical analysis suggests that a firm’s decision to invest into innovative activities is sensitive to government interventions. A firm’s likelihood to invest into innovative activities increases by 3% to 8%, if the innovation eco-system noticeably improves (measured by an increase of 1 level in the SME Policy Index). At the same time, a better eco-system encourages SMEs to invest more in R&D. Government reforms in establishing a dedicated policy framework (IP legislation), institutional infrastructure (science and technology parks, incubators) and financial support (public R&D grants, innovation vouchers) are particularly relevant to stimulate innovation performance in SMEs. Particular segments of the SME population, namely micro and manufacturing firms, are more likely to benefit from an increased innovation framework conditions. The marginal effects are particularly strong on product innovation, process innovation, and marketing innovation, but less on management innovation. In conclusion, government interventions supporting innovation will likely lead to higher innovation performance of SMEs. They increase productivity at both firm and country level, which is a vital step in transitioning towards knowledge-based market economies.

Keywords: innovation, research and development, government interventions, economic development, small and medium-sized enterprises, transition countries

Procedia PDF Downloads 312
9951 Formal Models of Sanitary Inspections Teams Activities

Authors: Tadeusz Nowicki, Radosław Pytlak, Robert Waszkowski, Jerzy Bertrandt, Anna Kłos

Abstract:

This paper presents methods for formal modeling of activities in the area of sanitary inspectors outbreak of food-borne diseases. The models allow you to measure the characteristics of the activities of sanitary inspection and as a result allow improving the performance of sanitary services and thus food security.

Keywords: food-borne disease, epidemic, sanitary inspection, mathematical models

Procedia PDF Downloads 294
9950 A Machine Learning Approach for Detecting and Locating Hardware Trojans

Authors: Kaiwen Zheng, Wanting Zhou, Nan Tang, Lei Li, Yuanhang He

Abstract:

The integrated circuit industry has become a cornerstone of the information society, finding widespread application in areas such as industry, communication, medicine, and aerospace. However, with the increasing complexity of integrated circuits, Hardware Trojans (HTs) implanted by attackers have become a significant threat to their security. In this paper, we proposed a hardware trojan detection method for large-scale circuits. As HTs introduce physical characteristic changes such as structure, area, and power consumption as additional redundant circuits, we proposed a machine-learning-based hardware trojan detection method based on the physical characteristics of gate-level netlists. This method transforms the hardware trojan detection problem into a machine-learning binary classification problem based on physical characteristics, greatly improving detection speed. To address the problem of imbalanced data, where the number of pure circuit samples is far less than that of HTs circuit samples, we used the SMOTETomek algorithm to expand the dataset and further improve the performance of the classifier. We used three machine learning algorithms, K-Nearest Neighbors, Random Forest, and Support Vector Machine, to train and validate benchmark circuits on Trust-Hub, and all achieved good results. In our case studies based on AES encryption circuits provided by trust-hub, the test results showed the effectiveness of the proposed method. To further validate the method’s effectiveness for detecting variant HTs, we designed variant HTs using open-source HTs. The proposed method can guarantee robust detection accuracy in the millisecond level detection time for IC, and FPGA design flows and has good detection performance for library variant HTs.

Keywords: hardware trojans, physical properties, machine learning, hardware security

Procedia PDF Downloads 135
9949 Investigation of Different Machine Learning Algorithms in Large-Scale Land Cover Mapping within the Google Earth Engine

Authors: Amin Naboureh, Ainong Li, Jinhu Bian, Guangbin Lei, Hamid Ebrahimy

Abstract:

Large-scale land cover mapping has become a new challenge in land change and remote sensing field because of involving a big volume of data. Moreover, selecting the right classification method, especially when there are different types of landscapes in the study area is quite difficult. This paper is an attempt to compare the performance of different machine learning (ML) algorithms for generating a land cover map of the China-Central Asia–West Asia Corridor that is considered as one of the main parts of the Belt and Road Initiative project (BRI). The cloud-based Google Earth Engine (GEE) platform was used for generating a land cover map for the study area from Landsat-8 images (2017) by applying three frequently used ML algorithms including random forest (RF), support vector machine (SVM), and artificial neural network (ANN). The selected ML algorithms (RF, SVM, and ANN) were trained and tested using reference data obtained from MODIS yearly land cover product and very high-resolution satellite images. The finding of the study illustrated that among three frequently used ML algorithms, RF with 91% overall accuracy had the best result in producing a land cover map for the China-Central Asia–West Asia Corridor whereas ANN showed the worst result with 85% overall accuracy. The great performance of the GEE in applying different ML algorithms and handling huge volume of remotely sensed data in the present study showed that it could also help the researchers to generate reliable long-term land cover change maps. The finding of this research has great importance for decision-makers and BRI’s authorities in strategic land use planning.

Keywords: land cover, google earth engine, machine learning, remote sensing

Procedia PDF Downloads 106
9948 The Analysis of Changes in Urban Hierarchy of Isfahan Province in the Fifty-Year Period (1956-2006)

Authors: Hamidreza Joudaki, Yousefali Ziari

Abstract:

The appearance of city and urbanism is one of the important processes which have affected social communities. Being industrialized urbanism developed along with each other in the history. In addition, they have had simple relationship for more than six thousand years, that is, from the appearance of the first cities. In 18th century by coming out of industrial capitalism, progressive development took place in urbanism in the world. In Iran, the city of each region made its decision by itself and the capital of region (downtown) was the only central part and also the regional city without any hierarchy, controlled its realm. However, this method of ruling during these three decays, because of changing in political, social and economic issues that have caused changes in rural and urban relationship. Moreover, it has changed the variety of performance of cities and systematic urban network in Iran. Today, urban system has very vast imbalanced apace and performance. In Isfahan, the trend of urbanism is like the other part of Iran and systematic urban hierarchy is not suitable and normal. This article is a quantitative and analytical. The statistical communities are Isfahan Province cities and the changes in urban network and its hierarchy during the period of fifty years (1956 -2006) has been surveyed. In addition, those data have been analyzed by model of Rank and size and Entropy index. In this article Iran cities and also the factor of entropy of primate city and urban hierarchy of Isfahan Province have been introduced. Urban residents of this Province have been reached from 55 percent to 83% (2006). As we see the analytical data reflects that there is mismatching and imbalance between cities. Because the entropy index was.91 in 1956.And it decreased to.63 in 2006. Isfahan city is the primate city in the whole of these periods. Moreover, the second and the third cities have population gap with regard to the other cities and finally, they do not follow the system of rank-size.

Keywords: urban network, urban hierarchy, primate city, Isfahan province, urbanism, first cities

Procedia PDF Downloads 240
9947 Evaluating the Dosimetric Performance for 3D Treatment Planning System for Wedged and Off-Axis Fields

Authors: Nashaat A. Deiab, Aida Radwan, Mohamed S. Yahiya, Mohamed Elnagdy, Rasha Moustafa

Abstract:

This study is to evaluate the dosimetric performance of our institution's 3D treatment planning system for wedged and off-axis 6MV photon beams, guided by the recommended QA tests documented in the AAPM TG53; NCS report 15 test packages, IAEA TRS 430 and ESTRO booklet no.7. The study was performed for Elekta Precise linear accelerator designed for clinical range of 4, 6 and 15 MV photon beams with asymmetric jaws and fully integrated multileaf collimator that enables high conformance to target with sharp field edges. Ten tests were applied on solid water equivalent phantom along with 2D array dose detection system. The calculated doses using 3D treatment planning system PrecisePLAN were compared with measured doses to make sure that the dose calculations are accurate for simple situations such as square and elongated fields, different SSD, beam modifiers e.g. wedges, blocks, MLC-shaped fields and asymmetric collimator settings. The QA results showed dosimetric accuracy of the TPS within the specified tolerance limits. Except for large elongated wedged field, the central axis and outside central axis have errors of 0.2% and 0.5%, respectively, and off- planned and off-axis elongated fields the region outside the central axis of the beam errors are 0.2% and 1.1%, respectively. The dosimetric investigated results yielded differences within the accepted tolerance level as recommended. Differences between dose values predicted by the TPS and measured values at the same point are the result from limitations of the dose calculation, uncertainties in the measurement procedure, or fluctuations in the output of the accelerator.

Keywords: quality assurance, dose calculation, wedged fields, off-axis fields, 3D treatment planning system, photon beam

Procedia PDF Downloads 429
9946 The Predictors of Head and Neck Cancer-Head and Neck Cancer-Related Lymphedema in Patients with Resected Advanced Head and Neck Cancer

Authors: Shu-Ching Chen, Li-Yun Lee

Abstract:

The purpose of the study was to identify the factors associated with head and neck cancer-related lymphoedema (HNCRL)-related symptoms, body image, and HNCRL-related functional outcomes among patients with resected advanced head and neck cancer. A cross-sectional correlational design was conducted to examine the predictors of HNCRL-related functional outcomes in patients with resected advanced head and neck cancer. Eligible patients were recruited from a single medical center in northern Taiwan. Consecutive patients were approached and recruited from the Radiation Head and Neck Outpatient Department of this medical center. Eligible subjects were assessed for the Symptom Distress Scale–Modified for Head and Neck Cancer (SDS-mhnc), Brief International Classification of Functioning, Disability and Health (ICF) Core Set for Head and Neck Cancer (BCSQ-H&N), Body Image Scale–Modified (BIS-m), The MD Anderson Head and Neck Lymphedema Rating Scale (MDAHNLRS), The Foldi’s Stages of Lymphedema (Foldi’s Scale), Patterson’s Scale, UCLA Shoulder Rating Scale (UCLA SRS), and Karnofsky’s Performance Status Index (KPS). The results showed that the worst problems with body HNCRL functional outcomes. Patients’ HNCRL symptom distress and performance status are robust predictors across over for overall HNCRL functional outcomes, problems with body HNCRL functional outcomes, and activity and social functioning HNCRL functional outcomes. Based on the results of this period research program, we will develop a Cancer Rehabilitation and Lymphedema Care Program (CRLCP) to use in the care of patients with resected advanced head and neck cancer.

Keywords: head and neck cancer, resected, lymphedema, symptom, body image, functional outcome

Procedia PDF Downloads 240
9945 The Design and Implementation of an Enhanced 2D Mesh Switch

Authors: Manel Langar, Riad Bourguiba, Jaouhar Mouine

Abstract:

In this paper, we propose the design and implementation of an enhanced wormhole virtual channel on chip router. It is a heart of a mesh NoC using the XY deterministic routing algorithm. It is characterized by its simple virtual channel allocation strategy which allows reducing area and complexity of connections without affecting the performance. We implemented our router on a Tezzaron process to validate its performances. This router is a basic element that will be used later to design a 3D mesh NoC.

Keywords: NoC, mesh, router, 3D NoC

Procedia PDF Downloads 554
9944 Design of a Remote Radiation Sensing Module Based on Portable Gamma Spectrometer

Authors: Young Gil Kim, Hye Min Park, Chan Jong Park, Koan Sik Joo

Abstract:

A personal gamma spectrometer has to be sensitive, pocket-sized, and carriable on the users. To serve these requirements, we developed the SiPM-based portable radiation detectors. The prototype uses a Ce:GAGG scintillator coupled to a silicon photomultiplier and a radio frequency(RF) module to measure gamma-ray, and can be accessed wirelessly or remotely by mobile equipment. The prototype device consumes roughly 4.4W, weighs about 180g (including battery), and measures 5.0 7.0. It is able to achieve 5.8% FWHM energy resolution at 662keV.

Keywords: Ce:GAGG, gamma-ray, radio frequency, silicon photomultiplier

Procedia PDF Downloads 315
9943 Digital Twins in the Built Environment: A Systematic Literature Review

Authors: Bagireanu Astrid, Bros-Williamson Julio, Duncheva Mila, Currie John

Abstract:

Digital Twins (DT) are an innovative concept of cyber-physical integration of data between an asset and its virtual replica. They have originated in established industries such as manufacturing and aviation and have garnered increasing attention as a potentially transformative technology within the built environment. With the potential to support decision-making, real-time simulations, forecasting abilities and managing operations, DT do not fall under a singular scope. This makes defining and leveraging the potential uses of DT a potential missed opportunity. Despite its recognised potential in established industries, literature on DT in the built environment remains limited. Inadequate attention has been given to the implementation of DT in construction projects, as opposed to its operational stage applications. Additionally, the absence of a standardised definition has resulted in inconsistent interpretations of DT in both industry and academia. There is a need to consolidate research to foster a unified understanding of the DT. Such consolidation is indispensable to ensure that future research is undertaken with a solid foundation. This paper aims to present a comprehensive systematic literature review on the role of DT in the built environment. To accomplish this objective, a review and thematic analysis was conducted, encompassing relevant papers from the last five years. The identified papers are categorised based on their specific areas of focus, and the content of these papers was translated into a through classification of DT. In characterising DT and the associated data processes identified, this systematic literature review has identified 6 DT opportunities specifically relevant to the built environment: Facilitating collaborative procurement methods, Supporting net-zero and decarbonization goals, Supporting Modern Methods of Construction (MMC) and off-site manufacturing (OSM), Providing increased transparency and stakeholders collaboration, Supporting complex decision making (real-time simulations and forecasting abilities) and Seamless integration with Internet of Things (IoT), data analytics and other DT. Finally, a discussion of each area of research is provided. A table of definitions of DT across the reviewed literature is provided, seeking to delineate the current state of DT implementation in the built environment context. Gaps in knowledge are identified, as well as research challenges and opportunities for further advancements in the implementation of DT within the built environment. This paper critically assesses the existing literature to identify the potential of DT applications, aiming to harness the transformative capabilities of data in the built environment. By fostering a unified comprehension of DT, this paper contributes to advancing the effective adoption and utilisation of this technology, accelerating progress towards the realisation of smart cities, decarbonisation, and other envisioned roles for DT in the construction domain.

Keywords: built environment, design, digital twins, literature review

Procedia PDF Downloads 64
9942 Comparison of Mcgrath, Pentax, and Macintosh Laryngoscope in Normal and Cervical Immobilized Manikin by Novices

Authors: Jong Yeop Kim, In Kyong Yi, Hyun Jeong Kwak, Sook Young Lee, Sung Yong Park

Abstract:

Background: Several video laryngoscopes (VLs) were used to facilitate tracheal intubation in the normal and potentially difficult airway, especially by novice personnel. The aim of this study was to compare tracheal intubation performance regarding the time to intubation, glottic view, difficulty, and dental click, by a novice using McGrath VL, Pentax Airway Scope (AWS) and Macintosh laryngoscope in normal and cervical immobilized manikin models. Methods: Thirty-five anesthesia nurses without previous intubation experience were recruited. The participants performed endotracheal intubation in a manikin model at two simulated neck positions (normal and fixed neck via cervical immobilization), using three different devices (McGrath VL, Pentax AWS, and Macintosh direct laryngoscope) at three times each. Performance parameters included intubation time, success rate of intubation, Cormack Lehane laryngoscope grading, dental click, and subjective difficulty score. Results: Intubation time and success rate at the first attempt were not significantly different between the 3 groups in normal airway manikin. In the cervical immobilized manikin, the intubation time was shorter (p = 0.012) and the success rate with the first attempt was significantly higher (p < 0.001) when using McGrath VL and Pentax AWS compared with Macintosh laryngoscope. Both VLs showed less difficulty score (p < 0.001) and more Cormack Lehane grade I (p < 0.001). The incidence of dental clicks was higher with McGrath VL than Macintosh laryngoscope in the normal and cervical immobilized airway (p = 0.005, p < 0.001, respectively). Conclusion: McGrath VL and Pentax AWS resulted in shorter intubation time, higher first attempt success rate, compared with Macintosh laryngoscope by a novice intubator in a cervical immobilized manikin model. McGrath VL could be reduced the risk of dental injury compared with Macintosh laryngoscope in this scenario.

Keywords: intubation, manikin, novice, videolaryngoscope

Procedia PDF Downloads 143
9941 Study on the Electrochemical Performance of Graphene Effect on Cadmium Oxide in Lithium Battery

Authors: Atef Y. Shenouda, Anton A. Momchilov

Abstract:

Graphene and CdO with different stoichiometric ratios of Cd(CH₃COO)₂ and graphene samples were prepared by hydrothermal reaction. The crystalline phases of pure CdO and 3CdO:1graphene were identified by X-ray diffraction (XRD). The particle morphology was studied with SEM. Furthermore, impedance measurements were applied. Galvanostatic measurements for the cells were carried out using potential limits between 0.01 and 3 V vs. Li/Li⁺. The current cycling intensity was 10⁻⁴ A. The specific discharge capacity of 3CdO-1G cell was about 450 Ah.Kg⁻¹ up to more than 100 cycles.

Keywords: CdO, graphene, negative electrode, lithium battery

Procedia PDF Downloads 148
9940 Effectiveness of Jute Geotextiles for Hill Slope Stabilization in Adverse Climatic Condition

Authors: Pradip Choudhury, Tapobrata Sanyal

Abstract:

Effectiveness of Jute Geotextiles (JGT) in hill slope management now stands substantiated. The reasons of its efficacy are attributed to its bio-degradability, hygroscopic property and its thickness. Usually open weave JGT is used for slope management. Thickness of JGT helps in reducing the velocity of surface run-off, thus curbing the extent of migration of soil particles detached as a result of kinetic energy of rain-drops and also of wind effects. Initially JGT acts as cover of the surface of slope thus protect movement of loose soil particles. Hygroscopic property of jute effects overland storage of the flow. JGT acts as mulch and creates a congenial micro-climate that fosters quick growth of vegetation on bio-degradation. In fact JGT plays an important role in bio-remediation of slope-erosion problems. Considering the environmental aftermath, JGT is the preferred option in developed countries for surface soil conservation against erosion. In India JGT has not been tried in low temperature zones at high altitudes where temperature goes below the freezing point (even below - 25° Celsius). The behavior of JGT in such low-temperature zones is not precisely known. The 16th BRTF of Project Himank of Border Roads Organization (BRO) has recently taken the initiative to try two varieties of JGT , ie, 292 gsm and 500 gsm at two different places for hill slope management in Leh, a high altitude place of about 2,660 mtrs and 4900 mtrs above MSL respectively in Jammu & Kashmir where erosion is caused more as a result of rapid movement of sand particles due to high wind (wind erosion. Soil particles of the region formed naturally by weathering of fragile rocks are usually loosely bonded (non-cohesive), undergo dissociation with the rise in wind force and kinetic energy of rain drops and are blown away by wind. Open weave JGT interestingly was observed to contain the dissociated soil particles within its pores and lend stability the affected soil mass to a great extent thus preventing its movement by extraneous agents such as wind. The paper delineates about climatic factors, type of JGT used and the prevailing site conditions with an attempt to analyze the mechanism of functioning of JGT in low temperature zones.

Keywords: climate, erosion, jutegeotextile, stabilize

Procedia PDF Downloads 417
9939 Effect of Garlic Extract on Growth Performance and Immune System of Broiler

Authors: Merry Muspita Dyah Utami

Abstract:

The positive effect of garlic extract have been reported by many studies. It has antibiotical potential, antibacterial, antiviral, antiparasitic, antifungal, and growth promoting. Supplementary garlic for broilers could mediate in getting the bioactive compounds in garlic. The avian bursa must be essential for antibody-mediated immunity. The size of bursa of fabricius must be some sort of endocrine or lymphoid gland associated with growth and sexual development. The research was conducted to evaluate the effects of garlic extract on growth performance and immune system of broiler. Seventy-two day old chick were equally divided into four group, three replication and six chicks each. Group I was control without garlic extract, then garlic extraxt was administrated to the experimental group II, III and IV (2, 4, 6% in ration). The experiment was conducted for three weeks period from day old chick to 21 days. Body weight of broiler were determined at day 1 and 21, feed intake was determined at the same period, feed conversion ratio was calculated accordingly. At 21 day age, four birds per replicate were slaughtered , bursa was collected, weight and calculated as a percentage of live body weight. Mortality was recorded as it occurred and was used to ajust the total number of broiler to determine the total feed intake and feed conversion rasio. Data were expressed as the mean was compare by one way analysis of variance (Anova) follow by Duncan Test, which used to identify differences between groups. A value of P<0.05 was accepted as significance. The body weight, feed conversion rasio, and the weight of bursa of fabricius showed a significant differences, but feed consumption and the percentage of bursa of live body weight were not significantly different (P > 0.05) influenced by dietary treatments. The results of this research, garlic extract has a potential role as natural growth promoter and immunomodulatory system in broiler.

Keywords: garlic extract, growth, immunity, broiler

Procedia PDF Downloads 318
9938 Constructing the Cult of the Self: on White, Working-class Males And The Neoliberalisation Of Identities – An Autoethnographic Study

Authors: Dane Morace-Court

Abstract:

This paper offers a reflective and reflexive examination of the lived experience of a group of young, white, working-class males engaging in secondary-education in England at a time when this population is widely recognised as the lowest attaining ethnic group within British schools. The focus of the paper is an exploration of the development of identities and aspirations, alongside contemporary demographic and ideological shifts in the British population, in their intersection with neoliberal education policies and the emerging ideological conflict between identity conservatism and liberalism. The construction and performance of intersecting social-class, gender, ethnic and national identities is considered as well as the process through which socially constructed narratives inform identities, values, and aspirations. Evocative autoethnography is then employed to offer reflections on working-class habitus and, in particular, classed and gendered codes that underpin expectations of manhood in post-industrial culture within an education system which seemingly requires the abandonment of aspects of a working-class background. Findings from the study identify the emergence of a culture of hyper-individualisation amongst white, working-class males in schools and a belief in the meritocratic ideologies of the New Right. In particular, the breakdown of the social contract, including notions of political and civic responsibility, coupled with the symbolic violence perpetrated against working-class culture and solidarity in British schools, have all informed the construction of a working-class masculinity which values the individual entrepreneur over the collective, and depoliticizes students to an extent where a focus on the spectacle and performance of success has replaced individual and collective investment in community.

Keywords: education, identity, masculinity, neoliberalism, working-class, intersectionality, autoethnography

Procedia PDF Downloads 98
9937 Estimation of the Exergy-Aggregated Value Generated by a Manufacturing Process Using the Theory of the Exergetic Cost

Authors: German Osma, Gabriel Ordonez

Abstract:

The production of metal-rubber spares for vehicles is a sequential process that consists in the transformation of raw material through cutting activities and chemical and thermal treatments, which demand electricity and fossil fuels. The energy efficiency analysis for these cases is mostly focused on studying of each machine or production step, but is not common to study of the quality of the production process achieves from aggregated value viewpoint, which can be used as a quality measurement for determining of impact on the environment. In this paper, the theory of exergetic cost is used for determining of aggregated exergy to three metal-rubber spares, from an exergy analysis and thermoeconomic analysis. The manufacturing processing of these spares is based into batch production technique, and therefore is proposed the use of this theory for discontinuous flows from of single models of workstations; subsequently, the complete exergy model of each product is built using flowcharts. These models are a representation of exergy flows between components into the machines according to electrical, mechanical and/or thermal expressions; they determine the demanded exergy to produce the effective transformation in raw materials (aggregated exergy value), the exergy losses caused by equipment and irreversibilities. The energy resources of manufacturing process are electricity and natural gas. The workstations considered are lathes, punching presses, cutters, zinc machine, chemical treatment tanks, hydraulic vulcanizing presses and rubber mixer. The thermoeconomic analysis was done by workstation and by spare; first of them describes the operation of the components of each machine and where the exergy losses are; while the second of them estimates the exergy-aggregated value for finished product and wasted feedstock. Results indicate that exergy efficiency of a mechanical workstation is between 10% and 60% while this value in the thermal workstations is less than 5%; also that each effective exergy-aggregated value is one-thirtieth of total exergy required for operation of manufacturing process, which amounts approximately to 2 MJ. These troubles are caused mainly by technical limitations of machines, oversizing of metal feedstock that demands more mechanical transformation work, and low thermal insulation of chemical treatment tanks and hydraulic vulcanizing presses. From established information, in this case, it is possible to appreciate the usefulness of theory of exergetic cost for analyzing of aggregated value in manufacturing processes.

Keywords: exergy-aggregated value, exergy efficiency, thermoeconomics, exergy modeling

Procedia PDF Downloads 158
9936 Effects of Foam Rolling with Different Application Volumes on the Isometric Force of the Calf Muscle with Consideration of Muscle Activity

Authors: T. Poppendieker, H. Maurer, C. Segieth

Abstract:

Over the past ten years, foam rolling has become a new trend in the fitness and health market. It is also a frequently used technique for self-massage. However, the scope of effects from foam rolling has only recently started to be researched and understood. The focus of this study is to examine the effects of prolonged foam rolling on muscle performance. Isometric muscle force was used as a parameter to determine an improving impact of the myofascial roller in two different application volumes. Besides the maximal muscle force, data were also collected on muscle activation during all tests. Twenty-four (17 females, 7 males) healthy students with an average age of 23.4 ± 2.8 years were recruited. The study followed a cross-over pre-/post design in which the order of conditions was counterbalanced. The subjects performed a one-minute and three-minute foam rolling application set on two separate days. Isometric maximal muscle force of the dominant calf was tested before and after the self-myofascial release application. The statistic software program SPSS 22 was used to analyze the data of the maximal isometric force of the calf muscle by a 2 x 2 (time of measurement x intervention) analysis of variance with repeated measures. The statistic significance level was set at p ≤ 0.05. Neither for the main effect of time of measurement (F(1,23) = .93, p = .36, f = .20) nor for the interaction of time of measurement x intervention (F(1,23) = 1.99, p = .17, f = 0.29) significant p-values were found. However, the effect size indicates a mean interaction effect with a tendency of greater pre-post improvements under the three-minute foam rolling condition. Changes in maximal force did not correlate with changes in EMG-activity (r = .02, p = .95 in the short and r = -.11, p = .65 in the long rolling condition). Results support findings of previous studies and suggest a positive potential for use of the foam roll as a means for keeping muscle force at least at the same performance level while leading to an increase in flexibility.

Keywords: application volume differences, foam rolling, isometric maximal force, self-myofascial release

Procedia PDF Downloads 278
9935 Aerosol Direct Radiative Forcing Over the Indian Subcontinent: A Comparative Analysis from the Satellite Observation and Radiative Transfer Model

Authors: Shreya Srivastava, Sagnik Dey

Abstract:

Aerosol direct radiative forcing (ADRF) refers to the alteration of the Earth's energy balance from the scattering and absorption of solar radiation by aerosol particles. India experiences substantial ADRF due to high aerosol loading from various sources. These aerosols' radiative impact depends on their physical characteristics (such as size, shape, and composition) and atmospheric distribution. Quantifying ADRF is crucial for understanding aerosols’ impact on the regional climate and the Earth's radiative budget. In this study, we have taken radiation data from Clouds and the Earth’s Radiant Energy System (CERES, spatial resolution=1ox1o) for 22 years (2000-2021) over the Indian subcontinent. Except for a few locations, the short-wave DARF exhibits aerosol cooling at the TOA (values ranging from +2.5 W/m2 to -22.5W/m2). Cooling due to aerosols is more pronounced in the absence of clouds. Being an aerosol hotspot, higher negative ADRF is observed over the Indo-Gangetic Plain (IGP). Aerosol Forcing Efficiency (AFE) shows a decreasing seasonal trend in winter (DJF) over the entire study region while an increasing trend over IGP and western south India during the post-monsoon season (SON) in clear-sky conditions. Analysing atmospheric heating and AOD trends, we found that only the aerosol loading is not governing the change in atmospheric heating but also the aerosol composition and/or their vertical profile. We used a Multi-angle Imaging Spectro-Radiometer (MISR) Level-2 Version 23 aerosol products to look into aerosol composition. MISR incorporates 74 aerosol mixtures in its retrieval algorithm based on size, shape, and absorbing properties. This aerosol mixture information was used for analysing long-term changes in aerosol composition and dominating aerosol species corresponding to the aerosol forcing value. Further, ADRF derived from this method is compared with around 35 studies across India, where a plane parallel Radiative transfer model was used, and the model inputs were taken from the OPAC (Optical Properties of Aerosols and Clouds) utilizing only limited aerosol parameter measurements. The result shows a large overestimation of TOA warming by the latter (i.e., Model-based method).

Keywords: aerosol radiative forcing (ARF), aerosol composition, MISR, CERES, SBDART

Procedia PDF Downloads 39
9934 An Efficient Algorithm for Solving the Transmission Network Expansion Planning Problem Integrating Machine Learning with Mathematical Decomposition

Authors: Pablo Oteiza, Ricardo Alvarez, Mehrdad Pirnia, Fuat Can

Abstract:

To effectively combat climate change, many countries around the world have committed to a decarbonisation of their electricity, along with promoting a large-scale integration of renewable energy sources (RES). While this trend represents a unique opportunity to effectively combat climate change, achieving a sound and cost-efficient energy transition towards low-carbon power systems poses significant challenges for the multi-year Transmission Network Expansion Planning (TNEP) problem. The objective of the multi-year TNEP is to determine the necessary network infrastructure to supply the projected demand in a cost-efficient way, considering the evolution of the new generation mix, including the integration of RES. The rapid integration of large-scale RES increases the variability and uncertainty in the power system operation, which in turn increases short-term flexibility requirements. To meet these requirements, flexible generating technologies such as energy storage systems must be considered within the TNEP as well, along with proper models for capturing the operational challenges of future power systems. As a consequence, TNEP formulations are becoming more complex and difficult to solve, especially for its application in realistic-sized power system models. To meet these challenges, there is an increasing need for developing efficient algorithms capable of solving the TNEP problem with reasonable computational time and resources. In this regard, a promising research area is the use of artificial intelligence (AI) techniques for solving large-scale mixed-integer optimization problems, such as the TNEP. In particular, the use of AI along with mathematical optimization strategies based on decomposition has shown great potential. In this context, this paper presents an efficient algorithm for solving the multi-year TNEP problem. The algorithm combines AI techniques with Column Generation, a traditional decomposition-based mathematical optimization method. One of the challenges of using Column Generation for solving the TNEP problem is that the subproblems are of mixed-integer nature, and therefore solving them requires significant amounts of time and resources. Hence, in this proposal we solve a linearly relaxed version of the subproblems, and trained a binary classifier that determines the value of the binary variables, based on the results obtained from the linearized version. A key feature of the proposal is that we integrate the binary classifier into the optimization algorithm in such a way that the optimality of the solution can be guaranteed. The results of a study case based on the HRP 38-bus test system shows that the binary classifier has an accuracy above 97% for estimating the value of the binary variables. Since the linearly relaxed version of the subproblems can be solved with significantly less time than the integer programming counterpart, the integration of the binary classifier into the Column Generation algorithm allowed us to reduce the computational time required for solving the problem by 50%. The final version of this paper will contain a detailed description of the proposed algorithm, the AI-based binary classifier technique and its integration into the CG algorithm. To demonstrate the capabilities of the proposal, we evaluate the algorithm in case studies with different scenarios, as well as in other power system models.

Keywords: integer optimization, machine learning, mathematical decomposition, transmission planning

Procedia PDF Downloads 72
9933 Solar Cell Degradation by Electron Irradiation Effect of Irradiation Fluence

Authors: H. Mazouz, A. Belghachi, F. Hadjaj

Abstract:

Solar cells used in orbit are exposed to radiation environment mainly protons and high energy electrons. These particles degrade the output parameters of the solar cell. The aim of this work is to characterize the effects of electron irradiation fluence on the J (V) characteristic and output parameters of gaAs solar cell by numerical simulation. The results obtained demonstrate that the electron irradiation-induced degradation of performances of the cells concerns mainly the short circuit current.

Keywords: gaAs solar cell, MeV electron irradiation, irradiation fluence, short circuit

Procedia PDF Downloads 449
9932 Employee Engagement

Authors: Jai Bakliya, Palak Dhamecha

Abstract:

Today customer satisfaction is given utmost priority be it any industry. But when it comes to hospitality industry this applies even more as they come in direct contact with customers while providing them services. Employee engagement is new concept adopted by Human Resource Department which impacts customer satisfactions. To satisfy your customers, it is necessary to see that the employees in the organisation are satisfied and engaged enough in their work that they meet the company’s expectations and contribute in the process of achieving company’s goals and objectives. After all employees is human capital of the organisation. Employee engagement has become a top business priority for every organisation. In this fast moving economy, business leaders know that having a potential and high-performing human resource is important for growth and survival. They recognize that a highly engaged manpower can increase innovation, productivity, and performance, while reducing costs related to retention and hiring in highly competitive talent markets. But while most executives see a clear need to improve employee engagement, many have yet to develop tangible ways to measure and tackle this goal. Employee Engagement is an approach which is applied to establish an emotional connection between an employee and the organisation which ensures the employee’s commitment towards his work which affects the productivity and overall performance of the organisation. The study was conducted in hospitality industry. A popular branded hotel was chosen as a sample unit. Data were collected, both qualitative and quantitative from respondents. It is found that employee engagement level of the organisation (Hotel) is quite low. This means that employees are not emotionally connected with the organisation which may in turn, affect performance of the employees it is important to note that in hospitality industry individual employee’s performance specifically in terms of emotional engagement is critical and, therefore, a low engagement level may contribute to low organisation performance. An attempt to this study was made to identify employee engagement level. Another objective to take this study was to explore the factors impeding employee engagement and to explore employee engagement facilitation. While in the hospitality industry where people tend to work for as long as 16 to 18 hours concepts like employee engagement is essential. Because employees get tired of their routine job and in case where job rotation cannot be done employee engagement acts as a solution. The study was conducted at Trident Hotel, Udaipur. It was conducted on the sample size of 30 in-house employees from 6 different departments. The various departments were: Accounts and General, Front Office, Food & Beverage Service, Housekeeping, Food & Beverage Production and Engineering. It was conducted with the help of research instrument. The research instrument was Questionnaire. Data collection source was primary source. Trident Udaipur is one of the busiest hotels in Udaipur. The occupancy rate of the guest over there is nearly 80%. Due the high occupancy rate employees or staff of the hotel used to remain very busy and occupied all the time in their work. They worked for their remuneration only. As a result, they do not have any encouragement for their work nor they are interested in going an extra mile for the organisation. The study result shows working environment factors including recognition and appreciation, opinions of the employee, counselling, feedback from superiors, treatment of managers and respect from the organisation are capable of increasing employee engagement level in the hotel. The above study result encouraged us to explore the factors contributed to low employee engagement. It is being found that factors such as recognition and appreciation, feedback from supervisors, opinion of the employee, counselling, feedback from supervisors, treatment from managers has contributed negatively to employee engagement level. Probable reasons for the low contribution are number of employees gave the negative feedback in accordance to the factors stated above of the organisation. It seems that the structure of organisation itself is responsible for the low contribution of employee engagement. The scope of this study is limited to trident hotel situated in the Udaipur. The limitation of the study was that that the results or findings were only based on the responses of respondents of Trident, Udaipur. And so the recommendations were also applicable in Trident, Udaipur and not to all the like organisations across the country. Through the data collected was further analysed, interpreted and concluded. On the basis of the findings, suggestions were provided to the hotel for improvisation.

Keywords: human resource, employee engagement, research, study

Procedia PDF Downloads 299
9931 High Temperature Volume Combustion Synthesis of Ti3Al with Low Porosities

Authors: Nese Ozturk Korpe, Muhammed H. Karas

Abstract:

Reaction synthesis, or combustion synthesis, is a processing technique in which the thermal activation energy of formation of a compound is sustained by its exothermic heat of reaction. The aim of the present study was to investigate the effect of high initial pressing pressures (420 MPa, 630 MPa, and 850 MPa) on porosity of Ti3Al which produced by volume combustion synthesis. Microstructure examinations were performed by optical microscope (OM) and scanning electron microscope (SEM). Phase analyses were performed with X-ray diffraction device (XRD). A significant decrease in porosity was obtained due to an increase in the initial pressing pressure.

Keywords: Titanium Aluminide, Volume Combustion Synthesis, Intermetallic, Porosity

Procedia PDF Downloads 160
9930 Vicarious Cues in Portraying Emotion: Musicians' Self-Appraisal

Authors: W. Linthicum-Blackhorse, P. Martens

Abstract:

This present study seeks to discover attitudinal commonalities and differences within a musician population relative to the communication of emotion via music. We hypothesized that instrument type, as well as age and gender, would bear significantly on musicians’ opinions. A survey was administered to 178 participants; 152 were current music majors (mean age 20.3 years, 62 female) and 26 were adult participants in a community choir (mean age 54.0 years, 12 female). The adult participants were all vocalists, while student participants represented the full range of orchestral instruments. The students were grouped by degree program, (performance, music education, or other) and instrument type (voice, brass, woodwinds, strings, percussion). The survey asked 'How important are each of the following areas to you for portraying emotion in music?' Participants were asked to rate each of 15 items on a scale of 1 (not at all important) to 10 (very important). Participants were also instructed to leave blank any item that they did not understand. The 15 items were: dynamic contrast, overall volume, phrasing, facial expression, staging (placement), pitch accuracy, tempo changes, bodily movement, your mood, your attitude, vibrato, rubato, stage/room lighting, clothing type, and clothing color. Contrary to our hypothesis, there was no overall effect of gender or age, and neither did any single response item show a significant difference due to these subject parameters. Among the student participants, however, one-way ANOVA revealed a significant effect of degree program on the rated importance of four items: dynamic contrast, tempo changes, vibrato, and rubato. Significant effects of instrument type were found in the responses to eight items: facial expression, staging, body movement, vibrato, rubato, lighting, clothing type, and clothing color. Post hoc comparisons (Tukey) show that some variation follows from obvious differences between instrument types (e.g. string players are more concerned with vibrato than everyone but woodwind players; vocalists are significantly more concerned with facial expression than everyone but string players), but other differences could point to communal mindsets toward vicarious cues within instrument type. These mindsets could be global (e.g. brass players deeming body movement significantly less important than string players, being less often featured as soloists and appearing less often at the front of the stage) or local (e.g. string players being significantly more concerned than all other groups about both clothing color and type, perhaps due to the strongly-expressed opinions of specific teachers). Future work will attempt to identify the source of these self-appraisals, whether enculturated via explicit pedagogy, or whether absorbed from individuals' observations and performance experience.

Keywords: performance, vicarious cues, communication, emotion

Procedia PDF Downloads 97
9929 Understanding the Interactive Nature in Auditory Recognition of Phonological/Grammatical/Semantic Errors at the Sentence Level: An Investigation Based upon Japanese EFL Learners’ Self-Evaluation and Actual Language Performance

Authors: Hirokatsu Kawashima

Abstract:

One important element of teaching/learning listening is intensive listening such as listening for precise sounds, words, grammatical, and semantic units. Several classroom-based investigations have been conducted to explore the usefulness of auditory recognition of phonological, grammatical and semantic errors in such a context. The current study reports the results of one such investigation, which targeted auditory recognition of phonological, grammatical, and semantic errors at the sentence level. 56 Japanese EFL learners participated in this investigation, in which their recognition performance of phonological, grammatical and semantic errors was measured on a 9-point scale by learners’ self-evaluation from the perspective of 1) two types of similar English sound (vowel and consonant minimal pair words), 2) two types of sentence word order (verb phrase-based and noun phrase-based word orders), and 3) two types of semantic consistency (verb-purpose and verb-place agreements), respectively, and their general listening proficiency was examined using standardized tests. A number of findings have been made about the interactive relationships between the three types of auditory error recognition and general listening proficiency. Analyses based on the OPLS (Orthogonal Projections to Latent Structure) regression model have disclosed, for example, that the three types of auditory error recognition are linked in a non-linear way: the highest explanatory power for general listening proficiency may be attained when quadratic interactions between auditory recognition of errors related to vowel minimal pair words and that of errors related to noun phrase-based word order are embraced (R2=.33, p=.01).

Keywords: auditory error recognition, intensive listening, interaction, investigation

Procedia PDF Downloads 500
9928 Self-Evaluation of the Foundation English Language Programme at the Center for Preparatory Studies Offered at the Sultan Qaboos University, Oman: Process and Findings

Authors: Meenalochana Inguva

Abstract:

The context: The Center for Preparatory study is one of the strongest and most vibrant academic teaching units of the Sultan Qaboos University (SQU). The Foundation Programme English Language (FPEL) is part of a larger foundation programme which was implemented at SQU in fall 2010. The programme has been designed to prepare the students who have been accepted to study in the university in order to achieve the required educational goals (the learning outcomes) that have been designed according to Oman Academic Standards and published by the Omani Authority for Academic Accreditation (OAAA) for the English language component. The curriculum: At the CPS, the English language curriculum is based on the learning outcomes drafted for each level. These learning outcomes guide the students in meeting what is expected of them by the end of each level. These six levels are progressive in nature and are seen as a continuum. The study: A periodic evaluation of language programmes is necessary to improve the quality of the programmes and to meet the set goals of the programmes. An evaluation may be carried out internally or externally depending on the purpose and context. A self-study programme was initiated at the beginning of spring semester 2015 with a team comprising a total of 11 members who worked with-in the assigned course areas (level and programme specific). Only areas specific to FPEL have been included in the study. The study was divided into smaller tasks and members focused on their assigned courses. The self-study primarily focused on analyzing the programme LOs, curriculum planning, materials used and their relevance against the GFP exit standards. The review team also reflected on the assessment methods and procedures followed to reflect on student learning. The team has paid attention to having standard criteria for assessment and transparency in procedures. A special attention was paid to the staging of LOs across levels to determine students’ language and study skills ability to cope with higher level courses. Findings: The findings showed that most of the LOs are met through the materials used for teaching. Students score low on objective tests and high on subjective tests. Motivated students take advantage of academic support activities others do not utilize the student support activities to their advantage. Reading should get more hours. In listening, the format of the listening materials in CT 2 does not match the test format. Some of the course materials need revision. For e.g. APA citation, referencing etc. No specific time is allotted for teaching grammar Conclusion: The findings resulted in taking actions in bridging gaps. It will also help the center to be better prepared for the external review of its FPEL curriculum. It will also provide a useful base to prepare for the self-study portfolio for GFP standards assessment and future audit.

Keywords: curriculum planning, learning outcomes, reflections, self-evaluation

Procedia PDF Downloads 209
9927 The Regulation on Human Exposure to Electromagnetic Fields for Brazilian Power System

Authors: Hugo Manoel Olivera Da Silva, Ricardo Silva Thé Pontes

Abstract:

In this work, is presented an analysis of the Brazilian regulation on human exposure to electromagnetic fields, which provides limits to electric fields, magnetic and electromagnetic fields. The regulations for the electricity sector was in charge of the Agência Nacional de Energia Elétrica-ANEEL, the Brazilian Electricity Regulatory Agency, that made it through the Normative Resolution Nº 398/2010, resulting in a series of obligations for the agents of the electricity sector, especially in the areas of generation, transmission, and distribution.

Keywords: adverse effects, electric energy, electric and magnetic fields, human health, regulation

Procedia PDF Downloads 588