Search results for: real time expression
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 21829

Search results for: real time expression

15649 Empirical Mode Decomposition Based Denoising by Customized Thresholding

Authors: Wahiba Mohguen, Raïs El’hadi Bekka

Abstract:

This paper presents a denoising method called EMD-Custom that was based on Empirical Mode Decomposition (EMD) and the modified Customized Thresholding Function (Custom) algorithms. EMD was applied to decompose adaptively a noisy signal into intrinsic mode functions (IMFs). Then, all the noisy IMFs got threshold by applying the presented thresholding function to suppress noise and to improve the signal to noise ratio (SNR). The method was tested on simulated data and real ECG signal, and the results were compared to the EMD-Based signal denoising methods using the soft and hard thresholding. The results showed the superior performance of the proposed EMD-Custom denoising over the traditional approach. The performances were evaluated in terms of SNR in dB, and Mean Square Error (MSE).

Keywords: customized thresholding, ECG signal, EMD, hard thresholding, soft-thresholding

Procedia PDF Downloads 292
15648 Colored Image Classification Using Quantum Convolutional Neural Networks Approach

Authors: Farina Riaz, Shahab Abdulla, Srinjoy Ganguly, Hajime Suzuki, Ravinesh C. Deo, Susan Hopkins

Abstract:

Recently, quantum machine learning has received significant attention. For various types of data, including text and images, numerous quantum machine learning (QML) models have been created and are being tested. Images are exceedingly complex data components that demand more processing power. Despite being mature, classical machine learning still has difficulties with big data applications. Furthermore, quantum technology has revolutionized how machine learning is thought of, by employing quantum features to address optimization issues. Since quantum hardware is currently extremely noisy, it is not practicable to run machine learning algorithms on it without risking the production of inaccurate results. To discover the advantages of quantum versus classical approaches, this research has concentrated on colored image data. Deep learning classification models are currently being created on Quantum platforms, but they are still in a very early stage. Black and white benchmark image datasets like MNIST and Fashion MINIST have been used in recent research. MNIST and CIFAR-10 were compared for binary classification, but the comparison showed that MNIST performed more accurately than colored CIFAR-10. This research will evaluate the performance of the QML algorithm on the colored benchmark dataset CIFAR-10 to advance QML's real-time applicability. However, deep learning classification models have not been developed to compare colored images like Quantum Convolutional Neural Network (QCNN) to determine how much it is better to classical. Only a few models, such as quantum variational circuits, take colored images. The methodology adopted in this research is a hybrid approach by using penny lane as a simulator. To process the 10 classes of CIFAR-10, the image data has been translated into grey scale and the 28 × 28-pixel image containing 10,000 test and 50,000 training images were used. The objective of this work is to determine how much the quantum approach can outperform a classical approach for a comprehensive dataset of color images. After pre-processing 50,000 images from a classical computer, the QCNN model adopted a hybrid method and encoded the images into a quantum simulator for feature extraction using quantum gate rotations. The measurements were carried out on the classical computer after the rotations were applied. According to the results, we note that the QCNN approach is ~12% more effective than the traditional classical CNN approaches and it is possible that applying data augmentation may increase the accuracy. This study has demonstrated that quantum machine and deep learning models can be relatively superior to the classical machine learning approaches in terms of their processing speed and accuracy when used to perform classification on colored classes.

Keywords: CIFAR-10, quantum convolutional neural networks, quantum deep learning, quantum machine learning

Procedia PDF Downloads 107
15647 Assessing the Effect of the Position of the Cavities on the Inner Plate of the Steel Shear Wall under Time History Dynamic Analysis

Authors: Masoud Mahdavi, Mojtaba Farzaneh Moghadam

Abstract:

The seismic forces caused by the waves created in the depths of the earth during the earthquake hit the structure and cause the building to vibrate. Creating large seismic forces will cause low-strength sections in the structure to suffer extensive surface damage. The use of new steel shear walls in steel structures has caused the strength of the building and its main members (columns) to increase due to the reduction and depreciation of seismic forces during earthquakes. In the present study, an attempt was made to evaluate a type of steel shear wall that has regular holes in the inner sheet by modeling the finite element model with Abacus software. The shear wall of the steel plate, measuring 6000 × 3000 mm (one floor) and 3 mm thickness, was modeled with four different pores with a cross-sectional area. The shear wall was dynamically subjected to a time history of 5 seconds by three accelerators, El Centro, Imperial Valley and Kobe. The results showed that increasing the distance between the geometric center of the hole and the geometric center of the inner plate in the steel shear wall (increasing the RCS index) caused the total maximum acceleration to be transferred from the perimeter of the hole to horizontal and vertical beams. The results also show that there is no direct relationship between RCS index and total acceleration in steel shear wall and RCS index is separate from the peak ground acceleration value of earthquake.

Keywords: hollow steel plate shear wall, time history analysis, finite element method, abaqus software

Procedia PDF Downloads 93
15646 A Comparative Evaluation of Finite Difference Methods for the Extended Boussinesq Equations and Application to Tsunamis Modelling

Authors: Aurore Cauquis, Philippe Heinrich, Mario Ricchiuto, Audrey Gailler

Abstract:

In this talk, we look for an accurate time scheme to model the propagation of waves. Several numerical schemes have been developed to solve the extended weakly nonlinear weakly dispersive Boussinesq Equations. The temporal schemes used are two Lax-Wendroff schemes, second or third order accurate, two Runge-Kutta schemes of second and third order and a simplified third order accurate Lax-Wendroff scheme. Spatial derivatives are evaluated with fourth order accuracy. The numerical model is applied to two monodimensional benchmarks on a flat bottom. It is also applied to the simulation of the Algerian tsunami generated by a Mw=6 seism on the 18th March 2021. The tsunami propagation was highly dispersive and propagated across the Mediterranean Sea. We study here the effects of the order of temporal discretization on the accuracy of the results and on the time of computation.

Keywords: numerical analysis, tsunami propagation, water wave, boussinesq equations

Procedia PDF Downloads 221
15645 Multi-Criteria Evolutionary Algorithm to Develop Efficient Schedules for Complex Maintenance Problems

Authors: Sven Tackenberg, Sönke Duckwitz, Andreas Petz, Christopher M. Schlick

Abstract:

This paper introduces an extension to the well-established Resource-Constrained Project Scheduling Problem (RCPSP) to apply it to complex maintenance problems. The problem is to assign technicians to a team which has to process several tasks with multi-level skill requirements during a work shift. Here, several alternative activities for a task allow both, the temporal shift of activities or the reallocation of technicians and tools. As a result, switches from one valid work process variant to another can be considered and may be selected by the developed evolutionary algorithm based on the present skill level of technicians or the available tools. An additional complication of the observed scheduling problem is that the locations of the construction sites are only temporarily accessible during a day. Due to intensive rail traffic, the available time slots for maintenance and repair works are extremely short and are often distributed throughout the day. To identify efficient working periods, a first concept of a Bayesian network is introduced and is integrated into the extended RCPSP with pre-emptive and non-pre-emptive tasks. Thereby, the Bayesian network is used to calculate the probability of a maintenance task to be processed during a specific period of the shift. Focusing on the domain of maintenance of the railway infrastructure in metropolitan areas as the most unproductive implementation process at construction site, the paper illustrates how the extended RCPSP can be applied for maintenance planning support. A multi-criteria evolutionary algorithm with a problem representation is introduced which is capable of revising technician-task allocations, whereas the duration of the task may be stochastic. The approach uses a novel activity list representation to ensure easily describable and modifiable elements which can be converted into detailed shift schedules. Thereby, the main objective is to develop a shift plan which maximizes the utilization of each technician due to a minimization of the waiting times caused by rail traffic. The results of the already implemented core algorithm illustrate a fast convergence towards an optimal team composition for a shift, an efficient sequence of tasks and a high probability of the subsequent implementation due to the stochastic durations of the tasks. In the paper, the algorithm for the extended RCPSP is analyzed in experimental evaluation using real-world example problems with various size, resource complexity, tightness and so forth.

Keywords: maintenance management, scheduling, resource constrained project scheduling problem, genetic algorithms

Procedia PDF Downloads 217
15644 The Impact of Gender Differences on the Expressions of Refusal in Jordanian Arabic

Authors: Hanan Yousef, Nisreen Naji Al-Khawaldeh

Abstract:

The present study investigates the use of the expression of refusal by native speakers of Jordanian Arabic (NSsJA) in different social situations (i.e. invitations, suggestions, and offers). It also investigates the influence of gender on the refusal realization patterns within the Jordanian culture to provide a better insight into the relation between situations, strategies and gender in the Jordanian culture. To that end, a group of 70 participants, including 35 male and 35 female students from different departments at the Hashemite University (HU) participated in this study using mixed methods (i.e. Discourse Completion Test (DCT), interviews and naturally occurring data). Data were analyzed in light of a developed coding scheme. The results showed that NSsJA preferred indirect strategies which mitigate the interaction such as "excuse, reason and, explanation" strategy more than other strategies which aggravate the interaction such as "face-threatening" strategy. Moreover, the analysis of this study has revealed a considerable impact of gender on the use of linguistic forms expressing refusal among NSsJA. Significant differences in the results of the Chi-square test relating the effect of participants' gender indicate that both males and females were conscious of the gender of their interlocutors. The findings provide worthwhile insights into the relation amongst types of communicative acts and the rapport between people in social interaction. They assert that refusal should not be labeled as face threatening act since it does not always pose a threat in some cases especially where refusal is expressed among friends, relatives and family members. They highlight some distinctive culture-specific features of the communicative acts of refusal.

Keywords: gender, Jordanian Arabic, politeness, refusals, speech act

Procedia PDF Downloads 147
15643 Analysis of the Annual Proficiency Testing Procedure for Intermediate Reference Laboratories Conducted by the National Reference Laboratory from 2013 to 2017

Authors: Reena K., Mamatha H. G., Somshekarayya, P. Kumar

Abstract:

Objectives: The annual proficiency testing of intermediate reference laboratories is conducted by the National Reference Laboratory (NRL) to assess the efficiency of the laboratories to correctly identify Mycobacterium tuberculosis and to determine its drug susceptibility pattern. The proficiency testing results from 2013 to 2017 were analyzed to determine laboratories that were consistent in reporting quality results and those that had difficulty in doing so. Methods: A panel of twenty cultures were sent out to each of these laboratories. The laboratories were expected to grow the cultures in their own laboratories, set up drug susceptibly testing by all the methods they were certified for and report the results within the stipulated time period. The turnaround time for reporting results, specificity, sensitivity positive and negative predictive values and efficiency of the laboratory in identifying the cultures were analyzed. Results: Most of the laboratories had reported their results within the stipulated time period. However, there was enormous delay in reporting results from few of the laboratories. This was mainly due to improper functioning of the biosafety level III laboratory. Only 40% of the laboratories had 100% efficiency in solid culture using Lowenstein Jensen medium. This was expected as a solid culture, and drug susceptibility testing is not used for diagnosing drug resistance. Rapid molecular methods such as Line probe assay and Genexpert are used to determine drug resistance. Automated liquid culture system such as the Mycobacterial growth indicator tube is used to determine prognosis of the patient while on treatment. It was observed that 90% of the laboratories had achieved 100% in the liquid culture method. Almost all laboratories had achieved 100% efficiency in the line probe assay method which is the method of choice for determining drug-resistant tuberculosis. Conclusion: Since the liquid culture and line probe assay technologies are routinely used for the detection of drug-resistant tuberculosis the laboratories exhibited higher level of efficiency as compared to solid culture and drug susceptibility testing which are rarely used. The infrastructure of the laboratory should be maintained properly so that samples can be processed safely and results could be declared on time.

Keywords: annual proficiency testing, drug susceptibility testing, intermediate reference laboratory, national reference laboratory

Procedia PDF Downloads 171
15642 Fighting for Equality in Early Buddhism

Authors: Kenneth Lee

Abstract:

During Buddha’s time in the 5th century BCE, the Indian society was organized by a social stratification system called “the caste system” (Skt. varna), which still exists today. The origination of the caste system can be traced back to 1500 BCE within the ancient Vedic texts of the Aryans, the Indo-European nomadic people who migrated and settled in the Indus Valley region. However, the four-tiered hierarchical nature of the caste system created inequality, privilege, and discrimination based on hereditary transmission. After renouncing his royal status as a prince, Siddhartha Gautama spent six years in the forest, practiced austerities, mastered meditation, and eventually realized enlightenment. Thereupon, now referred to as “Shakyamuni Buddha” or “sage from the tribe of Shakya who has become awake,” the Buddha founded the Sangha, a community of monks, nuns, and lay followers, where everyone was equal and treated equally. After providing a brief overview of Buddha’s time, this talk will examine Buddha’s Dharma or teachings on equality and his creation of the Sangha as “society within a society, which had a dissolving effect on society.

Keywords: equality, women, buddhism, discrimination

Procedia PDF Downloads 94
15641 Planning for Sustainability in the Built Environment

Authors: Adedayo Jeremiah Adeyekun, Samuel Oluwagbemiga Ishola

Abstract:

This paper aimed to identify the significance of sustainability in the built environment, the economic and environmental importance to building and construction projects. Sustainability in the built environment has been a key objective of research over the past several decades. Sustainability in the built environment requires reconciliation between economic, environmental and social impacts of design and planning decisions made during the life cycle of a project from inception to termination. Planning for sustainability in the built environment needs us to go beyond our individual disciplines to consider the variety of economic, social and environmental impacts of our decisions in the long term. A decision to build a green residential development in an isolated location may pass some of the test of sustainability through its reduction in stormwater runoff, energy efficiency, and ecological sustainability in the building, but it may fail to be sustainable from a transportation perspective. Sustainability is important to the planning, design, construction, and preservation of the built environment; because it helps these activities reflect multiple values and considerations. In fact, the arts and sciences of the built environment have traditionally integrated values and fostered creative expression, capabilities that can and should lead the sustainability movement as society seeks ways to live in dynamic balance with its own diverse needs and the natural world. This research aimed to capture the state-of-the-art in the development of innovative sustainable design and planning strategies for building and construction projects. Therefore, there is a need for a holistic selection and implication approach for identifying potential sustainable strategies applicable to a particular project and evaluating the overall life cycle impact of each alternative by accounting for different applicable impacts and making the final selection among various viable alternatives.

Keywords: sustainability, built environment, planning, design, construction

Procedia PDF Downloads 157
15640 Enhanced Thermal and Electrical Properties of Terbium Manganate-Polyvinyl Alcohol Nanocomposite Film

Authors: Monalisa Halder, Amit K. Das, Ajit K. Meikap

Abstract:

Polymer nanocomposites are very significant materials both in academia and industry for diverse potential applicability in electronics. Polymer plays the role of matrix element which has low density, flexibility, good mechanical strength and electrical properties. Use of nanosized multiferroic filler in the polymer matrix is suitable to achieve nanocomposites with enhanced magneto-dielectric effect and good mechanical properties both at the same time. Multiferroic terbium manganate (TbMnO₃) nanoparticles have been synthesized by sol-gel method using chloride precursors. Terbium manganate-polyvinyl alcohol (TbMnO₃-PVA) nanocomposite film has been prepared by solution casting method. Crystallite size of TbMnO₃ nanoparticle has been calculated to be ~ 40 nm from XRD analysis. Morphological study of the samples has been done by scanning electron microscopy and a well dispersion of the nanoparticles in the PVA matrix has been found. Thermogravimetric analysis (TGA) exhibits enhancement of thermal stability of the nanocomposite film with the inclusion of TbMnO₃ nanofiller in PVA matrix. The electrical transport properties of the nanocomposite film sample have been studied in the frequency range 20Hz - 2MHz at and above room temperature. The frequency dependent variation of ac conductivity follows universal dielectric response (UDR) obeying Jhonscher’s sublinear power law. Correlated barrier hopping (CBH) mechanism is the dominant charge transport mechanism with maximum barrier height 19 meV above room temperature. The variation of dielectric constant of the sample with frequency has been studied at different temperatures. Real part of dielectric constant at 1 KHz frequency at room temperature of the sample is found to be ~ 8 which is higher than that of the pure PVA film sample (~ 6). Dielectric constant decreases with the increase in frequency. Relaxation peaks have been observed in the variation of imaginary part of electric modulus with frequency. The relaxation peaks shift towards higher frequency as temperature increases probably due to the existence of interfacial polarization in the sample in presence of applied electric field. The current-voltage (I-V) characteristics of the nanocomposite film have been studied under ±40 V applied at different temperatures. I-V characteristic exhibits temperature dependent rectifying nature indicating the formation of Schottky barrier diode (SBD) with barrier height 23 meV. In conclusion, using multiferroic TbMnO₃ nanofiller in PVA matrix, enhanced thermal stability and electrical properties can be achieved.

Keywords: correlated barrier hopping, nanocomposite, schottky diode, TbMnO₃, TGA

Procedia PDF Downloads 114
15639 Agent-Base Modeling of IoT Applications by Using Software Product Line

Authors: Asad Abbas, Muhammad Fezan Afzal, Muhammad Latif Anjum, Muhammad Azmat

Abstract:

The Internet of Things (IoT) is used to link up real objects that use the internet to interact. IoT applications allow handling and operating the equipment in accordance with environmental needs, such as transportation and healthcare. IoT devices are linked together via a number of agents that act as a middleman for communications. The operation of a heat sensor differs indoors and outside because agent applications work with environmental variables. In this article, we suggest using Software Product Line (SPL) to model IoT agents and applications' features on an XML-based basis. The contextual diversity within the same domain of application can be handled, and the reusability of features is increased by XML-based feature modelling. For the purpose of managing contextual variability, we have embraced XML for modelling IoT applications, agents, and internet-connected devices.

Keywords: IoT agents, IoT applications, software product line, feature model, XML

Procedia PDF Downloads 77
15638 Benefits of Using Social Media and Collaborative Online Platforms in PBL

Authors: Susanna Graziano, Lydia Krstic Ward

Abstract:

The purpose of this presentation is to demonstrate the steps of using multimedia and collaborative platforms in project-based learning. The presentation will demonstrate the stages of the learning project with various components of independent and collaborative learning, where students research the topic, share information, prepare a survey, use social media (Facebook, Instagram, WhasApp) and collaborative platforms (wikispaces.com and Google docs) to collect, analyze and process data, then produce reports and logos to be displayed as a final product. At the beginning of the presentation participants will answer a questionnaire about project based learning and share their experience on using social media, real–world project work and collaborative learning. Using a PPP, the presentation will walk participants through the steps of a completed project where tertiary education students are involved in putting together a multimedia campaign for safe driving in Kuwait. The research component of the project entails taking a holistic view on the problem of the high death rate in traffic accidents. The final goal of the project is to lead students to raise public awareness about the importance of safe driving. The project steps involve using the social media and collaborative platforms for collecting data and sharing the required materials to be used in the final product – a display of written reports, slogans and videos, as well as oral presentations. The same structure can be used to organize a multimedia campaign focusing on other issues, whilst scaffolding on students’ ability to brainstorm, retrieve information, organize it and engage in collaborative/ cooperative learning whilst being immersed in content-based learning as well as in authentic tasks. More specifically, the project we carried out at Box Hill College was a real-world one and involved a multimedia Campaign for Safe Driving since reckless driving is one of the major problems in the country. The idea for the whole project started by a presentation given by a board member of the Kuwaiti Society for Traffic Safety who was invited to college and spoke about: • Driving laws in the country, • What causes car accidents, • Driving safety tips. The principal goal of this project was to let students consider problems of traffic in Kuwait from different points of view. They also had to address the number and causes of accidents, evaluate the effectiveness of the local traffic law in order to send a warning about the importance of safe driving and, finally, suggest ways of its improvement. Benefits included: • Engagement, • Autonomy, • Motivation, • Content knowledge, • Language mastery, • Enhanced critical thinking, • Increased metacognitive awareness, • Improved social skills, • Authentic experience.

Keywords: social media, online learning platforms, collaborative platforms, project based learning

Procedia PDF Downloads 412
15637 Development and Optimization of Colon Targeted Drug Delivery System of Ayurvedic Churna Formulation Using Eudragit L100 and Ethyl Cellulose as Coating Material

Authors: Anil Bhandari, Imran Khan Pathan, Peeyush K. Sharma, Rakesh K. Patel, Suresh Purohit

Abstract:

The purpose of this study was to prepare time and pH dependent release tablets of Ayurvedic Churna formulation and evaluate their advantages as colon targeted drug delivery system. The Vidangadi Churna was selected for this study which contains Embelin and Gallic acid. Embelin is used in Helminthiasis as therapeutic agent. Embelin is insoluble in water and unstable in gastric environment so it was formulated in time and pH dependent tablets coated with combination of two polymers Eudragit L100 and ethyl cellulose. The 150mg of core tablet of dried extract and lactose were prepared by wet granulation method. The compression coating was used in the polymer concentration of 150mg for both the layer as upper and lower coating tablet was investigated. The results showed that no release was found in 0.1 N HCl and pH 6.8 phosphate buffers for initial 5 hours and about 98.97% of the drug was released in pH 7.4 phosphate buffer in total 17 hours. The in vitro release profiles of drug from the formulation could be best expressed first order kinetics as highest linearity (r2= 0.9943). The results of the present study have demonstrated that the time and pH dependent tablets system is a promising vehicle for preventing rapid hydrolysis in gastric environment and improving oral bioavailability of Embelin and Gallic acid for treatment of Helminthiasis.

Keywords: embelin, gallic acid, Vidangadi Churna, colon targeted drug delivery

Procedia PDF Downloads 343
15636 Video Based Ambient Smoke Detection By Detecting Directional Contrast Decrease

Authors: Omair Ghori, Anton Stadler, Stefan Wilk, Wolfgang Effelsberg

Abstract:

Fire-related incidents account for extensive loss of life and material damage. Quick and reliable detection of occurring fires has high real world implications. Whereas a major research focus lies on the detection of outdoor fires, indoor camera-based fire detection is still an open issue. Cameras in combination with computer vision helps to detect flames and smoke more quickly than conventional fire detectors. In this work, we present a computer vision-based smoke detection algorithm based on contrast changes and a multi-step classification. This work accelerates computer vision-based fire detection considerably in comparison with classical indoor-fire detection.

Keywords: contrast analysis, early fire detection, video smoke detection, video surveillance

Procedia PDF Downloads 424
15635 Examining the Relations among Autobiographical Memory Recall Types, Quality of Descriptions, and Emotional Arousal in Psychotherapy for Depression

Authors: Jinny Hong, Jeanne C. Watson

Abstract:

Three types of autobiographical memory recall -specific, episodic, and generic- were examined in relation to the quality of descriptions and in-session levels of emotional arousal. Correlational analyses and general estimating equation were conducted to test the relationships between 1) quality of descriptions and type of memory, 2) type of memory and emotional arousal, and 3) quality of descriptions and emotional arousal. The data was transcripts drawn from an archival randomized-control study comparing cognitive-behavioral therapy and emotion-focused therapy in a 16-week treatment for depression. Autobiographical memory recall segments were identified and sorted into three categories: specific, episodic, and generic. Quality of descriptions of these segments was then operationalized and measured using the Referential Activity Scale, and each memory segment was rated on four dimensions: concreteness, specificity, clarity, and overall imagery. Clients’ level of emotional arousal for each recall was measured using the Client’s Expression Emotion Scale. Contrary to the predictions, generic memories are associated with higher emotional arousal ratings and descriptive language ratings compared to specific memories. However, a positive relationship emerged between the quality of descriptions and expressed emotional arousal, indicating that the quality of descriptions in which memories are described in sessions is more important than the type of memory recalled in predicting clients’ level of emotional arousal. The results from this study provide a clearer understanding of the role of memory recall types and use of language in activating emotional arousal in psychotherapy sessions in a depressed sample.

Keywords: autobiographical memory recall, emotional arousal, psychotherapy for depression, quality of descriptions, referential activity

Procedia PDF Downloads 148
15634 The Impact of CO2 on Learning and Memory Duration of Bombus terrestris

Authors: Gholizadeh F. F., Goldansaz S. H., Bandani A. R., A. Ashouri

Abstract:

This study aimed to investigate the direct effects of increasing carbon dioxide (CO₂) concentration on the behavior of Bombus terrestris bumblebees in laboratory conditions to understand the outcomes of the augmentation of this gas in the Earth's atmosphere on the decline of populations of these pollinators. Learning and memory duration of bumblebees were evaluated as two main behavioral factors in social insects at different concentrations of CO₂. In both series of experiments, the behavior of bees under the influence of CO₂ changes compared to the control. Insects kept at high CO₂ concentrations learn less than control bees and spend more time identifying and navigating to discover their food source and access time (nectar consumption). These results showed that bees maybe lose some of their food resources due to poorer identification and act weaker on searching due to less memory and avoiding the enemy in higher CO₂ concentration. Therefore, CO₂ increasing concentration can be one of the reasons for the decline of these pollinating insects' populations by negatively affecting their fitness.

Keywords: Bombus terrestris, CO₂, learning, memory duration

Procedia PDF Downloads 163
15633 Neologisms and Word-Formation Processes in Board Game Rulebook Corpus: Preliminary Results

Authors: Athanasios Karasimos, Vasiliki Makri

Abstract:

This research focuses on the design and development of the first text Corpus based on Board Game Rulebooks (BGRC) with direct application on the morphological analysis of neologisms and tendencies in word-formation processes. Corpus linguistics is a dynamic field that examines language through the lens of vast collections of texts. These corpora consist of diverse written and spoken materials, ranging from literature and newspapers to transcripts of everyday conversations. By morphologically analyzing these extensive datasets, morphologists can gain valuable insights into how language functions and evolves, as these extensive datasets can reflect the byproducts of inflection, derivation, blending, clipping, compounding, and neology. This entails scrutinizing how words are created, modified, and combined to convey meaning in a corpus of challenging, creative, and straightforward texts that include rules, examples, tutorials, and tips. Board games teach players how to strategize, consider alternatives, and think flexibly, which are critical elements in language learning. Their rulebooks reflect not only their weight (complexity) but also the language properties of each genre and subgenre of these games. Board games are a captivating realm where strategy, competition, and creativity converge. Beyond the excitement of gameplay, board games also spark the art of word creation. Word games, like Scrabble, Codenames, Bananagrams, Wordcraft, Alice in the Wordland, Once uUpona Time, challenge players to construct words from a pool of letters, thus encouraging linguistic ingenuity and vocabulary expansion. These games foster a love for language, motivating players to unearth obscure words and devise clever combinations. On the other hand, the designers and creators produce rulebooks, where they include their joy of discovering the hidden potential of language, igniting the imagination, and playing with the beauty of words, making these games a delightful fusion of linguistic exploration and leisurely amusement. In this research, more than 150 rulebooks in English from all types of modern board games, either language-independent or language-dependent, are used to create the BGRC. A representative sample of each genre (family, party, worker placement, deckbuilding, dice, and chance games, strategy, eurogames, thematic, role-playing, among others) was selected based on the score from BoardGameGeek, the size of the texts and the level of complexity (weight) of the game. A morphological model with morphological networks, multi-word expressions, and word-creation mechanics based on the complexity of the textual structure, difficulty, and board game category will be presented. In enabling the identification of patterns, trends, and variations in word formation and other morphological processes, this research aspires to make avail of this creative yet strict text genre so as to (a) give invaluable insight into morphological creativity and innovation that (re)shape the lexicon of the English language and (b) test morphological theories. Overall, it is shown that corpus linguistics empowers us to explore the intricate tapestry of language, and morphology in particular, revealing its richness, flexibility, and adaptability in the ever-evolving landscape of human expression.

Keywords: board game rulebooks, corpus design, morphological innovations, neologisms, word-formation processes

Procedia PDF Downloads 73
15632 Concussion: Clinical and Vocational Outcomes from Sport Related Mild Traumatic Brain Injury

Authors: Jack Nash, Chris Simpson, Holly Hurn, Ronel Terblanche, Alan Mistlin

Abstract:

There is an increasing incidence of mild traumatic brain injury (mTBI) cases throughout sport and with this, a growing interest from governing bodies to ensure these are managed appropriately and player welfare is prioritised. The Berlin consensus statement on concussion in sport recommends a multidisciplinary approach when managing those patients who do not have full resolution of mTBI symptoms. There are as of yet no standardised guideline to follow in the treatment of complex cases mTBI in athletes. The aim of this project was to analyse the outcomes, both clinical and vocational, of all patients admitted to the mild Traumatic Brain Injury (mTBI) service at the UK’s Defence Military Rehabilitation Centre Headley Court between 1st June 2008 and 1st February 2017, as a result of a sport induced injury, and evaluate potential predictive indicators of outcome. Patients were identified from a database maintained by the mTBI service. Clinical and occupational outcomes were ascertained from medical and occupational employment records, recorded prospectively, at time of discharge from the mTBI service. Outcomes were graded based on the vocational independence scale (VIS) and clinical documentation at discharge. Predictive indicators including referral time, age at time of injury, previous mental health diagnosis and a financial claim in place at time of entry to service were assessed using logistic regression. 45 Patients were treated for sport-related mTBI during this time frame. Clinically 96% of patients had full resolution of their mTBI symptoms after input from the mTBI service. 51% of patients returned to work at their previous vocational level, 4% had ongoing mTBI symptoms, 22% had ongoing physical rehabilitation needs, 11% required mental health input and 11% required further vestibular rehabilitation. Neither age, time to referral, pre-existing mental health condition nor compensation seeking had a significant impact on either vocational or clinical outcome in this population. The vast majority of patients reviewed in the mTBI clinic had persistent symptoms which could not be managed in primary care. A consultant-led, multidisciplinary approach to the diagnosis and management of mTBI has resulted in excellent clinical outcomes in these complex cases. High levels of symptom resolution suggest that this referral and treatment pathway is successful and is a model which could be replicated in other organisations with consultant led input. Further understanding of both predictive and individual factors would allow clinicians to focus treatments on those who are most likely to develop long-term complications following mTBI. A consultant-led, multidisciplinary service ensures a large number of patients will have complete resolution of mTBI symptoms after sport-related mTBI. Further research is now required to ascertain the key predictive indicators of outcome following sport-related mTBI.

Keywords: brain injury, concussion, neurology, rehabilitation, sports injury

Procedia PDF Downloads 139
15631 Immobilization of Lipase Enzyme by Low Cost Material: A Statistical Approach

Authors: Md. Z. Alam, Devi R. Asih, Md. N. Salleh

Abstract:

Immobilization of lipase enzyme produced from palm oil mill effluent (POME) by the activated carbon (AC) among the low cost support materials was optimized. The results indicated that immobilization of 94% was achieved by AC as the most suitable support material. A sequential optimization strategy based on a statistical experimental design, including one-factor-at-a-time (OFAT) method was used to determine the equilibrium time. Three components influencing lipase immobilization were optimized by the response surface methodology (RSM) based on the face-centered central composite design (FCCCD). On the statistical analysis of the results, the optimum enzyme concentration loading, agitation rate and carbon active dosage were found to be 30 U/ml, 300 rpm and 8 g/L respectively, with a maximum immobilization activity of 3732.9 U/g-AC after 2 hrs of immobilization. Analysis of variance (ANOVA) showed a high regression coefficient (R2) of 0.999, which indicated a satisfactory fit of the model with the experimental data. The parameters were statistically significant at p<0.05.

Keywords: activated carbon, POME based lipase, immobilization, adsorption

Procedia PDF Downloads 229
15630 The Relationships between Energy Consumption, Carbon Dioxide (CO2) Emissions, and GDP for Turkey: Time Series Analysis, 1980-2010

Authors: Jinhoa Lee

Abstract:

The relationships between environmental quality, energy use and economic output have created growing attention over the past decades among researchers and policy makers. Focusing on the empirical aspects of the role of carbon dioxide (CO2) emissions and energy use in affecting the economic output, this paper is an effort to fulfill the gap in a comprehensive case study at a country level using modern econometric techniques. To achieve the goal, this country-specific study examines the short-run and long-run relationships among energy consumption (using disaggregated energy sources: crude oil, coal, natural gas, and electricity), CO2 emissions and gross domestic product (GDP) for Turkey using time series analysis from the year 1980-2010. To investigate the relationships between the variables, this paper employs the Augmented Dickey-Fuller (ADF) test for stationarity, Johansen’s maximum likelihood method for cointegration and a Vector Error Correction Model (VECM) for both short- and long-run causality among the research variables for the sample. The long-run equilibrium in the VECM suggests no effects of the CO2 emissions and energy use on the GDP in Turkey. There exists a short-run bidirectional relationship between the electricity and natural gas consumption, and also there is a negative unidirectional causality running from the GDP to electricity use. Overall, the results partly support arguments that there are relationships between energy use and economic output; however, the effects may differ due to the source of energy such as in the case of Turkey for the period of 1980-2010. However, there is no significant relationship between the CO2 emissions and the GDP and between the CO2 emissions and the energy use both in the short term and long term.

Keywords: CO2 emissions, energy consumption, GDP, Turkey, time series analysis

Procedia PDF Downloads 496
15629 Through the Robot’s Eyes: A Comparison of Robot-Piloted, Virtual Reality, and Computer Based Exposure for Fear of Injections

Authors: Bonnie Clough, Tamara Ownsworth, Vladimir Estivill-Castro, Matt Stainer, Rene Hexel, Andrew Bulmer, Wendy Moyle, Allison Waters, David Neumann, Jayke Bennett

Abstract:

The success of global vaccination programs is reliant on the uptake of vaccines to achieve herd immunity. Yet, many individuals do not obtain vaccines or venipuncture procedures when needed. Whilst health education may be effective for those individuals who are hesitant due to safety or efficacy concerns, for many of these individuals, the primary concern relates to blood or injection fear or phobia (BII). BII is highly prevalent and associated with a range of negative health impacts, both at individual and population levels. Exposure therapy is an efficacious treatment for specific phobias, including BII, but has high patient dropout and low implementation by therapists. Whilst virtual reality approaches exposure therapy may be more acceptable, they have similarly low rates of implementation by therapists and are often difficult to tailor to an individual client’s needs. It was proposed that a piloted robot may be able to adequately facilitate fear induction and be an acceptable approach to exposure therapy. The current study examined fear induction responses, acceptability, and feasibility of a piloted robot for BII exposure. A Nao humanoid robot was programmed to connect with a virtual reality head-mounted display, enabling live streaming and exploration of real environments from a distance. Thirty adult participants with BII fear were randomly assigned to robot-pilot or virtual reality exposure conditions in a laboratory-based fear exposure task. All participants also completed a computer-based two-dimensional exposure task, with an order of conditions counterbalanced across participants. Measures included fear (heart rate variability, galvanic skin response, stress indices, and subjective units of distress), engagement with a feared stimulus (eye gaze: time to first fixation and a total number of fixations), acceptability, and perceived treatment credibility. Preliminary results indicate that fear responses can be adequately induced via a robot-piloted platform. Further results will be discussed, as will implications for the treatment of BII phobia and other fears. It is anticipated that piloted robots may provide a useful platform for facilitating exposure therapy, being more acceptable than in-vivo exposure and more flexible than virtual reality exposure.

Keywords: anxiety, digital mental health, exposure therapy, phobia, robot, virtual reality

Procedia PDF Downloads 64
15628 The Role of Urban Development Patterns for Mitigating Extreme Urban Heat: The Case Study of Doha, Qatar

Authors: Yasuyo Makido, Vivek Shandas, David J. Sailor, M. Salim Ferwati

Abstract:

Mitigating extreme urban heat is challenging in a desert climate such as Doha, Qatar, since outdoor daytime temperature area often too high for the human body to tolerate. Recent studies demonstrate that cities in arid and semiarid areas can exhibit ‘urban cool islands’ - urban areas that are cooler than the surrounding desert. However, the variation of temperatures as a result of the time of day and factors leading to temperature change remain at the question. To address these questions, we examined the spatial and temporal variation of air temperature in Doha, Qatar by conducting multiple vehicle-base local temperature observations. We also employed three statistical approaches to model surface temperatures using relevant predictors: (1) Ordinary Least Squares, (2) Regression Tree Analysis and (3) Random Forest for three time periods. Although the most important determinant factors varied by day and time, distance to the coast was the significant determinant at midday. A 70%/30% holdout method was used to create a testing dataset to validate the results through Pearson’s correlation coefficient. The Pearson’s analysis suggests that the Random Forest model more accurately predicts the surface temperatures than the other methods. We conclude with recommendations about the types of development patterns that show the greatest potential for reducing extreme heat in air climates.

Keywords: desert cities, tree-structure regression model, urban cool Island, vehicle temperature traverse

Procedia PDF Downloads 379
15627 The Utilization of Particle Swarm Optimization Method to Solve Nurse Scheduling Problem

Authors: Norhayati Mohd Rasip, Abd. Samad Hasan Basari , Nuzulha Khilwani Ibrahim, Burairah Hussin

Abstract:

The allocation of working schedule especially for shift environment is hard to fulfill its fairness among them. In the case of nurse scheduling, to set up the working time table for them is time consuming and complicated, which consider many factors including rules, regulation and human factor. The scenario is more complicated since most nurses are women which have personnel constraints and maternity leave factors. The undesirable schedule can affect the nurse productivity, social life and the absenteeism can significantly as well affect patient's life. This paper aimed to enhance the scheduling process by utilizing the particle swarm optimization in order to solve nurse scheduling problem. The result shows that the generated multiple initial schedule is fulfilled the requirements and produces the lowest cost of constraint violation.

Keywords: nurse scheduling, particle swarm optimisation, nurse rostering, hard and soft constraint

Procedia PDF Downloads 352
15626 Improving University Operations with Data Mining: Predicting Student Performance

Authors: Mladen Dragičević, Mirjana Pejić Bach, Vanja Šimičević

Abstract:

The purpose of this paper is to develop models that would enable predicting student success. These models could improve allocation of students among colleges and optimize the newly introduced model of government subsidies for higher education. For the purpose of collecting data, an anonymous survey was carried out in the last year of undergraduate degree student population using random sampling method. Decision trees were created of which two have been chosen that were most successful in predicting student success based on two criteria: Grade Point Average (GPA) and time that a student needs to finish the undergraduate program (time-to-degree). Decision trees have been shown as a good method of classification student success and they could be even more improved by increasing survey sample and developing specialized decision trees for each type of college. These types of methods have a big potential for use in decision support systems.

Keywords: data mining, knowledge discovery in databases, prediction models, student success

Procedia PDF Downloads 396
15625 Power Flow and Modal Analysis of a Power System Including Unified Power Flow Controller

Authors: Djilani Kobibi Youcef Islam, Hadjeri Samir, Djehaf Mohamed Abdeldjalil

Abstract:

The Flexible AC Transmission System (FACTS) technology is a new advanced solution that increases the reliability and provides more flexibility, controllability, and stability of a power system. The Unified Power Flow Controller (UPFC), as the most versatile FACTS device for regulating power flow, is able to control respectively transmission line real power, reactive power, and node voltage. The main purpose of this paper is to analyze the effect of the UPFC on the load flow, the power losses, and the voltage stability using NEPLAN software modules, Newton-Raphson load flow is used for the power flow analysis and the modal analysis is used for the study of the voltage stability. The simulation was carried out on the IEEE 14-bus test system.

Keywords: FACTS, load flow, modal analysis, UPFC, voltage stability

Procedia PDF Downloads 498
15624 Role of ABC Transporters in Non-Target Site Herbicide Resistance in Black Grass (Alopecurus myosuroides)

Authors: Alina Goldberg Cavalleri, Sara Franco Ortega, Nawaporn Onkokesung, Richard Dale, Melissa Brazier-Hicks, Robert Edwards

Abstract:

Non-target site based resistance (NTSR) to herbicides in weeds is a polygenic trait associated with the upregulation of proteins involved in xenobiotic detoxification and translocation we have termed the xenome. Among the xenome proteins, ABC transporters play a key role in enhancing herbicide metabolism by effluxing conjugated xenobiotics from the cytoplasm into the vacuole. The importance of ABC transporters is emphasized by the fact that they often contribute to multidrug resistance in human cells and antibiotic resistance in bacteria. They also play a key role in insecticide resistance in major vectors of human diseases and crop pests. By surveying available databases, transcripts encoding ABCs have been identified as being enhanced in populations exhibiting NTSR in several weed species. Based on a transcriptomics data in black grass (Alopecurus myosuroides, Am), we have identified three proteins from the ABC-C subfamily that are upregulated in NTSR populations. ABC-C transporters are poorly characterized proteins in plants, but in Arabidopsis localize to the vacuolar membrane and have functional roles in transporting glutathionylated (GSH)-xenobiotic conjugates. We found that the up-regulation of AmABCs strongly correlates with the up-regulation of a glutathione transferase termed AmGSTU2, which can conjugate GSH to herbicides. The expression profile of the ABC transcripts was profiled in populations of black grass showing different degree of resistance to herbicides. This, together with a phylogenetic analysis, revealed that AmABCs cluster in different groups which might indicate different substrate and roles in the herbicide resistance phenotype in the different populations

Keywords: black grass, herbicide, resistance, transporters

Procedia PDF Downloads 135
15623 Seismic Retrofits – A Catalyst for Minimizing the Building Sector’s Carbon Footprint

Authors: Juliane Spaak

Abstract:

A life-cycle assessment was performed, looking at seven retrofit projects in New Zealand using LCAQuickV3.5. The study found that retrofits save up to 80% of embodied carbon emissions for the structural elements compared to a new building. In other words, it is only a 20% carbon investment to transform and extend a building’s life. In addition, the systems were evaluated by looking at environmental impacts over the design life of these buildings and resilience using FEMA P58 and PACT software. With the increasing interest in Zero Carbon targets, significant changes in the building and construction sector are required. Emissions for buildings arise from both embodied carbon and operations. Based on the significant advancements in building energy technology, the focus is moving more toward embodied carbon, a large portion of which is associated with the structure. Since older buildings make up most of the real estate stock of our cities around the world, their reuse through structural retrofit and wider refurbishment plays an important role in extending the life of a building’s embodied carbon. New Zealand’s building owners and engineers have learned a lot about seismic issues following a decade of significant earthquakes. Recent earthquakes have brought to light the necessity to move away from constructing code-minimum structures that are designed for life safety but are frequently ‘disposable’ after a moderate earthquake event, especially in relation to a structure’s ability to minimize damage. This means weaker buildings sit as ‘carbon liabilities’, with considerably more carbon likely to be expended remediating damage after a shake. Renovating and retrofitting older assets plays a big part in reducing the carbon profile of the buildings sector, as breathing new life into a building’s structure is vastly more sustainable than the highest quality ‘green’ new builds, which are inherently more carbon-intensive. The demolition of viable older buildings (often including heritage buildings) is increasingly at odds with society’s desire for a lower carbon economy. Bringing seismic resilience and carbon best practice together in decision-making can open the door to commercially attractive outcomes, with retrofits that include structural and sustainability upgrades transforming the asset’s revenue generation. Across the global real estate market, tenants are increasingly demanding the buildings they occupy be resilient and aligned with their own climate targets. The relationship between seismic performance and ‘sustainable design’ has yet to fully mature, yet in a wider context is of profound consequence. A whole-of-life carbon perspective on a building means designing for the likely natural hazards within the asset’s expected lifespan, be that earthquake, storms, damage, bushfires, fires, and so on, ¬with financial mitigation (e.g., insurance) part, but not all, of the picture.

Keywords: retrofit, sustainability, earthquake, reuse, carbon, resilient

Procedia PDF Downloads 58
15622 Investigation of the Drying Times of Blood under Different Environmental Conditions and on Different Fabrics and the Transfer of Blood at Different Times of the Drying Process

Authors: Peter Parkinson

Abstract:

The research investigates the effects of temperature, humidity, wind speed, and fabric composition on the drying times of blood and assesses the degree of blood transfer that can occur during the drying process. An assortment of fabrics, of different composition and thicknesses, were collected and stained using two blood volumes and exposed to varying environmental conditions. The conclusion reached was that temperature, humidity, wind speed, and fabric thickness do have an effect on drying times. An increase in temperature and wind speed results in a decrease in drying times while an increase in fabric thickness and humidity extended the drying times of blood under similar conditions. Transfer experimentation utilized three donor fabrics, 100% white cotton, 100% acrylic, and 100% cotton denim, which were bloodstained using two blood volumes. The fabrics were subjected to both full and low/light force contact from the donor fabrics onto the recipient fabric, under different environmental conditions. Transfer times onto the 100% white cotton (recipient fabric) from all donor fabrics were shorter than the drying times observed. The intensities of the bloodstains decreased from high to low with time during the drying process. The degree of transfer at high, medium, and low intensities varied significantly between different materials and is dependent on the environmental conditions, fabric compositions, blood volumes, the type of contact (full or light force), and the drying times observed for the respective donor fabrics. These factors should be considered collectively and conservatively when assessing the time frame of secondary transfer in casework.

Keywords: blood, drying time, blood stain transfer, different environmental conditions, fabrics

Procedia PDF Downloads 134
15621 Effect of Different Ground Motion Scaling Methods on Behavior of 40 Story RC Core Wall Building

Authors: Muhammad Usman, Munir Ahmed

Abstract:

The demand of high-rise buildings has grown fast during the past decades. The design of these buildings by using RC core wall have been widespread nowadays in many countries. The RC core wall (RCCW) buildings encompasses central core wall and boundary columns joined through post tension slab at different floor levels. The core wall often provides greater stiffness as compared to the collective stiffness of the boundary columns. Hence, the core wall dominantly resists lateral loading i.e. wind or earthquake load. Non-linear response history analysis (NLRHA) procedure is the finest seismic design procedure of the times for designing high-rise buildings. The modern design tools for nonlinear response history analysis and performance based design has provided more confidence to design these structures for high-rise buildings. NLRHA requires selection and scaling of ground motions to match design spectrum for site specific conditions. Designers use several techniques for scaling ground motion records (time series). Time domain and frequency domain scaling are most commonly used which comprises their own benefits and drawbacks. Due to lengthy process of NLRHA, application of only one technique is conceivable. To the best of author’s knowledge, no consensus on the best procedures for the selection and scaling of the ground motions is available in literature. This research aims to provide the finest ground motion scaling technique specifically for designing 40 story high-rise RCCW buildings. Seismic response of 40 story RCCW building is checked by applying both the frequency domain and time domain scaling. Variable sites are selected in three critical seismic zones of Pakistan. The results indicates that there is extensive variation in seismic response of building for these scaling. There is still a need to build a consensus on the subjected research by investigating variable sites and buildings heights.

Keywords: 40-storied RC core wall building, nonlinear response history analysis, ground motions, time domain scaling, frequency domain scaling

Procedia PDF Downloads 122
15620 A Study of the Performance Parameter for Recommendation Algorithm Evaluation

Authors: C. Rana, S. K. Jain

Abstract:

The enormous amount of Web data has challenged its usage in efficient manner in the past few years. As such, a range of techniques are applied to tackle this problem; prominent among them is personalization and recommender system. In fact, these are the tools that assist user in finding relevant information of web. Most of the e-commerce websites are applying such tools in one way or the other. In the past decade, a large number of recommendation algorithms have been proposed to tackle such problems. However, there have not been much research in the evaluation criteria for these algorithms. As such, the traditional accuracy and classification metrics are still used for the evaluation purpose that provides a static view. This paper studies how the evolution of user preference over a period of time can be mapped in a recommender system using a new evaluation methodology that explicitly using time dimension. We have also presented different types of experimental set up that are generally used for recommender system evaluation. Furthermore, an overview of major accuracy metrics and metrics that go beyond the scope of accuracy as researched in the past few years is also discussed in detail.

Keywords: collaborative filtering, data mining, evolutionary, clustering, algorithm, recommender systems

Procedia PDF Downloads 398