Search results for: multimodal bio metrics
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 783

Search results for: multimodal bio metrics

423 Electrical Decomposition of Time Series of Power Consumption

Authors: Noura Al Akkari, Aurélie Foucquier, Sylvain Lespinats

Abstract:

Load monitoring is a management process for energy consumption towards energy savings and energy efficiency. Non Intrusive Load Monitoring (NILM) is one method of load monitoring used for disaggregation purposes. NILM is a technique for identifying individual appliances based on the analysis of the whole residence data retrieved from the main power meter of the house. Our NILM framework starts with data acquisition, followed by data preprocessing, then event detection, feature extraction, then general appliance modeling and identification at the final stage. The event detection stage is a core component of NILM process since event detection techniques lead to the extraction of appliance features. Appliance features are required for the accurate identification of the household devices. In this research work, we aim at developing a new event detection methodology with accurate load disaggregation to extract appliance features. Time-domain features extracted are used for tuning general appliance models for appliance identification and classification steps. We use unsupervised algorithms such as Dynamic Time Warping (DTW). The proposed method relies on detecting areas of operation of each residential appliance based on the power demand. Then, detecting the time at which each selected appliance changes its states. In order to fit with practical existing smart meters capabilities, we work on low sampling data with a frequency of (1/60) Hz. The data is simulated on Load Profile Generator software (LPG), which was not previously taken into consideration for NILM purposes in the literature. LPG is a numerical software that uses behaviour simulation of people inside the house to generate residential energy consumption data. The proposed event detection method targets low consumption loads that are difficult to detect. Also, it facilitates the extraction of specific features used for general appliance modeling. In addition to this, the identification process includes unsupervised techniques such as DTW. To our best knowledge, there exist few unsupervised techniques employed with low sampling data in comparison to the many supervised techniques used for such cases. We extract a power interval at which falls the operation of the selected appliance along with a time vector for the values delimiting the state transitions of the appliance. After this, appliance signatures are formed from extracted power, geometrical and statistical features. Afterwards, those formed signatures are used to tune general model types for appliances identification using unsupervised algorithms. This method is evaluated using both simulated data on LPG and real-time Reference Energy Disaggregation Dataset (REDD). For that, we compute performance metrics using confusion matrix based metrics, considering accuracy, precision, recall and error-rate. The performance analysis of our methodology is then compared with other detection techniques previously used in the literature review, such as detection techniques based on statistical variations and abrupt changes (Variance Sliding Window and Cumulative Sum).

Keywords: electrical disaggregation, DTW, general appliance modeling, event detection

Procedia PDF Downloads 52
422 Power Quality Issues: Power Supply Interruptions as Key Constraint to Development in Ekiti State, Nigeria

Authors: Oluwatosin S. Adeoye

Abstract:

The power quality issues in the world today are critical to the development of different nations. Prosperity of each nation depends on availability of constant power supply. Constant power supply is a major challenge in Africa particularly in Nigeria where the generated power is than thirty percent of the required power. The metrics of power quality are voltage dip, flickers, spikes, harmonics and interruptions. The level of interruptions in Ekiti State was examined through the investigation of the causes of power interruptions in the State. The method used was the collection of data from the Distribution Company, assessment through simple programming as a command for plotting the graphs through the use of MATLAB 2015 depicting the behavioural pattern of the interruption for a period of six months in 2016. The result shows that the interrelationship between the interruptions and development. Recommendations were suggested with the objective of solving the problems being set up by interruptions in the State and these include installation of reactors, automatic voltage regulators and effective tap changing system on the lines, busses and transformer substation respectively.

Keywords: development, frequency, interruption, power, quality

Procedia PDF Downloads 141
421 Measurement Tools of the Maturity Model for IT Service Outsourcing in Higher Education Institutions

Authors: Victoriano Valencia García, Luis Usero Aragonés, Eugenio J. Fernández Vicente

Abstract:

Nowadays, the successful implementation of ICTs is vital for almost any kind of organization. Good governance and ICT management are essential for delivering value, managing technological risks, managing resources and performance measurement. In addition, outsourcing is a strategic IT service solution which complements IT services provided internally in organizations. This paper proposes the measurement tools of a new holistic maturity model based on standards ISO/IEC 20000 and ISO/IEC 38500, and the frameworks and best practices of ITIL and COBIT, with a specific focus on IT outsourcing. These measurement tools allow independent validation and practical application in the field of higher education, using a questionnaire, metrics tables, and continuous improvement plan tables as part of the measurement process. Guidelines and standards are proposed in the model for facilitating adaptation to universities and achieving excellence in the outsourcing of IT services.

Keywords: IT governance, IT management, IT services, outsourcing, maturity model, measurement tools

Procedia PDF Downloads 567
420 A Hybrid Normalized Gradient Correlation Based Thermal Image Registration for Morphoea

Authors: L. I. Izhar, T. Stathaki, K. Howell

Abstract:

Analyzing and interpreting of thermograms have been increasingly employed in the diagnosis and monitoring of diseases thanks to its non-invasive, non-harmful nature and low cost. In this paper, a novel system is proposed to improve diagnosis and monitoring of morphoea skin disorder based on integration with the published lines of Blaschko. In the proposed system, image registration based on global and local registration methods are found inevitable. This paper presents a modified normalized gradient cross-correlation (NGC) method to reduce large geometrical differences between two multimodal images that are represented by smooth gray edge maps is proposed for the global registration approach. This method is improved further by incorporating an iterative-based normalized cross-correlation coefficient (NCC) method. It is found that by replacing the final registration part of the NGC method where translational differences are solved in the spatial Fourier domain with the NCC method performed in the spatial domain, the performance and robustness of the NGC method can be greatly improved. It is shown in this paper that the hybrid NGC method not only outperforms phase correlation (PC) method but also improved misregistration due to translation, suffered by the modified NGC method alone for thermograms with ill-defined jawline. This also demonstrates that by using the gradients of the gray edge maps and a hybrid technique, the performance of the PC based image registration method can be greatly improved.

Keywords: Blaschko’s lines, image registration, morphoea, thermal imaging

Procedia PDF Downloads 287
419 Making Creative Ethnography through Droned Mode of Engagements

Authors: Elin Linder

Abstract:

Ethnographic endeavors feature a long history of creative modes of engagements, and anthropology an equally long critique of its disciplinary attention to worded representations of beyond worded experiences. Curious and critical as our research comes about, takes place, unfolds, and develops, processes of documenting, exploring, experiencing, and producing knowledge commonly evolve as intrinsic parts of our situated wishes to make sense of the worlds we study. We may imagine to do one thing and to use a specific mode of fieldnoting, only to end up doing something else, such as to capture dynamics and dimensions otherwise not attentively engaged or even lost. This paper builds on such an experience, and it acts window to open the conversation for doing and representing ethnographic work as creatively as it was undertaken. Expressively and actively undertaken by means of sensuous scholarship, fieldworking in the world of olivicoltura in Apulia intriguingly advanced into resourcefully embodied research using a drone. While the drone first and foremost allowed perspectives that one as a human is largely and physically incapable of exploring, it rapidly emerged into a mode of engagement that probed critical question how one comes to learn how to see that which one watches, listen to that which one hears, smell that which one scents, feel that which one touch, and gather that which one experience. This paper develops how the drone incorporated a transition of a particularly situated ethnographic sense of attention, all while visualizing how imaginative conceptualizations enable unexpected modes of multimodal knowing in much multisensorial worlds of being.

Keywords: drone, multimodality, sensuous scholarship, critical creativity, ethnographic practice

Procedia PDF Downloads 53
418 Empirical Decomposition of Time Series of Power Consumption

Authors: Noura Al Akkari, Aurélie Foucquier, Sylvain Lespinats

Abstract:

Load monitoring is a management process for energy consumption towards energy savings and energy efficiency. Non Intrusive Load Monitoring (NILM) is one method of load monitoring used for disaggregation purposes. NILM is a technique for identifying individual appliances based on the analysis of the whole residence data retrieved from the main power meter of the house. Our NILM framework starts with data acquisition, followed by data preprocessing, then event detection, feature extraction, then general appliance modeling and identification at the final stage. The event detection stage is a core component of NILM process since event detection techniques lead to the extraction of appliance features. Appliance features are required for the accurate identification of the household devices. In this research work, we aim at developing a new event detection methodology with accurate load disaggregation to extract appliance features. Time-domain features extracted are used for tuning general appliance models for appliance identification and classification steps. We use unsupervised algorithms such as Dynamic Time Warping (DTW). The proposed method relies on detecting areas of operation of each residential appliance based on the power demand. Then, detecting the time at which each selected appliance changes its states. In order to fit with practical existing smart meters capabilities, we work on low sampling data with a frequency of (1/60) Hz. The data is simulated on Load Profile Generator software (LPG), which was not previously taken into consideration for NILM purposes in the literature. LPG is a numerical software that uses behaviour simulation of people inside the house to generate residential energy consumption data. The proposed event detection method targets low consumption loads that are difficult to detect. Also, it facilitates the extraction of specific features used for general appliance modeling. In addition to this, the identification process includes unsupervised techniques such as DTW. To our best knowledge, there exist few unsupervised techniques employed with low sampling data in comparison to the many supervised techniques used for such cases. We extract a power interval at which falls the operation of the selected appliance along with a time vector for the values delimiting the state transitions of the appliance. After this, appliance signatures are formed from extracted power, geometrical and statistical features. Afterwards, those formed signatures are used to tune general model types for appliances identification using unsupervised algorithms. This method is evaluated using both simulated data on LPG and real-time Reference Energy Disaggregation Dataset (REDD). For that, we compute performance metrics using confusion matrix based metrics, considering accuracy, precision, recall and error-rate. The performance analysis of our methodology is then compared with other detection techniques previously used in the literature review, such as detection techniques based on statistical variations and abrupt changes (Variance Sliding Window and Cumulative Sum).

Keywords: general appliance model, non intrusive load monitoring, events detection, unsupervised techniques;

Procedia PDF Downloads 56
417 The Importance of Applying Established Web Site Design Principles on an Online Performance Management System

Authors: R. W. Brown, P. J. Blignaut

Abstract:

An online performance management system was evaluated, and recommendations were made to improve the system. The study shows the effects of not adhering to the established web design principles and conventions. Furthermore, the study indicates that if the online performance management system is not well designed, it may have negative effects on the overall usability of the system and these negative effects will have consequences for both the employer and employees. The evaluation was done in terms of the usability metrics of effectiveness, efficiency and user satisfaction. Effectiveness was measured in terms of the success rate with which users could execute prescribed tasks in a sandbox system. Efficiency was expressed in terms of the time it took participants to understand what is expected of them and to execute the tasks. Post-test questionnaires were used in order to determine the satisfaction of the participants. Recommendations were made to improve the usability of the online performance management system.

Keywords: eye tracking, human resource management, performance management, usability

Procedia PDF Downloads 185
416 Design and Preliminary Evaluation of Benzoxazolone-Based Agents for Targeting Mitochondrial-Located Translocator Protein

Authors: Nidhi Chadha, A. K. Tiwari, Marilyn D. Milton, Anil K. Mishra

Abstract:

Translocator protein (18 kDa) TSPO is highly expressed during microglia activation in neuroinflammation. Although a number of PET ligands have been developed for the visualization of activated microglia, one of the advantageous approaches is to develop potential optical imaging (OI) probe. Our study involves computational screening, synthesis and evaluation of TSPO ligand through various imaging modalities namely PET/SPECT/Optical. The initial computational screening involves pharmacophore modeling from the library designing having oxo-benzooxazol-3-yl-N-phenyl-acetamide groups and synthesis for visualization of efficacy of these compounds as multimodal imaging probes. Structure modeling of monomer, Ala147Thr mutated, parallel and anti-parallel TSPO dimers was performed and docking analysis was performed for distinct binding sites. Computational analysis showed pattern of variable binding profile of known diagnostic ligands and NBMP via interactions with conserved residues along with TSPO’s natural polymorphism of Ala147→Thr, which showed alteration in the binding affinity due to considerable changes in tertiary structure. Preliminary in vitro binding studies shows binding affinity in the range of 1-5 nm and selectivity was also certified by blocking studies. In summary, this skeleton was found to be potential probe for TSPO imaging due to ease in synthesis, appropriate lipophilicity and reach to specific region of brain.

Keywords: TSPO, molecular modeling, imaging, docking

Procedia PDF Downloads 436
415 Methodology to Achieve Non-Cooperative Target Identification Using High Resolution Range Profiles

Authors: Olga Hernán-Vega, Patricia López-Rodríguez, David Escot-Bocanegra, Raúl Fernández-Recio, Ignacio Bravo

Abstract:

Non-Cooperative Target Identification has become a key research domain in the Defense industry since it provides the ability to recognize targets at long distance and under any weather condition. High Resolution Range Profiles, one-dimensional radar images where the reflectivity of a target is projected onto the radar line of sight, are widely used for identification of flying targets. According to that, to face this problem, an approach to Non-Cooperative Target Identification based on the exploitation of Singular Value Decomposition to a matrix of range profiles is presented. Target Identification based on one-dimensional radar images compares a collection of profiles of a given target, namely test set, with the profiles included in a pre-loaded database, namely training set. The classification is improved by using Singular Value Decomposition since it allows to model each aircraft as a subspace and to accomplish recognition in a transformed domain where the main features are easier to extract hence, reducing unwanted information such as noise. Singular Value Decomposition permits to define a signal subspace which contain the highest percentage of the energy, and a noise subspace which will be discarded. This way, only the valuable information of each target is used in the recognition process. The identification algorithm is based on finding the target that minimizes the angle between subspaces and takes place in a transformed domain. Two metrics, F1 and F2, based on Singular Value Decomposition are accomplished in the identification process. In the case of F2, the angle is weighted, since the top vectors set the importance in the contribution to the formation of a target signal, on the contrary F1 simply shows the evolution of the unweighted angle. In order to have a wide database or radar signatures and evaluate the performance, range profiles are obtained through numerical simulation of seven civil aircraft at defined trajectories taken from an actual measurement. Taking into account the nature of the datasets, the main drawback of using simulated profiles instead of actual measured profiles is that the former implies an ideal identification scenario, since measured profiles suffer from noise, clutter and other unwanted information and simulated profiles don't. In this case, the test and training samples have similar nature and usually a similar high signal-to-noise ratio, so as to assess the feasibility of the approach, the addition of noise has been considered before the creation of the test set. The identification results applying the unweighted and weighted metrics are analysed for demonstrating which algorithm provides the best robustness against noise in an actual possible scenario. So as to confirm the validity of the methodology, identification experiments of profiles coming from electromagnetic simulations are conducted, revealing promising results. Considering the dissimilarities between the test and training sets when noise is added, the recognition performance has been improved when weighting is applied. Future experiments with larger sets are expected to be conducted with the aim of finally using actual profiles as test sets in a real hostile situation.

Keywords: HRRP, NCTI, simulated/synthetic database, SVD

Procedia PDF Downloads 334
414 Development of a Decision-Making Method by Using Machine Learning Algorithms in the Early Stage of School Building Design

Authors: Rajaian Hoonejani Mohammad, Eshraghi Pegah, Zomorodian Zahra Sadat, Tahsildoost Mohammad

Abstract:

Over the past decade, energy consumption in educational buildings has steadily increased. The purpose of this research is to provide a method to quickly predict the energy consumption of buildings using separate evaluation of zones and decomposing the building to eliminate the complexity of geometry at the early design stage. To produce this framework, machine learning algorithms such as Support vector regression (SVR) and Artificial neural network (ANN) are used to predict energy consumption and thermal comfort metrics in a school as a case. The database consists of more than 55000 samples in three climates of Iran. Cross-validation evaluation and unseen data have been used for validation. In a specific label, cooling energy, it can be said the accuracy of prediction is at least 84% and 89% in SVR and ANN, respectively. The results show that the SVR performed much better than the ANN.

Keywords: early stage of design, energy, thermal comfort, validation, machine learning

Procedia PDF Downloads 42
413 Node Pair Selection Scheme in Relay-Aided Communication Based on Stable Marriage Problem

Authors: Tetsuki Taniguchi, Yoshio Karasawa

Abstract:

This paper describes a node pair selection scheme in relay-aided multiple source multiple destination communication system based on stable marriage problem. A general case is assumed in which all of source, relay and destination nodes are equipped with multiantenna and carry out multistream transmission. Based on several metrics introduced from inter-node channel condition, the preference order is determined about all source-relay and relay-destination relations, and then the node pairs are determined using Gale-Shapley algorithm. The computer simulations show that the effectiveness of node pair selection is larger in multihop communication. Some additional aspects which are different from relay-less case are also investigated.

Keywords: relay, multiple input multiple output (MIMO), multiuser, amplify and forward, stable marriage problem, Gale-Shapley algorithm

Procedia PDF Downloads 379
412 An Energy Efficient Clustering Approach for Underwater ‎Wireless Sensor Networks

Authors: Mohammad Reza Taherkhani‎

Abstract:

Wireless sensor networks that are used to monitor a special environment, are formed from a large number of sensor nodes. The role of these sensors is to sense special parameters from ambient and to make a connection. In these networks, the most important challenge is the management of energy usage. Clustering is one of the methods that are broadly used to face this challenge. In this paper, a distributed clustering protocol based on learning automata is proposed for underwater wireless sensor networks. The proposed algorithm that is called LA-Clustering forms clusters in the same energy level, based on the energy level of nodes and the connection radius regardless of size and the structure of sensor network. The proposed approach is simulated and is compared with some other protocols with considering some metrics such as network lifetime, number of alive nodes, and number of transmitted data. The simulation results demonstrate the efficiency of the proposed approach.

Keywords: underwater sensor networks, clustering, learning automata, energy consumption

Procedia PDF Downloads 337
411 A New Concept for Deriving the Expected Value of Fuzzy Random Variables

Authors: Liang-Hsuan Chen, Chia-Jung Chang

Abstract:

Fuzzy random variables have been introduced as an imprecise concept of numeric values for characterizing the imprecise knowledge. The descriptive parameters can be used to describe the primary features of a set of fuzzy random observations. In fuzzy environments, the expected values are usually represented as fuzzy-valued, interval-valued or numeric-valued descriptive parameters using various metrics. Instead of the concept of area metric that is usually adopted in the relevant studies, the numeric expected value is proposed by the concept of distance metric in this study based on two characters (fuzziness and randomness) of FRVs. Comparing with the existing measures, although the results show that the proposed numeric expected value is same with those using the different metric, if only triangular membership functions are used. However, the proposed approach has the advantages of intuitiveness and computational efficiency, when the membership functions are not triangular types. An example with three datasets is provided for verifying the proposed approach.

Keywords: fuzzy random variables, distance measure, expected value, descriptive parameters

Procedia PDF Downloads 322
410 Modelling of Atomic Force Microscopic Nano Robot's Friction Force on Rough Surfaces

Authors: M. Kharazmi, M. Zakeri, M. Packirisamy, J. Faraji

Abstract:

Micro/Nanorobotics or manipulation of nanoparticles by Atomic Force Microscopic (AFM) is one of the most important solutions for controlling the movement of atoms, particles and micro/nano metrics components and assembling of them to design micro/nano-meter tools. Accurate modelling of manipulation requires identification of forces and mechanical knowledge in the Nanoscale which are different from macro world. Due to the importance of the adhesion forces and the interaction of surfaces at the nanoscale several friction models were presented. In this research, friction and normal forces that are applied on the AFM by using of the dynamic bending-torsion model of AFM are obtained based on Hurtado-Kim friction model (HK), Johnson-Kendall-Robert contact model (JKR) and Greenwood-Williamson roughness model (GW). Finally, the effect of standard deviation of asperities height on the normal load, friction force and friction coefficient are studied.

Keywords: atomic force microscopy, contact model, friction coefficient, Greenwood-Williamson model

Procedia PDF Downloads 179
409 Low-Temperature Fabrication of Reaction Bonded Composites, Based on Sic and (Sic+B4C) Mixture, Infiltrated with Si-Al Alloy

Authors: Helen Dilman, Eyal Oz, Shmuel Hayun, Nahum Frage

Abstract:

The conventional approach for manufacturing silicon carbide and boron carbide reaction bonded composites is based on infiltrating a ceramic porous preform with molten silicon. The relatively high melting temperature of the silicon infiltrating medium is a drawback of the process. The present contribution is concerned with an approach that allows obtaining reaction bonded composites by pressure-less infiltration at a significantly lower (850-1000oC) temperature range. This approach was applied for the fabrication of fully dense SiC/(Si-Al) and (SiC+B4C)/(Si-Al) composites. The key feature of the approach is based on using Si alloys with low melting temperature and the Mg-vapor atmosphere, under which an adequate wetting between ceramics and liquid alloys for the infiltration process is achieved. In the first set of the experiments ceramic performs compacted from multimodal SiC powders (with the green density of about 27 vol. %) without free carbon addition were infiltrated by Si-20%Al alloy at 950oC. In the second set, 19 vol. % of a fine boron carbide powder was added to SiC powders as a source of carbon. The green density of the SiC-B4C preforms was about 23-25 vol. %. In both cases, successful infiltration was achieved and the composites were fully dense. The density of the composites was about 3g/cm3. For the SiC based composites the hardness value was 750±150HV, Young modulus-280GPa and bending strength-240±30MPa. These values for (SiC-B4C)/(Si-Al) composites (1460±200HV, 317GPa and 360±20MPa) were significantly higher due to the formation of novel ceramics phases. Microstructural characteristics of the composites and their phase composition will be discussed.

Keywords: boron carbide, composites, infiltration, low temperatures, silicon carbide

Procedia PDF Downloads 533
408 Analysis of Solvent Effect on the Mechanical Properties of Poly(Ether Ether Ketone) Using Nano-Indentation

Authors: Tanveer Iqbal, Saima Yasin, Muhammad Zafar, Ahmad Shakeel, Fahad Nazir, Paul F. Luckham

Abstract:

The contact performance of polymeric composites is dependent on the localized mechanical properties of materials. This is particularly important for fiber oriented polymeric materials where self-lubrication from top layers has been the basic requirement. The nanoindentation response of fiber reinforced poly(etheretherketone), PEEK, composites have been evaluated to determine the near-surface mechanical characteristics. Load-displacement compliance, hardness and elastic modulus data based on contact compliance mode (CSM) indentation of carbon fiber oriented and glass fiber oriented PEEK composites are reported as a function of indentation contact displacement. The composite surfaces were indented to a maximum penetration depth of 5µm using Berkovich tip indenter. A typical multiphase response of the composite surface is depicted from analysis of the indentation data for the composites, showing presence of polymer matrix, fibers, and interphase regions. The observed experimental results show that although the surface mechanical properties of carbon fiber based PEEK composite were comparatively higher, the properties of matrix material were seen to be increased in the presence of glass fibers. The experimental methodology may provide a convenient means to understand morphological description of the multimodal polymeric composites.

Keywords: nanoindentation, PEEK, modulus, hardness, plasticization

Procedia PDF Downloads 171
407 Adaption Model for Building Agile Pronunciation Dictionaries Using Phonemic Distance Measurements

Authors: Akella Amarendra Babu, Rama Devi Yellasiri, Natukula Sainath

Abstract:

Where human beings can easily learn and adopt pronunciation variations, machines need training before put into use. Also humans keep minimum vocabulary and their pronunciation variations are stored in front-end of their memory for ready reference, while machines keep the entire pronunciation dictionary for ready reference. Supervised methods are used for preparation of pronunciation dictionaries which take large amounts of manual effort, cost, time and are not suitable for real time use. This paper presents an unsupervised adaptation model for building agile and dynamic pronunciation dictionaries online. These methods mimic human approach in learning the new pronunciations in real time. A new algorithm for measuring sound distances called Dynamic Phone Warping is presented and tested. Performance of the system is measured using an adaptation model and the precision metrics is found to be better than 86 percent.

Keywords: pronunciation variations, dynamic programming, machine learning, natural language processing

Procedia PDF Downloads 150
406 A Network Approach to Analyzing Financial Markets

Authors: Yusuf Seedat

Abstract:

The necessity to understand global financial markets has increased following the unfortunate spread of the recent financial crisis around the world. Financial markets are considered to be complex systems consisting of highly volatile move-ments whose indexes fluctuate without any clear pattern. Analytic methods of stock prices have been proposed in which financial markets are modeled using common network analysis tools and methods. It has been found that two key components of social network analysis are relevant to modeling financial markets, allowing us to forecast accurate predictions of stock prices within the financial market. Financial markets have a number of interacting components, leading to complex behavioral patterns. This paper describes a social network approach to analyzing financial markets as a viable approach to studying the way complex stock markets function. We also look at how social network analysis techniques and metrics are used to gauge an understanding of the evolution of financial markets as well as how community detection can be used to qualify and quantify in-fluence within a network.

Keywords: network analysis, social networks, financial markets, stocks, nodes, edges, complex networks

Procedia PDF Downloads 172
405 Transforming ESL Teaching and Learning with ICT

Authors: Helena Sit

Abstract:

Developing skills in using ICT in the language classroom has been discussed at all educational levels. Digital tools and learning management systems enable teachers to transform their instructional activities while giving learners the opportunity to engage with virtual communities. In the field of English as a second language (ESL) teaching and learning, the use of technology-enhanced learning and diverse pedagogical practices continues to grow. Whilst technology and multimodal learning is a way of the future for education, second language teachers now face the predicament as to whether implementing these newer ways of learning is, in fact, beneficial or disadvantageous to learners. Research has shown that integrating multimodality and technology can improve students’ engagement and participation in their English language learning. However, students can experience anxiety or misunderstanding when engaging with E-learning or digital-mediated learning. This paper aims to explore how ESL teaching and learning are transformed via the use of educational technology and what impact it has had on student teachers. Case study is employed in this research. The study reviews the growing presence of technology and multimodality in university language classrooms, discusses their impact on teachers’ pedagogical practices, and proposes scaffolding strategies to help design effective English language courses in the Australian education context. The study sheds light on how pedagogical integration today may offer a way forward for language teachers of tomorrow and provides implications to implement an evidence-informed approach that blends knowledge from research, practice and people experiencing the practice in the digital era.

Keywords: educational technology, ICT in higher education, curriculum design and innovation, teacher education, multiliteracies pedagogy

Procedia PDF Downloads 53
404 Using Greywolf Optimized Machine Learning Algorithms to Improve Accuracy for Predicting Hospital Readmission for Diabetes

Authors: Vincent Liu

Abstract:

Machine learning algorithms (ML) can achieve high accuracy in predicting outcomes compared to classical models. Metaheuristic, nature-inspired algorithms can enhance traditional ML algorithms by optimizing them such as by performing feature selection. We compare ten ML algorithms to predict 30-day hospital readmission rates for diabetes patients in the US using a dataset from UCI Machine Learning Repository with feature selection performed by Greywolf nature-inspired algorithm. The baseline accuracy for the initial random forest model was 65%. After performing feature engineering, SMOTE for class balancing, and Greywolf optimization, the machine learning algorithms showed better metrics, including F1 scores, accuracy, and confusion matrix with improvements ranging in 10%-30%, and a best model of XGBoost with an accuracy of 95%. Applying machine learning this way can improve patient outcomes as unnecessary rehospitalizations can be prevented by focusing on patients that are at a higher risk of readmission.

Keywords: diabetes, machine learning, 30-day readmission, metaheuristic

Procedia PDF Downloads 33
403 Design of Digital IIR Filter Using Opposition Learning and Artificial Bee Colony Algorithm

Authors: J. S. Dhillon, K. K. Dhaliwal

Abstract:

In almost all the digital filtering applications the digital infinite impulse response (IIR) filters are preferred over finite impulse response (FIR) filters because they provide much better performance, less computational cost and have smaller memory requirements for similar magnitude specifications. However, the digital IIR filters are generally multimodal with respect to the filter coefficients and therefore, reliable methods that can provide global optimal solutions are required. The artificial bee colony (ABC) algorithm is one such recently introduced meta-heuristic optimization algorithm. But in some cases it shows insufficiency while searching the solution space resulting in a weak exchange of information and hence is not able to return better solutions. To overcome this deficiency, the opposition based learning strategy is incorporated in ABC and hence a modified version called oppositional artificial bee colony (OABC) algorithm is proposed in this paper. Duplication of members is avoided during the run which also augments the exploration ability. The developed algorithm is then applied for the design of optimal and stable digital IIR filter structure where design of low-pass (LP) and high-pass (HP) filters is carried out. Fuzzy theory is applied to achieve maximize satisfaction of minimum magnitude error and stability constraints. To check the effectiveness of OABC, the results are compared with some well established filter design techniques and it is observed that in most cases OABC returns better or atleast comparable results.

Keywords: digital infinite impulse response filter, artificial bee colony optimization, opposition based learning, digital filter design, multi-parameter optimization

Procedia PDF Downloads 451
402 Hyper Tuned RBF SVM: Approach for the Prediction of the Breast Cancer

Authors: Surita Maini, Sanjay Dhanka

Abstract:

Machine learning (ML) involves developing algorithms and statistical models that enable computers to learn and make predictions or decisions based on data without being explicitly programmed. Because of its unlimited abilities ML is gaining popularity in medical sectors; Medical Imaging, Electronic Health Records, Genomic Data Analysis, Wearable Devices, Disease Outbreak Prediction, Disease Diagnosis, etc. In the last few decades, many researchers have tried to diagnose Breast Cancer (BC) using ML, because early detection of any disease can save millions of lives. Working in this direction, the authors have proposed a hybrid ML technique RBF SVM, to predict the BC in earlier the stage. The proposed method is implemented on the Breast Cancer UCI ML dataset with 569 instances and 32 attributes. The authors recorded performance metrics of the proposed model i.e., Accuracy 98.24%, Sensitivity 98.67%, Specificity 97.43%, F1 Score 98.67%, Precision 98.67%, and run time 0.044769 seconds. The proposed method is validated by K-Fold cross-validation.

Keywords: breast cancer, support vector classifier, machine learning, hyper parameter tunning

Procedia PDF Downloads 53
401 The Masterplan for the Urban Regeneration of the Heritage District of Msheireb Downtown Doha, State of Qatar

Authors: Raffaello Furlan

Abstract:

In the 21st century, the sustainable urban development of GCC-cities is challenged by inhabitants’ over-dependency on private-use vehicles. In turn, this habit has generated problems of urban inefficiency, contributing to traffic congestion, pollution, urban sprawling, fragmentation of the urban fabric, and various environmental and social challenges. In the context of Doha, the capital city of the State of Qatar, the over-dependency on private-use vehicles is justified by the lack of alternative public modes of transportation that support the need to connect fragmented urban districts and provide an effective solution to urban sprawl. Therefore, the current construction of the Qatar Metro Rail is offering the potential for investigating and defining a strategy for the sustainable urban development and/or urban regeneration of transit villages (TODs) in Qatar. Namely, the aim of this research study is (i) to investigate the development of transit villages (TODs) in the cultural-heritage district of Msheireb, Downtown Doha, (ii) to explore how the introduction of the new public transport system of Doha Metro can be effectively utilized as means of urban regeneration of the cultural core of the city, (iii) to propose a masterplan for TOD suitable for the district, suiting and responding to regional cultural and societal values. The findings reveal that the strategies for the sustainable urban regeneration of Msheireb are based on (i) the integration of land-use and multimodal transportation systems, (ii) the implementation of the public realm, and (iii) conservation of culture and urban identity.

Keywords: sustainable urbanism, smart growth, TODs, cultural district, Msheireb Downtown Doha

Procedia PDF Downloads 212
400 Text Emotion Recognition by Multi-Head Attention based Bidirectional LSTM Utilizing Multi-Level Classification

Authors: Vishwanath Pethri Kamath, Jayantha Gowda Sarapanahalli, Vishal Mishra, Siddhesh Balwant Bandgar

Abstract:

Recognition of emotional information is essential in any form of communication. Growing HCI (Human-Computer Interaction) in recent times indicates the importance of understanding of emotions expressed and becomes crucial for improving the system or the interaction itself. In this research work, textual data for emotion recognition is used. The text being the least expressive amongst the multimodal resources poses various challenges such as contextual information and also sequential nature of the language construction. In this research work, the proposal is made for a neural architecture to resolve not less than 8 emotions from textual data sources derived from multiple datasets using google pre-trained word2vec word embeddings and a Multi-head attention-based bidirectional LSTM model with a one-vs-all Multi-Level Classification. The emotions targeted in this research are Anger, Disgust, Fear, Guilt, Joy, Sadness, Shame, and Surprise. Textual data from multiple datasets were used for this research work such as ISEAR, Go Emotions, Affect datasets for creating the emotions’ dataset. Data samples overlap or conflicts were considered with careful preprocessing. Our results show a significant improvement with the modeling architecture and as good as 10 points improvement in recognizing some emotions.

Keywords: text emotion recognition, bidirectional LSTM, multi-head attention, multi-level classification, google word2vec word embeddings

Procedia PDF Downloads 152
399 A Blending Analysis of Metaphors and Metonymies Used to Depict the Deal of the Century by Jordanian Cartoonists

Authors: Aseel Zibin, Abdel Rahman Altakhaineh

Abstract:

This study analyses 30 cartoons depicting THE DEAL OF THE CENTURY as envisaged by two Jordanian cartoonists, namely, EmadHajjaj and Osama Hajjaj. Conceptual Blending Theory (CBT) and Multimodal Metaphor Theory (MMT) are adopted as a theoretical framework to interpret the metaphors and metonymies used in the target cartoons. The results reveal that the target domain THE DEAL OF THE CENTURY was conceptualized mainly through layered metaphors that have metonymic basis and event metaphors\allegories. Specifically, 6 groups were identified: OBJECT or a situation involving OBJECTS, situations involving HUMANS\HYBRIDS of HUMANS and OBJECTS, an ANIMAL OR situation involving an ANIMAL, hybrids of WEAPONS and humans, and event metaphors used to build a story\allegory. The target domain was also depicted via event metaphors used to build a story; some of which are embedded in the Jordanian culture, while others could be perceivable cross-culturally. The results also demonstrate that the most widely used configurations to construe the metaphors was the pictorial source–verbal target in line with Lan and Zuo (2016); the motivation was probably the greater conceptual density and concreteness of visual representation since the target is better captured verbally because of its abstractness. The use of cross-modal mappings of this type was attributed to the abstractness of the target domain, THE DEAL OF THE CENTURY, which makes it more construable via verbal cues rather than visual ones. In contrast, the source domains used were mainly concrete and thus perceivable pictorially rather than verbally.

Keywords: semiotics, cognitive semantics, metaphor, culture, blending, cartoon

Procedia PDF Downloads 135
398 What Smart Can Learn about Art

Authors: Faten Hatem

Abstract:

This paper explores the associated understanding of the role and meaning of art and whether it is perceived to be separate from smart city construction. The paper emphasises the significance of fulfilling the inherent need for discovery and interaction, driving people to explore new places and think of works of art. This is done by exploring the ways of thinking and types of art in Milton Keynes by illustrating a general pattern of misunderstanding that relies on the separation between smart, art, and architecture, promoting a better and deeper understanding of the interconnections between neuroscience, art, and architecture. A reflective approach is used to clarify the potential and impact of using art-based research, methodology, and ways of knowing when approaching global phenomena and knowledge production while examining the process of making and developing smart cities, in particular, asserting that factors can severely impact it in the process of conducting the study itself. It follows a case study as a research strategy. The qualitative methods included data collection and analysis that involved interviews and observations that depended on visuals.

Keywords: smart cities, art and smart, smart cities design, smart cities making, sustainability, city brain and smart cities metrics, smart cities standards, smart cities applications, governance, planning and policy

Procedia PDF Downloads 81
397 Determination of Water Pollution and Water Quality with Decision Trees

Authors: Çiğdem Bakır, Mecit Yüzkat

Abstract:

With the increasing emphasis on water quality worldwide, the search for and expanding the market for new and intelligent monitoring systems has increased. The current method is the laboratory process, where samples are taken from bodies of water, and tests are carried out in laboratories. This method is time-consuming, a waste of manpower, and uneconomical. To solve this problem, we used machine learning methods to detect water pollution in our study. We created decision trees with the Orange3 software we used in our study and tried to determine all the factors that cause water pollution. An automatic prediction model based on water quality was developed by taking many model inputs such as water temperature, pH, transparency, conductivity, dissolved oxygen, and ammonia nitrogen with machine learning methods. The proposed approach consists of three stages: preprocessing of the data used, feature detection, and classification. We tried to determine the success of our study with different accuracy metrics and the results. We presented it comparatively. In addition, we achieved approximately 98% success with the decision tree.

Keywords: decision tree, water quality, water pollution, machine learning

Procedia PDF Downloads 65
396 Performance of Neural Networks vs. Radial Basis Functions When Forming a Metamodel for Residential Buildings

Authors: Philip Symonds, Jon Taylor, Zaid Chalabi, Michael Davies

Abstract:

With the world climate projected to warm and major cities in developing countries becoming increasingly populated and polluted, governments are tasked with the problem of overheating and air quality in residential buildings. This paper presents the development of an adaptable model of these risks. Simulations are performed using the EnergyPlus building physics software. An accurate metamodel is formed by randomly sampling building input parameters and training on the outputs of EnergyPlus simulations. Metamodels are used to vastly reduce the amount of computation time required when performing optimisation and sensitivity analyses. Neural Networks (NNs) are compared to a Radial Basis Function (RBF) algorithm when forming a metamodel. These techniques were implemented using the PyBrain and scikit-learn python libraries, respectively. NNs are shown to perform around 15% better than RBFs when estimating overheating and air pollution metrics modelled by EnergyPlus.

Keywords: neural networks, radial basis functions, metamodelling, python machine learning libraries

Procedia PDF Downloads 424
395 Video-On-Demand QoE Evaluation across Different Age-Groups and Its Significance for Network Capacity

Authors: Mujtaba Roshan, John A. Schormans

Abstract:

Quality of Experience (QoE) drives churn in the broadband networks industry, and good QoE plays a large part in the retention of customers. QoE is known to be affected by the Quality of Service (QoS) factors packet loss probability (PLP), delay and delay jitter caused by the network. Earlier results have shown that the relationship between these QoS factors and QoE is non-linear, and may vary from application to application. We use the network emulator Netem as the basis for experimentation, and evaluate how QoE varies as we change the emulated QoS metrics. Focusing on Video-on-Demand, we discovered that the reported QoE may differ widely for users of different age groups, and that the most demanding age group (the youngest) can require an order of magnitude lower PLP to achieve the same QoE than is required by the most widely studied age group of users. We then used a bottleneck TCP model to evaluate the capacity cost of achieving an order of magnitude decrease in PLP, and found it be (almost always) a 3-fold increase in link capacity that was required.

Keywords: network capacity, packet loss probability, quality of experience, quality of service

Procedia PDF Downloads 256
394 Real-Time Lane Marking Detection Using Weighted Filter

Authors: Ayhan Kucukmanisa, Orhan Akbulut, Oguzhan Urhan

Abstract:

Nowadays, advanced driver assistance systems (ADAS) have become popular, since they enable safe driving. Lane detection is a vital step for ADAS. The performance of the lane detection process is critical to obtain a high accuracy lane departure warning system (LDWS). Challenging factors such as road cracks, erosion of lane markings, weather conditions might affect the performance of a lane detection system. In this paper, 1-D weighted filter based on row filtering to detect lane marking is proposed. 2-D input image is filtered by 1-D weighted filter considering four-pixel values located symmetrically around the center of candidate pixel. Performance evaluation is carried out by two metrics which are true positive rate (TPR) and false positive rate (FPR). Experimental results demonstrate that the proposed approach provides better lane marking detection accuracy compared to the previous methods while providing real-time processing performance.

Keywords: lane marking filter, lane detection, ADAS, LDWS

Procedia PDF Downloads 170