Search results for: multimedia learning tools
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 10249

Search results for: multimedia learning tools

4399 Failure Analysis and Fatigue Life Estimation of a Shaft of a Rotary Draw Bending Machine

Authors: B. Engel, Sara Salman Hassan Al-Maeeni

Abstract:

Human consumption of the Earth's resources increases the need for a sustainable development as an important ecological, social, and economic theme. Re-engineering of machine tools, in terms of design and failure analysis, is defined as steps performed on an obsolete machine to return it to a new machine with the warranty that matches the customer requirement. To understand the future fatigue behavior of the used machine components, it is important to investigate the possible causes of machine parts failure through design, surface, and material inspections. In this study, the failure modes of the shaft of the rotary draw bending machine are inspected. Furthermore, stress and deflection analysis of the shaft subjected to combined torsion and bending loads are carried out by an analytical method and compared with a finite element analysis method. The theoretical fatigue strength, correction factors, and fatigue life sustained by the shaft before damaged are estimated by creating a stress-cycle (S-N) diagram. In conclusion, it is seen that the shaft can work in the second life, but it needs some surface treatments to increase the reliability and fatigue life.

Keywords: failure analysis, fatigue life, FEM analysis, shaft, stress analysis

Procedia PDF Downloads 278
4398 Analysis of Atomic Models in High School Physics Textbooks

Authors: Meng-Fei Cheng, Wei Fneg

Abstract:

New Taiwan high school standards emphasize employing scientific models and modeling practices in physics learning. However, to our knowledge. Few studies address how scientific models and modeling are approached in current science teaching, and they do not examine the views of scientific models portrayed in the textbooks. To explore the views of scientific models and modeling in textbooks, this study investigated the atomic unit in different textbook versions as an example and provided suggestions for modeling curriculum. This study adopted a quantitative analysis of qualitative data in the atomic units of four mainstream version of Taiwan high school physics textbooks. The models were further analyzed using five dimensions of the views of scientific models (nature of models, multiple models, purpose of the models, testing models, and changing models); each dimension had three levels (low, medium, high). Descriptive statistics were employed to compare the frequency of describing the five dimensions of the views of scientific models in the atomic unit to understand the emphasis of the views and to compare the frequency of the eight scientific models’ use to investigate the atomic model that was used most often in the textbooks. Descriptive statistics were further utilized to investigate the average levels of the five dimensions of the views of scientific models to examine whether the textbooks views were close to the scientific view. The average level of the five dimensions of the eight atomic models were also compared to examine whether the views of the eight atomic models were close to the scientific views. The results revealed the following three major findings from the atomic unit. (1) Among the five dimensions of the views of scientific models, the most portrayed dimension was the 'purpose of models,' and the least portrayed dimension was 'multiple models.' The most diverse view was the 'purpose of models,' and the most sophisticated scientific view was the 'nature of models.' The least sophisticated scientific view was 'multiple models.' (2) Among the eight atomic models, the most mentioned model was the atomic nucleus model, and the least mentioned model was the three states of matter. (3) Among the correlations between the five dimensions, the dimension of 'testing models' was highly related to the dimension of 'changing models.' In short, this study examined the views of scientific models based on the atomic units of physics textbooks to identify the emphasized and disregarded views in the textbooks. The findings suggest how future textbooks and curriculum can provide a thorough view of scientific models to enhance students' model-based learning.

Keywords: atomic models, textbooks, science education, scientific model

Procedia PDF Downloads 145
4397 Towards Efficient Reasoning about Families of Class Diagrams Using Union Models

Authors: Tejush Badal, Sanaa Alwidian

Abstract:

Class diagrams are useful tools within the Unified Modelling Language (UML) to model and visualize the relationships between, and properties of objects within a system. As a system evolves over time and space (e.g., products), a series of models with several commonalities and variabilities create what is known as a model family. In circumstances where there are several versions of a model, examining each model individually, becomes expensive in terms of computation resources. To avoid performing redundant operations, this paper proposes an approach for representing a family of class diagrams into Union Models to represent model families using a single generic model. The paper aims to analyze and reason about a family of class diagrams using union models as opposed to individual analysis of each member model in the family. The union algorithm provides a holistic view of the model family, where the latter cannot be otherwise obtained from an individual analysis approach, this in turn, enhances the analysis performed in terms of speeding up the time needed to analyze a family of models together as opposed to analyzing individual models, one model at a time.

Keywords: analysis, class diagram, model family, unified modeling language, union model

Procedia PDF Downloads 58
4396 Expectation and Satisfaction of Health Spa Business Service, Ranong Province, Thailand

Authors: Supattra Pranee

Abstract:

The purposes of this research are to study the current business of health spa and to study the customers’ level of expectation as well as level of satisfaction of the health spa business in Ranong, Thailand. This paper drew upon data collected from health spa customers by using questionnaire. In addition, an in-depth interview was utilized to collect data from health spa entrepreneurs. The findings revealed that the health spa business is growing very fast and the coming ASEAN Economic Community (AEC) will ameliorate the business growth and increase the customer base. There is a need to improve staff’s ability to communicate in English. However, the economic size of Ranong province is still small which has resulted in the hesitation of investors to increase their investment in this business. The findings also revealed four categories of level of expectation and satisfaction as follows: (1) Service: overall, customers had a high expectation with a mean of 3.80 and 0.873 SD and a high level of satisfaction with a mean of 3.66 and 0.704 SD. (2) Staff: overall, customers had a high expectation with a mean of 3.95 and 0.865 SD and a high level of satisfaction with a mean of 3.84 and 0.783 SD. (3) Product, Equipment, and Tools: overall, customers had a high expectation with a mean of 4.02 and 0.913 SD and a high level of satisfaction with a mean of 3.88 and 0.772 SD. (4) Place, Atmosphere, and Environment: overall, customers had a high expectation with a mean of 3.95 and 0.906 SD and a high level of satisfaction with a mean of 3.86 and 0.785 SD.

Keywords: expectation, health spa business, satisfaction, ranong province

Procedia PDF Downloads 292
4395 Quantum Dot Biosensing for Advancing Precision Cancer Detection

Authors: Sourav Sarkar, Manashjit Gogoi

Abstract:

In the evolving landscape of cancer diagnostics, optical biosensing has emerged as a promising tool due to its sensitivity and specificity. This study explores the potential of CdS/ZnS core-shell quantum dots (QDs) capped with 3-Mercaptopropionic acid (3-MPA), which aids in the linking chemistry of QDs to various cancer antibodies. The QDs, with their unique optical and electronic properties, have been integrated into the biosensor design. Their high quantum yield and size-dependent emission spectra have been exploited to improve the sensor’s detection capabilities. The study presents the design of this QD-enhanced optical biosensor. The use of these QDs can also aid multiplexed detection, enabling simultaneous monitoring of different cancer biomarkers. This innovative approach holds significant potential for advancing cancer diagnostics, contributing to timely and accurate detection. Future work will focus on optimizing the biosensor design for clinical applications and exploring the potential of QDs in other biosensing applications. This study underscores the potential of integrating nanotechnology and biosensing for cancer research, paving the way for next-generation diagnostic tools. It is a step forward in our quest for achieving precision oncology.

Keywords: quantum dots, biosensing, cancer, device

Procedia PDF Downloads 43
4394 Content-Aware Image Augmentation for Medical Imaging Applications

Authors: Filip Rusak, Yulia Arzhaeva, Dadong Wang

Abstract:

Machine learning based Computer-Aided Diagnosis (CAD) is gaining much popularity in medical imaging and diagnostic radiology. However, it requires a large amount of high quality and labeled training image datasets. The training images may come from different sources and be acquired from different radiography machines produced by different manufacturers, digital or digitized copies of film radiographs, with various sizes as well as different pixel intensity distributions. In this paper, a content-aware image augmentation method is presented to deal with these variations. The results of the proposed method have been validated graphically by plotting the removed and added seams of pixels on original images. Two different chest X-ray (CXR) datasets are used in the experiments. The CXRs in the datasets defer in size, some are digital CXRs while the others are digitized from analog CXR films. With the proposed content-aware augmentation method, the Seam Carving algorithm is employed to resize CXRs and the corresponding labels in the form of image masks, followed by histogram matching used to normalize the pixel intensities of digital radiography, based on the pixel intensity values of digitized radiographs. We implemented the algorithms, resized the well-known Montgomery dataset, to the size of the most frequently used Japanese Society of Radiological Technology (JSRT) dataset and normalized our digital CXRs for testing. This work resulted in the unified off-the-shelf CXR dataset composed of radiographs included in both, Montgomery and JSRT datasets. The experimental results show that even though the amount of augmentation is large, our algorithm can preserve the important information in lung fields, local structures, and global visual effect adequately. The proposed method can be used to augment training and testing image data sets so that the trained machine learning model can be used to process CXRs from various sources, and it can be potentially used broadly in any medical imaging applications.

Keywords: computer-aided diagnosis, image augmentation, lung segmentation, medical imaging, seam carving

Procedia PDF Downloads 204
4393 Flood Early Warning and Management System

Authors: Yogesh Kumar Singh, T. S. Murugesh Prabhu, Upasana Dutta, Girishchandra Yendargaye, Rahul Yadav, Rohini Gopinath Kale, Binay Kumar, Manoj Khare

Abstract:

The Indian subcontinent is severely affected by floods that cause intense irreversible devastation to crops and livelihoods. With increased incidences of floods and their related catastrophes, an Early Warning System for Flood Prediction and an efficient Flood Management System for the river basins of India is a must. Accurately modeled hydrological conditions and a web-based early warning system may significantly reduce economic losses incurred due to floods and enable end users to issue advisories with better lead time. This study describes the design and development of an EWS-FP using advanced computational tools/methods, viz. High-Performance Computing (HPC), Remote Sensing, GIS technologies, and open-source tools for the Mahanadi River Basin of India. The flood prediction is based on a robust 2D hydrodynamic model, which solves shallow water equations using the finite volume method. Considering the complexity of the hydrological modeling and the size of the basins in India, it is always a tug of war between better forecast lead time and optimal resolution at which the simulations are to be run. High-performance computing technology provides a good computational means to overcome this issue for the construction of national-level or basin-level flash flood warning systems having a high resolution at local-level warning analysis with a better lead time. High-performance computers with capacities at the order of teraflops and petaflops prove useful while running simulations on such big areas at optimum resolutions. In this study, a free and open-source, HPC-based 2-D hydrodynamic model, with the capability to simulate rainfall run-off, river routing, and tidal forcing, is used. The model was tested for a part of the Mahanadi River Basin (Mahanadi Delta) with actual and predicted discharge, rainfall, and tide data. The simulation time was reduced from 8 hrs to 3 hrs by increasing CPU nodes from 45 to 135, which shows good scalability and performance enhancement. The simulated flood inundation spread and stage were compared with SAR data and CWC Observed Gauge data, respectively. The system shows good accuracy and better lead time suitable for flood forecasting in near-real-time. To disseminate warning to the end user, a network-enabled solution is developed using open-source software. The system has query-based flood damage assessment modules with outputs in the form of spatial maps and statistical databases. System effectively facilitates the management of post-disaster activities caused due to floods, like displaying spatial maps of the area affected, inundated roads, etc., and maintains a steady flow of information at all levels with different access rights depending upon the criticality of the information. It is designed to facilitate users in managing information related to flooding during critical flood seasons and analyzing the extent of the damage.

Keywords: flood, modeling, HPC, FOSS

Procedia PDF Downloads 80
4392 Through Integrated Project Management and Systems Engineering to Support System Design Development: A Project Management-based Systems Engineering Approach

Authors: Xiaojing Gao, James Njuguna

Abstract:

This paper emphasizes the importance of integrating project management and systems engineering for innovative system design and production development. The research highlights the need for a flexible approach that unifies these disciplines, as their isolation often leads to communication challenges and complexity within multidisciplinary teams. The paper aims to elucidate the intricate relationship between project management and systems engineering, recommending the consolidation of engineering disciplines into a single lifecycle for improved support of the design and development process. The research identifies a synergy between these disciplines, focusing on streamlining information communication during product design and development. The insights gained from this process can lead to product design optimization. Additionally, the paper introduces a proposed Project Management-Based Systems Engineering (PMBSE) framework, emphasizing effective communication, efficient processes, and advanced tools to enhance product development outcomes within the product lifecycle.

Keywords: system engineering, product design and development, project management, cross-disciplinary

Procedia PDF Downloads 56
4391 A Mechanism of Reusable, Portable, and Reliable Script Generator on Android

Authors: Kuei-Chun Liu, Yu-Yu Lai, Ching-Hong Wu

Abstract:

A good automated testing tool could reduce as much as possible the manual work done by testers. Traditional record-replay testing tool provides an automated testing solution by recording mouse coordinates as test scripts, but it will be easily broken if any change of resolutions. Therefore, more and more testers design multiple test scripts to automate the testing process for different devices. In order to improve the traditional record-replay approach and reduce the effort that the testers spending on writing test scripts, we propose an approach for generating the Android application test scripts based on accessibility service without connecting to a computer. This approach simulates user input actions and replays them correctly even at the different conditions such as the internet connection is unstable when the device under test, the different resolutions on Android devices. In this paper, we describe how to generate test scripts automatically and make a comparison with existing tools for Android such as Robotium, Appium, UIAutomator, and MonkeyTalk.

Keywords: accessibility service, Appium, automated testing, MonkeyTalk, Robotium, testing, UIAutomator

Procedia PDF Downloads 362
4390 Connecting Students and Faculty Research Efforts through the Research and Projects Portal

Authors: Havish Nalapareddy, Mark V. Albert, Ranak Bansal, Avi Udash, Lin Lin

Abstract:

Students engage in many course projects during their degree programs. However, impactful projects often need a time frame longer than a single semester. Ideally, projects are documented and structured to be readily accessible to future students who may choose to continue the project, with features that emphasize the local community, university, or course structure. The Research and Project Portal (RAPP) is a place where students can post both their completed and ongoing projects with all the resources and tools used. This portal allows students to see what other students have done in the past, in the same university environment, related to their domain of interest. Computer science instructors or students selecting projects can use this portal to assign or choose an incomplete project. Additionally, this portal allows non-computer science faculty and industry collaborators to document their project ideas for students in courses to prototype directly, rather than directly soliciting the help of instructors in engaging students. RAPP serves as a platform linking students across classes and faculty both in and out of computer science courses on joint projects to encourage long-term project efforts across semesters or years.

Keywords: education, technology, research, academic portal

Procedia PDF Downloads 126
4389 Developing a Modified Version of KIVA-3V, Enabling Gaseous Injections

Authors: Hossein Keshtkar, Ali Nasiri Toosi

Abstract:

With the growing concerns about gasoline environmental pollution and also the need for a more widely available fuel source, natural gas is finding its way to the automotive engines. But before this could happen industrially, simulations of natural gas direct injection need to take place to maximize and optimize power output. KIVA is one of the most powerful tools when it comes to engine simulation. Widely accepted by both researchers and the industry, KIVA an open-source code, offers great in-depth simulation and analyzation. KIVA can compute complex phenomena’s which can occur inside the chamber before, whilst and after ignition. One downside to KIVA, is its in-capability of simulating gaseous injections, making it useful for only liquidized fuel. In this study, we developed a numerical code, to enable the simulation of gaseous injection within the KIVA code. By introducing our code as a subroutine, we modified the original KIVA program. To ensure the correct application of gaseous fuel injection using our modified KIVA code, we simulated two different cases and compared them with their experimental data. We concluded our modified version of KIVA’s simulation results came in very close to those measured experimentally.

Keywords: gaseous injections, KIVA, natural gas direct injection, numerical code, simulation

Procedia PDF Downloads 271
4388 Improving the Competency of Undergraduate Nursing Students in Addressing a Timely Public Health Issue

Authors: Tsu-Yin Wu, Jenni Hoffman, Lydia McMurrows, Sarah Lally

Abstract:

Recent events of the Flint Water Crisis and elevated lead levels in Detroit public school water have highlighted a specific public health disparity and shown the need for better education of healthcare providers on lead education. Identifying children and pregnant women with a high risk for lead poisoning and ensuring lead testing is completed is critical. The purpose of this study is to explore the impact of an educational intervention on knowledge and confidence levels among nursing students enrolled in the prelicensure Bachelor of Science in Nursing (BSN) and Registered Nurse to BSN program (R2B). The study used both quantitative and qualitative research methods to assess the impact of multi-modal pedagogy on knowledge and confidence of lead screening and prevention among prelicensure and R2B nursing students. The students received lead poisoning and prevention content in addition to completing an e-learning module developed by the Pediatric Environmental Health Specialty Units. A total of 115 students completed the pre-and post-test instrument that consisted of demographic, lead knowledge, and confidence items. Despite the increase of total knowledge, three dimensions of lead poisoning, and confidence from pre- to post-test scores for both groups, there was no statistical significance on the increase between prelicensure and R2B students. Thematic analysis of qualitative data showed five themes from participants' learning experiences: lead exposure, signs and symptoms of lead poisoning, screening and diagnosis, prevention, and policy and statewide issues. The study is limited by a small sample and participants recalling some correct answers from the pretest, thus, scoring higher on the post-test. The results contribute to the minimally existent literature examining a critical public health concern regarding lead health exposure and prevention education of nursing students. Incorporating such content area into the nursing curriculum is essential in ensuring that such public health disparities are mitigated.

Keywords: lead poisoning, emerging public health issue, community health, nursing edducation

Procedia PDF Downloads 184
4387 Optimized Dynamic Bayesian Networks and Neural Verifier Test Applied to On-Line Isolated Characters Recognition

Authors: Redouane Tlemsani, Redouane, Belkacem Kouninef, Abdelkader Benyettou

Abstract:

In this paper, our system is a Markovien system which we can see it like a Dynamic Bayesian Networks. One of the major interests of these systems resides in the complete training of the models (topology and parameters) starting from training data. The Bayesian Networks are representing models of dubious knowledge on complex phenomena. They are a union between the theory of probability and the graph theory in order to give effective tools to represent a joined probability distribution on a set of random variables. The representation of knowledge bases on description, by graphs, relations of causality existing between the variables defining the field of study. The theory of Dynamic Bayesian Networks is a generalization of the Bayesians networks to the dynamic processes. Our objective amounts finding the better structure which represents the relationships (dependencies) between the variables of a dynamic bayesian network. In applications in pattern recognition, one will carry out the fixing of the structure which obliges us to admit some strong assumptions (for example independence between some variables).

Keywords: Arabic on line character recognition, dynamic Bayesian network, pattern recognition, networks

Procedia PDF Downloads 601
4386 SPARK: An Open-Source Knowledge Discovery Platform That Leverages Non-Relational Databases and Massively Parallel Computational Power for Heterogeneous Genomic Datasets

Authors: Thilina Ranaweera, Enes Makalic, John L. Hopper, Adrian Bickerstaffe

Abstract:

Data are the primary asset of biomedical researchers, and the engine for both discovery and research translation. As the volume and complexity of research datasets increase, especially with new technologies such as large single nucleotide polymorphism (SNP) chips, so too does the requirement for software to manage, process and analyze the data. Researchers often need to execute complicated queries and conduct complex analyzes of large-scale datasets. Existing tools to analyze such data, and other types of high-dimensional data, unfortunately suffer from one or more major problems. They typically require a high level of computing expertise, are too simplistic (i.e., do not fit realistic models that allow for complex interactions), are limited by computing power, do not exploit the computing power of large-scale parallel architectures (e.g. supercomputers, GPU clusters etc.), or are limited in the types of analysis available, compounded by the fact that integrating new analysis methods is not straightforward. Solutions to these problems, such as those developed and implemented on parallel architectures, are currently available to only a relatively small portion of medical researchers with access and know-how. The past decade has seen a rapid expansion of data management systems for the medical domain. Much attention has been given to systems that manage phenotype datasets generated by medical studies. The introduction of heterogeneous genomic data for research subjects that reside in these systems has highlighted the need for substantial improvements in software architecture. To address this problem, we have developed SPARK, an enabling and translational system for medical research, leveraging existing high performance computing resources, and analysis techniques currently available or being developed. It builds these into The Ark, an open-source web-based system designed to manage medical data. SPARK provides a next-generation biomedical data management solution that is based upon a novel Micro-Service architecture and Big Data technologies. The system serves to demonstrate the applicability of Micro-Service architectures for the development of high performance computing applications. When applied to high-dimensional medical datasets such as genomic data, relational data management approaches with normalized data structures suffer from unfeasibly high execution times for basic operations such as insert (i.e. importing a GWAS dataset) and the queries that are typical of the genomics research domain. SPARK resolves these problems by incorporating non-relational NoSQL databases that have been driven by the emergence of Big Data. SPARK provides researchers across the world with user-friendly access to state-of-the-art data management and analysis tools while eliminating the need for high-level informatics and programming skills. The system will benefit health and medical research by eliminating the burden of large-scale data management, querying, cleaning, and analysis. SPARK represents a major advancement in genome research technologies, vastly reducing the burden of working with genomic datasets, and enabling cutting edge analysis approaches that have previously been out of reach for many medical researchers.

Keywords: biomedical research, genomics, information systems, software

Procedia PDF Downloads 256
4385 New Possibilities for Testing UX and UI Design on Mobile Devices

Authors: Jakub Berčík, Anna Mravcová, Jana Gálová, Katarína Neomániová

Abstract:

In an era when everything is increasingly digital, consumers are always looking for new options in solutions to their everyday needs. In this context, mobile apps are developing at an exponential pace. One of the fastest growing segments of mobile technologies is, obviously, e-commerce. It can be predicted that mobile commerce will record nearly three times the global growth of e-commerce across all platforms, which indicates its importance in the given segment. The current coronavirus pandemic is also changing many of the existing paradigms both socially, economically, and technologically, which has a major impact on changing consumer behaviour and the emphasis on simplification and clarity of mobile solutions. This is the area that user experience (UX) and user interface (UI) designers deal with. Their task is to design a sufficiently attractive and interesting solution that will be available on all mobile devices and at the same time will be easy enough for the customer/visitor to get to the destination or to get the necessary information in a few clicks. The basis for changes in UX design can now be obtained not only through online analytical tools but also through neuromarketing, especially in the case of mobile devices. The paper highlights new possibilities for testing UX design applications on mobile devices using a special platform that combines a stationary eye camera (eye tracking) and facial analysis (facial coding).

Keywords: emotions, mobile design, user experience, visual attention

Procedia PDF Downloads 116
4384 Maturity Model for Agro-Industrial Logistics

Authors: Erika Tatiana Ruiz, Wilson Adarme Jaimes

Abstract:

This abstract presents the methodology for improving the logistics processes of agricultural production units belonging to the coffee, cocoa, and fruit sectors, starting from the fundamental concepts and detailing each of the phases to carry out the diagnosis, which will be the basis for the formulation of its action plan and implementation of the maturity model. As a result of this work, the maturity model is formulated to improve logistics processes. This model seeks to: generate a progressive model that is useful for all productive units belonging to these sectors at the national level, regardless of their initial conditions, focus on the improvement of logistics processes as a strategy that contributes to improving the competitiveness of the agricultural sector in Colombia and spread the implementation of good logistics practices in postharvest in all departments of the country through autonomous tools. This model has been built through a series of steps that allow the evaluation and improvement of the logistics dimensions or indicators. The potential improvements for each dimension provide the foundation on which to advance to the next level. Within the maturity model, a methodology is indicated for the design and execution of strategies to improve its logistics processes, taking into account the current state of each production unit.

Keywords: agroindustrial, characterization, logistics, maturity model, processes

Procedia PDF Downloads 126
4383 Transmission Design That Eliminates Gradual System Problems in Gearboxes

Authors: Ömer Ateş, Atilla Savaş

Abstract:

Reducers and transmission systems are power and speed transfer tools that have been used for many years in the technology world and in all engineering fields. Since today's transmissions have a threaded tap system, torque interruption occurs during tap change. besides, breakdown and manufacturing costs are high. Another problem is the limited torque and rpm setting in stepped gearbox systems. In this study, a new type of transmission system is designed to solve these problems. This new type of transmission system has been called the Continuously Variable Pulley. The most important feature of the transmission system in the study is that it can be adjusted Revolutions Per Minute-wise and torque-wise at the millimeter (precision) adjustment level. In order to make adjustments at this level, an adjustable pulley with the help of hydraulic piston is designed. The efficiency of the designed transmission system is 97 percent, the efficiency of today's transmissions is in the range of 85-95 percent. examined at the analysis and calculations, it is seen that the designed system gives realistic results and can be compared with today's transmissions and reducers. Therefore, this new type of transmission has been proven to be usable in production areas and the world of technology.

Keywords: gearbox, reducer, transmission, torque

Procedia PDF Downloads 113
4382 Using the Smith-Waterman Algorithm to Extract Features in the Classification of Obesity Status

Authors: Rosa Figueroa, Christopher Flores

Abstract:

Text categorization is the problem of assigning a new document to a set of predetermined categories, on the basis of a training set of free-text data that contains documents whose category membership is known. To train a classification model, it is necessary to extract characteristics in the form of tokens that facilitate the learning and classification process. In text categorization, the feature extraction process involves the use of word sequences also known as N-grams. In general, it is expected that documents belonging to the same category share similar features. The Smith-Waterman (SW) algorithm is a dynamic programming algorithm that performs a local sequence alignment in order to determine similar regions between two strings or protein sequences. This work explores the use of SW algorithm as an alternative to feature extraction in text categorization. The dataset used for this purpose, contains 2,610 annotated documents with the classes Obese/Non-Obese. This dataset was represented in a matrix form using the Bag of Word approach. The score selected to represent the occurrence of the tokens in each document was the term frequency-inverse document frequency (TF-IDF). In order to extract features for classification, four experiments were conducted: the first experiment used SW to extract features, the second one used unigrams (single word), the third one used bigrams (two word sequence) and the last experiment used a combination of unigrams and bigrams to extract features for classification. To test the effectiveness of the extracted feature set for the four experiments, a Support Vector Machine (SVM) classifier was tuned using 20% of the dataset. The remaining 80% of the dataset together with 5-Fold Cross Validation were used to evaluate and compare the performance of the four experiments of feature extraction. Results from the tuning process suggest that SW performs better than the N-gram based feature extraction. These results were confirmed by using the remaining 80% of the dataset, where SW performed the best (accuracy = 97.10%, weighted average F-measure = 97.07%). The second best was obtained by the combination of unigrams-bigrams (accuracy = 96.04, weighted average F-measure = 95.97) closely followed by the bigrams (accuracy = 94.56%, weighted average F-measure = 94.46%) and finally unigrams (accuracy = 92.96%, weighted average F-measure = 92.90%).

Keywords: comorbidities, machine learning, obesity, Smith-Waterman algorithm

Procedia PDF Downloads 285
4381 Path Integrals and Effective Field Theory of Large Scale Structure

Authors: Revant Nayar

Abstract:

In this work, we recast the equations describing large scale structure, and by extension all nonlinear fluids, in the path integral formalism. We first calculate the well known two and three point functions using Schwinger Keldysh formalism used commonly to perturbatively solve path integrals in non- equilibrium systems. Then we include EFT corrections due to pressure, viscosity, and noise as effects on the time-dependent propagator. We are able to express results for arbitrary two and three point correlation functions in LSS in terms of differential operators acting on a triple K master intergral. We also, for the first time, get analytical results for more general initial conditions deviating from the usual power law P∝kⁿ by introducing a mass scale in the initial conditions. This robust field theoretic formalism empowers us with tools from strongly coupled QFT to study the strongly non-linear regime of LSS and turbulent fluid dynamics such as OPE and holographic duals. These could be used to capture fully the strongly non-linear dynamics of fluids and move towards solving the open problem of classical turbulence.

Keywords: quantum field theory, cosmology, effective field theory, renormallisation

Procedia PDF Downloads 124
4380 Data Mining Meets Educational Analysis: Opportunities and Challenges for Research

Authors: Carla Silva

Abstract:

Recent development of information and communication technology enables us to acquire, collect, analyse data in various fields of socioeconomic – technological systems. Along with the increase of economic globalization and the evolution of information technology, data mining has become an important approach for economic data analysis. As a result, there has been a critical need for automated approaches to effective and efficient usage of massive amount of educational data, in order to support institutions to a strategic planning and investment decision-making. In this article, we will address data from several different perspectives and define the applied data to sciences. Many believe that 'big data' will transform business, government, and other aspects of the economy. We discuss how new data may impact educational policy and educational research. Large scale administrative data sets and proprietary private sector data can greatly improve the way we measure, track, and describe educational activity and educational impact. We also consider whether the big data predictive modeling tools that have emerged in statistics and computer science may prove useful in educational and furthermore in economics. Finally, we highlight a number of challenges and opportunities for future research.

Keywords: data mining, research analysis, investment decision-making, educational research

Procedia PDF Downloads 343
4379 Urban Ecological Interaction: Air, Water, Light and New Transit at the Human Scale of Barcelona’s Superilles

Authors: Philip Speranza

Abstract:

As everyday transit options are shifting from autocentric to pedestrian and bicycle oriented modes for healthy living, downtown streets are becoming more attractive places to live. However, tools and methods to measure the natural environment at the small scale of streets do not exist. Fortunately, a combination of mobile data collection technology and parametric urban design software now allows an interface to relate urban ecological conditions. This paper describes creation of an interactive tool to measure urban phenomena of air, water, and heat/light at the scale of new three-by-three block pedestrianized areas in Barcelona called Superilles. Each Superilla limits transit to the exterior of the blocks and to create more walkable and bikeable interior streets for healthy living. The research will describe the integration of data collection, analysis, and design output via a live interface using parametric software Rhino Grasshopper and the Human User Interface (UI) plugin.

Keywords: transit, urban design, GIS, parametric design, Superilles, Barcelona, urban ecology

Procedia PDF Downloads 235
4378 Data and Model-based Metamodels for Prediction of Performance of Extended Hollo-Bolt Connections

Authors: M. Cabrera, W. Tizani, J. Ninic, F. Wang

Abstract:

Open section beam to concrete-filled tubular column structures has been increasingly utilized in construction over the past few decades due to their enhanced structural performance, as well as economic and architectural advantages. However, the use of this configuration in construction is limited due to the difficulties in connecting the structural members as there is no access to the inner part of the tube to install standard bolts. Blind-bolted systems are a relatively new approach to overcome this limitation as they only require access to one side of the tubular section to tighten the bolt. The performance of these connections in concrete-filled steel tubular sections remains uncharacterized due to the complex interactions between concrete, bolt, and steel section. Over the last years, research in structural performance has moved to a more sophisticated and efficient approach consisting of machine learning algorithms to generate metamodels. This method reduces the need for developing complex, and computationally expensive finite element models, optimizing the search for desirable design variables. Metamodels generated by a data fusion approach use numerical and experimental results by combining multiple models to capture the dependency between the simulation design variables and connection performance, learning the relations between different design parameters and predicting a given output. Fully characterizing this connection will transform high-rise and multistorey construction by means of the introduction of design guidance for moment-resisting blind-bolted connections, which is currently unavailable. This paper presents a review of the steps taken to develop metamodels generated by means of artificial neural network algorithms which predict the connection stress and stiffness based on the design parameters when using Extended Hollo-Bolt blind bolts. It also provides consideration of the failure modes and mechanisms that contribute to the deformability as well as the feasibility of achieving blind-bolted rigid connections when using the blind fastener.

Keywords: blind-bolted connections, concrete-filled tubular structures, finite element analysis, metamodeling

Procedia PDF Downloads 148
4377 Recognizing Human Actions by Multi-Layer Growing Grid Architecture

Authors: Z. Gharaee

Abstract:

Recognizing actions performed by others is important in our daily lives since it is necessary for communicating with others in a proper way. We perceive an action by observing the kinematics of motions involved in the performance. We use our experience and concepts to make a correct recognition of the actions. Although building the action concepts is a life-long process, which is repeated throughout life, we are very efficient in applying our learned concepts in analyzing motions and recognizing actions. Experiments on the subjects observing the actions performed by an actor show that an action is recognized after only about two hundred milliseconds of observation. In this study, hierarchical action recognition architecture is proposed by using growing grid layers. The first-layer growing grid receives the pre-processed data of consecutive 3D postures of joint positions and applies some heuristics during the growth phase to allocate areas of the map by inserting new neurons. As a result of training the first-layer growing grid, action pattern vectors are generated by connecting the elicited activations of the learned map. The ordered vector representation layer receives action pattern vectors to create time-invariant vectors of key elicited activations. Time-invariant vectors are sent to second-layer growing grid for categorization. This grid creates the clusters representing the actions. Finally, one-layer neural network developed by a delta rule labels the action categories in the last layer. System performance has been evaluated in an experiment with the publicly available MSR-Action3D dataset. There are actions performed by using different parts of human body: Hand Clap, Two Hands Wave, Side Boxing, Bend, Forward Kick, Side Kick, Jogging, Tennis Serve, Golf Swing, Pick Up and Throw. The growing grid architecture was trained by applying several random selections of generalization test data fed to the system during on average 100 epochs for each training of the first-layer growing grid and around 75 epochs for each training of the second-layer growing grid. The average generalization test accuracy is 92.6%. A comparison analysis between the performance of growing grid architecture and self-organizing map (SOM) architecture in terms of accuracy and learning speed show that the growing grid architecture is superior to the SOM architecture in action recognition task. The SOM architecture completes learning the same dataset of actions in around 150 epochs for each training of the first-layer SOM while it takes 1200 epochs for each training of the second-layer SOM and it achieves the average recognition accuracy of 90% for generalization test data. In summary, using the growing grid network preserves the fundamental features of SOMs, such as topographic organization of neurons, lateral interactions, the abilities of unsupervised learning and representing high dimensional input space in the lower dimensional maps. The architecture also benefits from an automatic size setting mechanism resulting in higher flexibility and robustness. Moreover, by utilizing growing grids the system automatically obtains a prior knowledge of input space during the growth phase and applies this information to expand the map by inserting new neurons wherever there is high representational demand.

Keywords: action recognition, growing grid, hierarchical architecture, neural networks, system performance

Procedia PDF Downloads 149
4376 Crack Growth Life Prediction of a Fighter Aircraft Wing Splice Joint Under Spectrum Loading Using Random Forest Regression and Artificial Neural Networks with Hyperparameter Optimization

Authors: Zafer Yüce, Paşa Yayla, Alev Taşkın

Abstract:

There are heaps of analytical methods to estimate the crack growth life of a component. Soft computing methods have an increasing trend in predicting fatigue life. Their ability to build complex relationships and capability to handle huge amounts of data are motivating researchers and industry professionals to employ them for challenging problems. This study focuses on soft computing methods, especially random forest regressors and artificial neural networks with hyperparameter optimization algorithms such as grid search and random grid search, to estimate the crack growth life of an aircraft wing splice joint under variable amplitude loading. TensorFlow and Scikit-learn libraries of Python are used to build the machine learning models for this study. The material considered in this work is 7050-T7451 aluminum, which is commonly preferred as a structural element in the aerospace industry, and regarding the crack type; corner crack is used. A finite element model is built for the joint to calculate fastener loads and stresses on the structure. Since finite element model results are validated with analytical calculations, findings of the finite element model are fed to AFGROW software to calculate analytical crack growth lives. Based on Fighter Aircraft Loading Standard for Fatigue (FALSTAFF), 90 unique fatigue loading spectra are developed for various load levels, and then, these spectrums are utilized as inputs to the artificial neural network and random forest regression models for predicting crack growth life. Finally, the crack growth life predictions of the machine learning models are compared with analytical calculations. According to the findings, a good correlation is observed between analytical and predicted crack growth lives.

Keywords: aircraft, fatigue, joint, life, optimization, prediction.

Procedia PDF Downloads 155
4375 Critical Path Segments Method for Scheduling Technique

Authors: Sherif M. Hafez, Remon F. Aziz, May S. A. Elalim

Abstract:

Project managers today rely on scheduling tools based on the Critical Path Method (CPM) to determine the overall project duration and the activities’ float times which lead to greater efficiency in planning and control of projects. CPM was useful for scheduling construction projects, but researchers had highlighted a number of serious drawbacks that limit its use as a decision support tool and lacks the ability to clearly record and represent detailed information. This paper discusses the drawbacks of CPM as a scheduling technique and presents a modified critical path method (CPM) model which is called critical path segments (CPS). The CPS scheduling mechanism addresses the problems of CPM in three ways: decomposing the activity duration of separated but connected time segments; all relationships among activities are converted into finish–to–start relationship; and analysis and calculations are made with forward path. Sample cases are included to illustrate the shortages in CPM, CPS full analysis and calculations are explained in details, and how schedules can be handled better with the CPS technique.

Keywords: construction management, scheduling, critical path method, critical path segments, forward pass, float, project control

Procedia PDF Downloads 343
4374 Data Mining As A Tool For Knowledge Management: A Review

Authors: Maram Saleh

Abstract:

Knowledge has become an essential resource in today’s economy and become the most important asset of maintaining competition advantage in organizations. The importance of knowledge has made organizations to manage their knowledge assets and resources through all multiple knowledge management stages such as: Knowledge Creation, knowledge storage, knowledge sharing and knowledge use. Researches on data mining are continues growing over recent years on both business and educational fields. Data mining is one of the most important steps of the knowledge discovery in databases process aiming to extract implicit, unknown but useful knowledge and it is considered as significant subfield in knowledge management. Data miming have the great potential to help organizations to focus on extracting the most important information on their data warehouses. Data mining tools and techniques can predict future trends and behaviors, allowing businesses to make proactive, knowledge-driven decisions. This review paper explores the applications of data mining techniques in supporting knowledge management process as an effective knowledge discovery technique. In this paper, we identify the relationship between data mining and knowledge management, and then focus on introducing some application of date mining techniques in knowledge management for some real life domains.

Keywords: Data Mining, Knowledge management, Knowledge discovery, Knowledge creation.

Procedia PDF Downloads 193
4373 Indexing and Incremental Approach Using Map Reduce Bipartite Graph (MRBG) for Mining Evolving Big Data

Authors: Adarsh Shroff

Abstract:

Big data is a collection of dataset so large and complex that it becomes difficult to process using data base management tools. To perform operations like search, analysis, visualization on big data by using data mining; which is the process of extraction of patterns or knowledge from large data set. In recent years, the data mining applications become stale and obsolete over time. Incremental processing is a promising approach to refreshing mining results. It utilizes previously saved states to avoid the expense of re-computation from scratch. This project uses i2MapReduce, an incremental processing extension to Map Reduce, the most widely used framework for mining big data. I2MapReduce performs key-value pair level incremental processing rather than task level re-computation, supports not only one-step computation but also more sophisticated iterative computation, which is widely used in data mining applications, and incorporates a set of novel techniques to reduce I/O overhead for accessing preserved fine-grain computation states. To optimize the mining results, evaluate i2MapReduce using a one-step algorithm and three iterative algorithms with diverse computation characteristics for efficient mining.

Keywords: big data, map reduce, incremental processing, iterative computation

Procedia PDF Downloads 334
4372 An Investigation into the Views of Gifted Children on the Effects of Computer and Information Technologies on Their Lives and Education

Authors: Ahmet Kurnaz, Eyup Yurt, Ümit Çiftci

Abstract:

In this study, too, an attempt was made to reveal the place and effects of information technologies on the lives and education of gifted children based on the views of gifted. To this end, the effects of information technologies on gifted are general skills, technology use, academic and social skills, and cooperative and personal skills were investigated. These skills were explored depending on whether or not gifted had their own computers, had internet connection at home, or how often they use the internet, average time period they spent at the computer, how often they played computer games and their use of social media. The study was conducted using the screening model with a quantitative approach. The sample of the study consisted of 129 gifted attending 5-12th classes in 12 provinces in different regions of Turkey. 64 of the participants were female while 65 were male. The research data were collected using the using computer of gifted and information technologies (UCIT) questionnaire which was developed by the researchers and given its final form after receiving expert view. As a result of the study, it was found that UCIT use improved foreign language speaking skills of gifted, enabled them to get to know and understand different cultures, and made use of computer and information technologies while they study. At the end of the study these result were obtained: Gifted have positive idea using computer and communication technology. There are differences whether using the internet about the ideas UCIT. But there are not differences whether having computer, inhabited city, grade level, having internet at home, daily and weekly internet usage durations, playing the computer and internet game, having Facebook and Twitter account about the UCIT. UCIT contribute to the development of gifted vocabulary, allows knowing and understand different cultures, developing foreign language speaking skills, gifted do not give up computer when they do their homework, improve their reading, listening, understanding and writing skills in a foreign language. Gifted children want to have transition to the use of tablets in education. They think UCIT facilitates doing their homework, contributes learning more information in a shorter time. They'd like to use computer-assisted instruction programs at courses. They think they will be more successful in the future if their computer skills are good. But gifted students prefer teacher instead of teaching with computers and they said that learning can be run from home without going to school.

Keywords: gifted, using computer, communication technology, information technologies

Procedia PDF Downloads 380
4371 Optimization of Bills Assignment to Different Skill-Levels of Data Entry Operators in a Business Process Outsourcing Industry

Authors: M. S. Maglasang, S. O. Palacio, L. P. Ogdoc

Abstract:

Business Process Outsourcing has been one of the fastest growing and emerging industry in the Philippines today. Unlike most of the contact service centers, more popularly known as "call centers", The BPO Industry’s primary outsourced service is performing audits of the global clients' logistics. As a service industry, manpower is considered as the most important yet the most expensive resource in the company. Because of this, there is a need to maximize the human resources so people are effectively and efficiently utilized. The main purpose of the study is to optimize the current manpower resources through effective distribution and assignment of different types of bills to the different skill-level of data entry operators. The assignment model parameters include the average observed time matrix gathered from through time study, which incorporates the learning curve concept. Subsequently, a simulation model was made to duplicate the arrival rate of demand which includes the different batches and types of bill per day. Next, a mathematical linear programming model was formulated. Its objective is to minimize direct labor cost per bill by allocating the different types of bills to the different skill-levels of operators. Finally, a hypothesis test was done to validate the model, comparing the actual and simulated results. The analysis of results revealed that the there’s low utilization of effective capacity because of its failure to determine the product-mix, skill-mix, and simulated demand as model parameters. Moreover, failure to consider the effects of learning curve leads to overestimation of labor needs. From 107 current number of operators, the proposed model gives a result of 79 operators. This results to an increase of utilization of effective capacity to 14.94%. It is recommended that the excess 28 operators would be reallocated to the other areas of the department. Finally, a manpower capacity planning model is also recommended in support to management’s decisions on what to do when the current capacity would reach its limit with the expected increasing demand.

Keywords: optimization modelling, linear programming, simulation, time and motion study, capacity planning

Procedia PDF Downloads 501
4370 The Restoration of the Old District in the Urbanization: The Case Study of Samsen Riverside Community, Dusit District, Bangkok

Authors: Tikhanporn Punluekdej, Saowapa Phaithayawat

Abstract:

The objectives of this research are: 1) to discover the mechanism in the restoration process of the old district, and 2) to study the people participation in the community with related units. This research utilizes qualitative research method together with the tools used in the study of historical and anthropological disciplines. The research revealed that the restoration process of the old district started with the needs of the local people in the community. These people are considered as a young generation in the community. The leading group of the community played a vital role in the restoration process by igniting the whole idea and followed by the help from those who have lived in the area of more than fifty years. The restoration process is the genuine desire of the local people without the intervention of the local politics. The core group would coordinate with the related units in which there were, for instance, the academic institutions in order to find out the most dominant historical features of the community including its settlement. The Crown Property Bureau, as the sole-owner of the land, joined the restoration in the physical development dimension. The restoration was possible due to the cooperation between local people and related units, under the designated plans, budget, and social activities.

Keywords: restoration, urban area, old district, people participation

Procedia PDF Downloads 397