Search results for: software component and interfaces
6832 Continuous Improvement of Teaching Quality through Course Evaluation by the Students
Authors: Valerie Follonier, Henrike Hamelmann, Jean-Michel Jullien
Abstract:
The Distance Learning University in Switzerland (UniDistance) is offering bachelor and master courses as well as further education programs. The professors and their assistants work at traditional Swiss universities and are giving their courses at UniDistance following a blended learning and flipped classroom approach. A standardized course evaluation by the students has been established as a component of a quality improvement process. The students’ feedback enables the stakeholders to identify areas of improvement, initiate professional development for the teaching teams and thus continuously augment the quality of instruction. This paper describes the evaluation process, the tools involved and how the approach involving all stakeholders helps forming a culture of quality in teaching. Additionally, it will present the first evaluation results following the new process. Two software tools have been developed to support all stakeholders in the process of the semi-annual formative evaluation. The first tool allows to create the survey and to assign it to the relevant courses and students. The second tool presents the results of the evaluation to the stakeholders, providing specific features for the teaching teams, the dean, the directorate and EDUDL+ (Educational development unit distance learning). The survey items were selected in accordance with the e-learning strategy of the institution and are formulated to support the professional development of the teaching teams. By reviewing the results the teaching teams become aware of the opinion of the students and are asked to write a feedback for the attention of their dean. The dean reviews the results of the faculty and writes a general report about the situation of the faculty and the possible improvements intended. Finally, EDUDL+ writes a final report summarising the evaluation results. A mechanism of adjustable warnings allows it to generate quality indicators for each module. These are summarised for each faculty and globally for the whole institution in order to increase the vigilance of the responsible. The quality process involves changing the indicators regularly to focus on different areas each semester, to facilitate the professional development of the teaching teams and to progressively augment the overall teaching quality of the institution.Keywords: continuous improvement process, course evaluation, distance learning, software tools, teaching quality
Procedia PDF Downloads 2596831 Disability and Quality of Life in Low Back Pain: A Cross-Sectional Study
Authors: Zarina Zahari, Maria Justine, Kamaria Kamaruddin
Abstract:
Low back pain (LBP) is a major musculoskeletal problem in global population. This study aimed to examine the relationship between pain, disability and quality of life in patients with non-specific low back pain (LBP). One hundred LBP participants were recruited in this cross-sectional study (mean age = 42.23±11.34 years old). Pain was measured using Numerical Rating Scale (11-point). Disability was assessed using the revised Oswestry low back pain disability questionnaire (ODQ) and quality of life (QoL) was evaluated using the SF-36 v2. Majority of participants (58%) presented with moderate pain and 49% experienced severe disability. Thus, the pain and disability were found significant with negative correlation (r= -0.712, p<0.05). The pain and QoL also showed significant and positive correlation with both Physical Health Component Summary (PHCS) (r= .840, p<0.05) and Mental Health Component Summary (MHCS) (r= 0.446, p<0.05). Regression analysis indicated that pain emerged as an indicator of both disability and QoL (PHCS and MHCS) accounting for 51%, 71% and 21% of the variances respectively. This indicates that pain is an important factor in predicting disability and QoL in LBP sufferers.Keywords: disability, low back pain, pain, quality of life
Procedia PDF Downloads 5356830 Software Development for Both Small Wind Performance Optimization and Structural Compliance Analysis with International Safety Regulations
Authors: K. M. Yoo, M. H. Kang
Abstract:
Conventional commercial wind turbine design software is limited to large wind turbines due to not incorporating with low Reynold’s Number aerodynamic characteristics typically for small wind turbines. To extract maximum annual energy product from an intermediately designed small wind turbine associated with measured wind data, numerous simulation is highly recommended to have a best fitting planform design with proper airfoil configuration. Since depending upon wind distribution with average wind speed, an optimal wind turbine planform design changes accordingly. It is theoretically not difficult, though, it is very inconveniently time-consuming design procedure to finalize conceptual layout of a desired small wind turbine. Thus, to help simulations easier and faster, a GUI software is developed to conveniently iterate and change airfoil types, wind data, and geometric blade data as well. With magnetic generator torque curve, peak power tracking simulation is also available to better match with the magnetic generator. Small wind turbine often lacks starting torque due to blade optimization. Thus this simulation is also embedded along with yaw design. This software provides various blade cross section details at user’s design convenience such as skin thickness control with fiber direction option, spar shape, and their material properties. Since small wind turbine is under international safety regulations with fatigue damage during normal operations and safety load analyses with ultimate excessive loads, load analyses are provided with each category mandated in the safety regulations.Keywords: GUI software, Low Reynold’s number aerodynamics, peak power tracking, safety regulations, wind turbine performance optimization
Procedia PDF Downloads 3066829 Don't Just Guess and Slip: Estimating Bayesian Knowledge Tracing Parameters When Observations Are Scant
Authors: Michael Smalenberger
Abstract:
Intelligent tutoring systems (ITS) are computer-based platforms which can incorporate artificial intelligence to provide step-by-step guidance as students practice problem-solving skills. ITS can replicate and even exceed some benefits of one-on-one tutoring, foster transactivity in collaborative environments, and lead to substantial learning gains when used to supplement the instruction of a teacher or when used as the sole method of instruction. A common facet of many ITS is their use of Bayesian Knowledge Tracing (BKT) to estimate parameters necessary for the implementation of the artificial intelligence component, and for the probability of mastery of a knowledge component relevant to the ITS. While various techniques exist to estimate these parameters and probability of mastery, none directly and reliably ask the user to self-assess these. In this study, 111 undergraduate students used an ITS in a college-level introductory statistics course for which detailed transaction-level observations were recorded, and users were also routinely asked direct questions that would lead to such a self-assessment. Comparisons were made between these self-assessed values and those obtained using commonly used estimation techniques. Our findings show that such self-assessments are particularly relevant at the early stages of ITS usage while transaction level data are scant. Once a user’s transaction level data become available after sufficient ITS usage, these can replace the self-assessments in order to eliminate the identifiability problem in BKT. We discuss how these findings are relevant to the number of exercises necessary to lead to mastery of a knowledge component, the associated implications on learning curves, and its relevance to instruction time.Keywords: Bayesian Knowledge Tracing, Intelligent Tutoring System, in vivo study, parameter estimation
Procedia PDF Downloads 1746828 Robust ResNets for Chemically Reacting Flows
Authors: Randy Price, Harbir Antil, Rainald Löhner, Fumiya Togashi
Abstract:
Chemically reacting flows are common in engineering applications such as hypersonic flow, combustion, explosions, manufacturing process, and environmental assessments. The number of reactions in combustion simulations can exceed 100, making a large number of flow and combustion problems beyond the capabilities of current supercomputers. Motivated by this, deep neural networks (DNNs) will be introduced with the goal of eventually replacing the existing chemistry software packages with DNNs. The DNNs used in this paper are motivated by the Residual Neural Network (ResNet) architecture. In the continuum limit, ResNets become an optimization problem constrained by an ODE. Such a feature allows the use of ODE control techniques to enhance the DNNs. In this work, DNNs are constructed, which update the species un at the nᵗʰ timestep to uⁿ⁺¹ at the n+1ᵗʰ timestep. Parallel DNNs are trained for each species, taking in uⁿ as input and outputting one component of uⁿ⁺¹. These DNNs are applied to multiple species and reactions common in chemically reacting flows such as H₂-O₂ reactions. Experimental results show that the DNNs are able to accurately replicate the dynamics in various situations and in the presence of errors.Keywords: chemical reacting flows, computational fluid dynamics, ODEs, residual neural networks, ResNets
Procedia PDF Downloads 1216827 Annotation Ontology for Semantic Web Development
Authors: Hadeel Al Obaidy, Amani Al Heela
Abstract:
The main purpose of this paper is to examine the concept of semantic web and the role that ontology and semantic annotation plays in the development of semantic web services. The paper focuses on semantic web infrastructure illustrating how ontology and annotation work to provide the learning capabilities for building content semantically. To improve productivity and quality of software, the paper applies approaches, notations and techniques offered by software engineering. It proposes a conceptual model to develop semantic web services for the infrastructure of web information retrieval system of digital libraries. The developed system uses ontology and annotation to build a knowledge based system to define and link the meaning of a web content to retrieve information for users’ queries. The results are more relevant through keywords and ontology rule expansion that will be more accurate to satisfy the requested information. The level of results accuracy would be enhanced since the query semantically analyzed work with the conceptual architecture of the proposed system.Keywords: semantic web services, software engineering, semantic library, knowledge representation, ontology
Procedia PDF Downloads 1746826 UWB Open Spectrum Access for a Smart Software Radio
Authors: Hemalatha Rallapalli, K. Lal Kishore
Abstract:
In comparison to systems that are typically designed to provide capabilities over a narrow frequency range through hardware elements, the next generation cognitive radios are intended to implement a broader range of capabilities through efficient spectrum exploitation. This offers the user the promise of greater flexibility, seamless roaming possible on different networks, countries, frequencies, etc. It requires true paradigm shift i.e., liberalization over a wide band of spectrum as well as a growth path to more and greater capability. This work contributes towards the design and implementation of an open spectrum access (OSA) feature to unlicensed users thus offering a frequency agile radio platform that is capable of performing spectrum sensing over a wideband. Thus, an ultra-wideband (UWB) radio, which has the intelligence of spectrum sensing only, unlike the cognitive radio with complete intelligence, is named as a Smart Software Radio (SSR). The spectrum sensing mechanism is implemented based on energy detection. Simulation results show the accuracy and validity of this method.Keywords: cognitive radio, energy detection, software radio, spectrum sensing
Procedia PDF Downloads 4296825 Review of the Software Used for 3D Volumetric Reconstruction of the Liver
Authors: P. Strakos, M. Jaros, T. Karasek, T. Kozubek, P. Vavra, T. Jonszta
Abstract:
In medical imaging, segmentation of different areas of human body like bones, organs, tissues, etc. is an important issue. Image segmentation allows isolating the object of interest for further processing that can lead for example to 3D model reconstruction of whole organs. Difficulty of this procedure varies from trivial for bones to quite difficult for organs like liver. The liver is being considered as one of the most difficult human body organ to segment. It is mainly for its complexity, shape versatility and proximity of other organs and tissues. Due to this facts usually substantial user effort has to be applied to obtain satisfactory results of the image segmentation. Process of image segmentation then deteriorates from automatic or semi-automatic to fairly manual one. In this paper, overview of selected available software applications that can handle semi-automatic image segmentation with further 3D volume reconstruction of human liver is presented. The applications are being evaluated based on the segmentation results of several consecutive DICOM images covering the abdominal area of the human body.Keywords: image segmentation, semi-automatic, software, 3D volumetric reconstruction
Procedia PDF Downloads 2916824 Bridging the Gap between Different Interfaces for Business Process Modeling
Authors: Katalina Grigorova, Kaloyan Mironov
Abstract:
The paper focuses on the benefits of business process modeling. Although this discipline is developing for many years, there is still necessity of creating new opportunities to meet the ever-increasing users’ needs. Because one of these needs is related to the conversion of business process models from one standard to another, the authors have developed a converter between BPMN and EPC standards using workflow patterns as intermediate tool. Nowadays there are too many systems for business process modeling. The variety of output formats is almost the same as the systems themselves. This diversity additionally hampers the conversion of the models. The presented study is aimed at discussing problems due to differences in the output formats of various modeling environments.Keywords: business process modeling, business process modeling standards, workflow patterns, converting models
Procedia PDF Downloads 5886823 Optimal Management of Internal Capital of Company
Authors: S. Sadallah
Abstract:
In this paper, dynamic programming is used to determine the optimal management of financial resources in company. Solution of the problem by consider into simpler substructures is constructed. The optimal management of internal capital of company are simulated. The tools applied in this development are based on graph theory. The software of given problems is built by using greedy algorithm. The obtained model and program maintenance enable us to define the optimal version of management of proper financial flows by using visual diagram on each level of investment.Keywords: management, software, optimal, greedy algorithm, graph-diagram
Procedia PDF Downloads 2856822 Human Kinetics Education and the Computer Operations, Effects and Merits
Authors: Kehinde Adeyeye Adelabu
Abstract:
Computer applications has completely revolutionized the way of life of people which does not exclude the field of sport education. There are computer technologies which help to enhance teaching in every field of education. Invention of computers has done great to the field of education. This study was therefore carried out to examine the effects and merits of computer operations in Human Kinetics Education and Sports. The study was able to identify the component of computer, uses of computer in Human Kinetics education (sports), computer applications in some branches of human kinetics education. A qualitative research method was employed by the author in gathering experts’ views and used to analyze the effects and merits of computer applications in the field of human kinetics education. No experiment was performed in the cause of carrying out the study. The source of information for the study was text-books, journal, articles, past project reports, internet i.e. Google search engine. Computer has significantly helped to improve Education (Human Kinetic), it has complemented the basic physical fitness testing and gave a more scientific basis to the testing. The use of the software and packages has made cost projections, database applications, inventory control, management of events, word processing, electronic mailing and record keeping easier than the pasts.Keywords: application, computer operation, education, human kinetics
Procedia PDF Downloads 1866821 Simple Ways to Enhance the Security of Web Services
Authors: Majid Azarniush, Soroush Mokallaei
Abstract:
Although robust security software, including anti-viruses, anti spy wares, anti-spam and firewalls, are amalgamated with new technologies such as Safe Zone, Hybrid Cloud, Sand Box etc., and it can be said that they have managed to prepare highest level of security against viruses, spy wares and other malwares in 2012, but in fact hackers' attacks to websites are increasingly becoming more and more complicated. Because of security matters and developments, it can be said that it was expected to happen so. Here in this work, we try to point out to some functional and vital notes to enhance security on the web enabling the user to browse safely in no limit web world and to use virtual space securely.Keywords: firewalls, security, web services, software
Procedia PDF Downloads 5146820 Interactive Garments: Flexible Technologies for Textile Integration
Authors: Anupam Bhatia
Abstract:
Upon reviewing the literature and the pragmatic work done in the field of E- textiles, it is observed that the applications of wearable technologies have found a steady growth in the field of military, medical, industrial, sports; whereas fashion is at a loss to know how to treat this technology and bring it to market. The purpose of this paper is to understand the practical issues of integration of electronics in garments; cutting patterns for mass production, maintaining the basic properties of textiles and daily maintenance of garments that hinder the wide adoption of interactive fabric technology within Fashion and leisure wear. To understand the practical hindrances an experimental and laboratory approach is taken. “Techno Meets Fashion” has been an interactive fashion project where sensor technologies have been embedded with textiles that result in set of ensembles that are light emitting garments, sound sensing garments, proximity garments, shape memory garments etc. Smart textiles, especially in the form of textile interfaces, are drastically underused in fashion and other lifestyle product design. Clothing and some other textile products must be washable, which subjects to the interactive elements to water and chemical immersion, physical stress, and extreme temperature. The current state of the art tends to be too fragile for this treatment. The process for mass producing traditional textiles becomes difficult in interactive textiles. As cutting patterns from larger rolls of cloth and sewing them together to make garments breaks and reforms electronic connections in an uncontrolled manner. Because of this, interactive fabric elements are integrated by hand into textiles produced by standard methods. The Arduino has surely made embedding electronics into textiles much easier than before; even then electronics are not integral to the daily wear garments. Soft and flexible interfaces of MEMS (micro sensors and Micro actuators) can be an option to make this possible by blending electronics within E-textiles in a way that’s seamless and still retains functions of the circuits as well as the garment. Smart clothes, which offer simultaneously a challenging design and utility value, can be only mass produced if the demands of the body are taken care of i.e. protection, anthropometry, ergonomics of human movement, thermo- physiological regulation.Keywords: ambient intelligence, proximity sensors, shape memory materials, sound sensing garments, wearable technology
Procedia PDF Downloads 3946819 Statistical Model of Water Quality in Estero El Macho, Machala-El Oro
Authors: Rafael Zhindon Almeida
Abstract:
Surface water quality is an important concern for the evaluation and prediction of water quality conditions. The objective of this study is to develop a statistical model that can accurately predict the water quality of the El Macho estuary in the city of Machala, El Oro province. The methodology employed in this study is of a basic type that involves a thorough search for theoretical foundations to improve the understanding of statistical modeling for water quality analysis. The research design is correlational, using a multivariate statistical model involving multiple linear regression and principal component analysis. The results indicate that water quality parameters such as fecal coliforms, biochemical oxygen demand, chemical oxygen demand, iron and dissolved oxygen exceed the allowable limits. The water of the El Macho estuary is determined to be below the required water quality criteria. The multiple linear regression model, based on chemical oxygen demand and total dissolved solids, explains 99.9% of the variance of the dependent variable. In addition, principal component analysis shows that the model has an explanatory power of 86.242%. The study successfully developed a statistical model to evaluate the water quality of the El Macho estuary. The estuary did not meet the water quality criteria, with several parameters exceeding the allowable limits. The multiple linear regression model and principal component analysis provide valuable information on the relationship between the various water quality parameters. The findings of the study emphasize the need for immediate action to improve the water quality of the El Macho estuary to ensure the preservation and protection of this valuable natural resource.Keywords: statistical modeling, water quality, multiple linear regression, principal components, statistical models
Procedia PDF Downloads 1006818 An Efficient Machine Learning Model to Detect Metastatic Cancer in Pathology Scans Using Principal Component Analysis Algorithm, Genetic Algorithm, and Classification Algorithms
Authors: Bliss Singhal
Abstract:
Machine learning (ML) is a branch of Artificial Intelligence (AI) where computers analyze data and find patterns in the data. The study focuses on the detection of metastatic cancer using ML. Metastatic cancer is the stage where cancer has spread to other parts of the body and is the cause of approximately 90% of cancer-related deaths. Normally, pathologists spend hours each day to manually classifying whether tumors are benign or malignant. This tedious task contributes to mislabeling metastasis being over 60% of the time and emphasizes the importance of being aware of human error and other inefficiencies. ML is a good candidate to improve the correct identification of metastatic cancer, saving thousands of lives and can also improve the speed and efficiency of the process, thereby taking fewer resources and time. So far, the deep learning methodology of AI has been used in research to detect cancer. This study is a novel approach to determining the potential of using preprocessing algorithms combined with classification algorithms in detecting metastatic cancer. The study used two preprocessing algorithms: principal component analysis (PCA) and the genetic algorithm, to reduce the dimensionality of the dataset and then used three classification algorithms: logistic regression, decision tree classifier, and k-nearest neighbors to detect metastatic cancer in the pathology scans. The highest accuracy of 71.14% was produced by the ML pipeline comprising of PCA, the genetic algorithm, and the k-nearest neighbor algorithm, suggesting that preprocessing and classification algorithms have great potential for detecting metastatic cancer.Keywords: breast cancer, principal component analysis, genetic algorithm, k-nearest neighbors, decision tree classifier, logistic regression
Procedia PDF Downloads 836817 Research Attitude: Its Factor Structure and Determinants in the Graduate Level
Authors: Janet Lynn S. Montemayor
Abstract:
Dropping survivability and rising drop-out rate in the graduate school is attributed to the demands that come along with research-related requirements. Graduate students tend to withdraw from their studies when confronted with such requirements. This act of succumbing to the challenge is primarily due to a negative mindset. An understanding of students’ view towards research is essential for teachers in facilitating research activities in the graduate school. This study aimed to develop a tool that accurately measures attitude towards research. Psychometric properties of the Research Attitude Inventory (RAIn) was assessed. A pool of items (k=50) was initially constructed and was administered to a development sample composed of Masters and Doctorate degree students (n=159). Results show that the RAIn is a reliable measure of research attitude (k=41, αmax = 0.894). Principal component analysis using orthogonal rotation with Kaiser normalization identified four underlying factors of research attitude, namely predisposition, purpose, perspective, and preparation. Research attitude among the respondents was analyzed using this measure.Keywords: graduate education, principal component analysis, research attitude, scale development
Procedia PDF Downloads 1916816 Multi-Label Approach to Facilitate Test Automation Based on Historical Data
Authors: Warda Khan, Remo Lachmann, Adarsh S. Garakahally
Abstract:
The increasing complexity of software and its applicability in a wide range of industries, e.g., automotive, call for enhanced quality assurance techniques. Test automation is one option to tackle the prevailing challenges by supporting test engineers with fast, parallel, and repetitive test executions. A high degree of test automation allows for a shift from mundane (manual) testing tasks to a more analytical assessment of the software under test. However, a high initial investment of test resources is required to establish test automation, which is, in most cases, a limitation to the time constraints provided for quality assurance of complex software systems. Hence, a computer-aided creation of automated test cases is crucial to increase the benefit of test automation. This paper proposes the application of machine learning for the generation of automated test cases. It is based on supervised learning to analyze test specifications and existing test implementations. The analysis facilitates the identification of patterns between test steps and their implementation with test automation components. For the test case generation, this approach exploits historical data of test automation projects. The identified patterns are the foundation to predict the implementation of unknown test case specifications. Based on this support, a test engineer solely has to review and parameterize the test automation components instead of writing them manually, resulting in a significant time reduction for establishing test automation. Compared to other generation approaches, this ML-based solution can handle different writing styles, authors, application domains, and even languages. Furthermore, test automation tools require expert knowledge by means of programming skills, whereas this approach only requires historical data to generate test cases. The proposed solution is evaluated using various multi-label evaluation criteria (EC) and two small-sized real-world systems. The most prominent EC is ‘Subset Accuracy’. The promising results show an accuracy of at least 86% for test cases, where a 1:1 relationship (Multi-Class) between test step specification and test automation component exists. For complex multi-label problems, i.e., one test step can be implemented by several components, the prediction accuracy is still at 60%. It is better than the current state-of-the-art results. It is expected the prediction quality to increase for larger systems with respective historical data. Consequently, this technique facilitates the time reduction for establishing test automation and is thereby independent of the application domain and project. As a work in progress, the next steps are to investigate incremental and active learning as additions to increase the usability of this approach, e.g., in case labelled historical data is scarce.Keywords: machine learning, multi-class, multi-label, supervised learning, test automation
Procedia PDF Downloads 1336815 The Laser Line Detection for Autonomous Mapping Based on Color Segmentation
Authors: Pavel Chmelar, Martin Dobrovolny
Abstract:
Laser projection or laser footprint detection is today widely used in many fields of robotics, measurement, or electronics. The system accuracy strictly depends on precise laser footprint detection on target objects. This article deals with the laser line detection based on the RGB segmentation and the component labeling. As a measurement device was used the developed optical rangefinder. The optical rangefinder is equipped with vertical sweeping of the laser beam and high quality camera. This system was developed mainly for automatic exploration and mapping of unknown spaces. In the first section is presented a new detection algorithm. In the second section are presented measurements results. The measurements were performed in variable light conditions in interiors. The last part of the article present achieved results and their differences between day and night measurements.Keywords: color segmentation, component labelling, laser line detection, automatic mapping, distance measurement, vector map
Procedia PDF Downloads 4336814 The Impact of Experiential Learning on the Success of Upper Division Mechanical Engineering Students
Authors: Seyedali Seyedkavoosi, Mohammad Obadat, Seantorrion Boyle
Abstract:
The purpose of this study is to assess the effectiveness of a nontraditional experiential learning strategy in improving the success and interest of mechanical engineering students, using the Kinematics/Dynamics of Machine course as a case study. This upper-division technical course covers a wide range of topics, including mechanism and machine system analysis and synthesis, yet the complexities of ideas like acceleration, motion, and machine component relationships are hard to explain using standard teaching techniques. To solve this problem, a thorough design project was created that gave students hands-on experience developing, manufacturing, and testing their inventions. The main goals of the project were to improve students' grasp of machine design and kinematics, to develop problem-solving and presenting abilities, and to familiarize them with professional software. A questionnaire survey was done to evaluate the effect of this technique on students' performance and interest in mechanical engineering. The outcomes of the study shed light on the usefulness of nontraditional experiential learning approaches in engineering education.Keywords: experiential learning, nontraditional teaching, hands-on design project, engineering education
Procedia PDF Downloads 986813 Understanding Factors that May Affect Survival and Productivity of Pacific Salmonids
Authors: Julia B. Kischkat, Charlie D. Waters
Abstract:
This research aims to understand the factors that may affect the survival and productivity of Pacific salmonids through two components. The first component is lab-based and aims to improve high-performance liquid chromatography to better quantify vitamin deficiencies such as thiamine. The lab work is conducted at the National Oceanic and Atmospheric Administration (NOAA) Ted Stevens Marine Research Institute in Juneau, Alaska. Deficiencies in thiamine have been shown to reduce the survival of salmonids at early life stages. The second component involves the analysis of a 22-year data set of migration timing of juvenile Coho Salmon, Dolly Varden, Steelhead, and returning adult Steelhead at Little Port Walter, Alaska. The statistical analysis quantifies their migration fluctuations and whether they correlate to various environmental conditions such as temperature, salinity, and precipitation.Keywords: climate change, smolt timing, phenology, migration timing, salmon, time series analysis, ecology, chemistry, fisheries science
Procedia PDF Downloads 1186812 A Formal Property Verification for Aspect-Oriented Programs in Software Development
Authors: Moustapha Bande, Hakima Ould-Slimane, Hanifa Boucheneb
Abstract:
Software development for complex systems requires efficient and automatic tools that can be used to verify the satisfiability of some critical properties such as security ones. With the emergence of Aspect-Oriented Programming (AOP), considerable work has been done in order to better modularize the separation of concerns in the software design and implementation. The goal is to prevent the cross-cutting concerns to be scattered across the multiple modules of the program and tangled with other modules. One of the key challenges in the aspect-oriented programs is to be sure that all the pieces put together at the weaving time ensure the satisfiability of the overall system requirements. Our paper focuses on this problem and proposes a formal property verification approach for a given property from the woven program. The approach is based on the control flow graph (CFG) of the woven program, and the use of a satisfiability modulo theories (SMT) solver to check whether each property (represented par one aspect) is satisfied or not once the weaving is done.Keywords: aspect-oriented programming, control flow graph, property verification, satisfiability modulo theories
Procedia PDF Downloads 1776811 Digital Image Correlation: Metrological Characterization in Mechanical Analysis
Authors: D. Signore, M. Ferraiuolo, P. Caramuta, O. Petrella, C. Toscano
Abstract:
The Digital Image Correlation (DIC) is a newly developed optical technique that is spreading in all engineering sectors because it allows the non-destructive estimation of the entire surface deformation without any contact with the component under analysis. These characteristics make the DIC very appealing in all the cases the global deformation state is to be known without using strain gages, which are the most used measuring device. The DIC is applicable to any material subjected to distortion caused by either thermal or mechanical load, allowing to obtain high-definition mapping of displacements and deformations. That is why in the civil and the transportation industry, DIC is very useful for studying the behavior of metallic materials as well as of composite materials. DIC is also used in the medical field for the characterization of the local strain field of the vascular tissues surface subjected to uniaxial tensile loading. DIC can be carried out in the two dimension mode (2D DIC) if a single camera is used or in a three dimension mode (3D DIC) if two cameras are involved. Each point of the test surface framed by the cameras can be associated with a specific pixel of the image, and the coordinates of each point are calculated knowing the relative distance between the two cameras together with their orientation. In both arrangements, when a component is subjected to a load, several images related to different deformation states can be are acquired through the cameras. A specific software analyzes the images via the mutual correlation between the reference image (obtained without any applied load) and those acquired during the deformation giving the relative displacements. In this paper, a metrological characterization of the digital image correlation is performed on aluminum and composite targets both in static and dynamic loading conditions by comparison between DIC and strain gauges measures. In the static test, interesting results have been obtained thanks to an excellent agreement between the two measuring techniques. In addition, the deformation detected by the DIC is compliant with the result of a FEM simulation. In the dynamic test, the DIC was able to follow with a good accuracy the periodic deformation of the specimen giving results coherent with the ones given by FEM simulation. In both situations, it was seen that the DIC measurement accuracy depends on several parameters such as the optical focusing, the parameters chosen to perform the mutual correlation between the images and, finally, the reference points on image to be analyzed. In the future, the influence of these parameters will be studied, and a method to increase the accuracy of the measurements will be developed in accordance with the requirements of the industries especially of the aerospace one.Keywords: accuracy, deformation, image correlation, mechanical analysis
Procedia PDF Downloads 3116810 Method and System of Malay Traditional Women Apparel Pattern Drafting for Hazi Attire
Authors: Haziyah Hussin
Abstract:
Hazi Attire software is purposely designed to be used for pattern drafting of the Malay Traditional Women Apparel. It is software created using LISP Program that works under AutoCAD engine and able to draft various patterns for Malay women apparels from fitted, semi-fitted and loose silhouettes. It is fully automatic and the user can select styles from the menu on the screen and enter the measurements. Within five seconds patterns are ready to be printed and sewn. Hazi Attire is different from other programmes available in the market since it is fully automatic, user-friendly and able to print selected pattern chosen quickly and accurately. With this software (Hazi Attire), the selected styles can be generated the pattern according to made-to-measure or standard sizes. It would benefit the apparel industries by reducing manufacturing lead time and cycle time.Keywords: basic pattern, pattern drafting, toile, Malay traditional women apparel, the measurement parameters, fitted, semi-fitted and loose silhouette
Procedia PDF Downloads 2726809 The Effect of Incorporation of Inulin as a Fat Replacer on the Quality of Milk Products Vis-À-Vis Ice Cream
Authors: Harish Kumar Sharma
Abstract:
The influence of different levels of inulin as a fat replacer on the quality of ice cream was investigated. The physicochemical, rheological and textural properties of control ice cream and ice cream prepared with inulin in different proportions were determined and correlated to the different parameters using Pearson correlation and Principle Component Analysis (PCA). Based on the overall acepectability, ice cream with 4% inulin was found best and was selected for preparation of ice cream with inulin:SPI in different proportions. Compared with control ice cream, Inulin:SPI showed different rheological properties, resulting in significantly higher apparent viscosities, consistency coefficient and greater deviations from Newtonian flow. In addition, both hardness and melting resistance significantly increased with increase in the SPI content in ice cream prepared with inulin: SPI. Also hardness value increased for inulin based ice cream compared to control ice cream but it melted significantly faster than the latter. Colour value significantly decreased in both the cases compared to the control sample. The deliberation shall focus on the effect of incorporation of inulin on the quality of ice-cream.Keywords: fat replacer, inulin, ice cream, viscosity, principal component analysis
Procedia PDF Downloads 3846808 Internal Power Recovery in Cryogenic Cooling Plants Part I: Expander Development
Authors: Ambra Giovannelli, Erika Maria Archilei
Abstract:
The amount of the electrical power required by refrigeration systems is relevant worldwide. It is evaluated in the order of 15% of the total electricity production taking refrigeration and air-conditioning into consideration. For this reason, in the last years several energy saving techniques have been proposed to reduce the power demand of such plants. The paper deals with the development of an innovative internal recovery system for cryogenic cooling plants. Such a system consists in a Compressor-Expander Group (CEG) designed on the basis of the automotive turbocharging technology. In particular, the paper is focused on the design of the expander, the critical component of the CEG system. Due to the low volumetric flow entering the expander and the high expansion ratio, a commercial turbocharger expander wheel was strongly modified. It was equipped with a transonic nozzle, designed to have a radially inflow full admission. To verify the performance of such a machine and suggest improvements, two different set of nozzles have been designed and modelled by means of the commercial Ansys-CFX software. steady-state 3D CFD simulations of the second-generation prototype are presented and compared with the initial ones.Keywords: vapour cCompression systems, energy saving, refrigeration plant, organic fluids, radial turbine
Procedia PDF Downloads 2086807 A Stable Method for Determination of the Number of Independent Components
Authors: Yuyan Yi, Jingyi Zheng, Nedret Billor
Abstract:
Independent component analysis (ICA) is one of the most commonly used blind source separation (BSS) techniques for signal pre-processing, such as noise reduction and feature extraction. The main parameter in the ICA method is the number of independent components (IC). Although there have been several methods for the determination of the number of ICs, it has not been given sufficient attentionto this important parameter. In this study, wereview the mostused methods fordetermining the number of ICs and providetheir advantages and disadvantages. Further, wepropose an improved version of column-wise ICAByBlock method for the determination of the number of ICs.To assess the performance of the proposed method, we compare the column-wise ICAbyBlock with several existing methods through different ICA methods by using simulated and real signal data. Results show that the proposed column-wise ICAbyBlock is an effective and stable method for determining the optimal number of components in ICA. This method is simple, and results can be demonstrated intuitively with good visualizations.Keywords: independent component analysis, optimal number, column-wise, correlation coefficient, cross-validation, ICAByblock
Procedia PDF Downloads 996806 Local Tax Map Software System Development
Authors: Smithinun Thairoongrojana
Abstract:
This research is a qualitative research with three main purposes: (1) to develop the local tax map software system to be linked to the main Local Tax Map System (LTAX3000) system; (2) to design and develop a program for tax data fieldwork on wireless devices and link it to LTAX3000 database of Surat Thani Municipality; (3) to develop the human resource responsible for the fieldwork to be able to use the program and maintain the system and also to be able to work with the dynamic of technologies. In-depth interviews with the two groups of samples, the board of Surat Thani Municipality and operation staff responsible for observing and taxing fieldworks were conducted. The result of this study demonstrates the new developed fieldworks system that can be used both stand-alone usage and networking usage. The fieldworks system to collect and store the variety of taxing information within Surat Thani Municipality will be explained. Then the fieldwork operation process development and the replacement of transferring and storing the information via the network communication.Keywords: Local tax map, software system development, wireless devices, human resource
Procedia PDF Downloads 1926805 Assessment of the Psychoemotional State and Quality of Life at Women Teachers of the Senior Age Group
Authors: Meruyert Burumbayeva, Aiman Mussina, Gulnoza Aldabekova, Aiymtory Abildaeva, Gulshat Yerdenova, Aigul Kairgeldina
Abstract:
this article introduces results of a research which purpose is evaluation the quality of life, the psychophysiological status, expressiveness of uneasiness at women teachers of the senior age group. At a research of quality of life of teachers the lowest values have been received from the indicators of the general state of health, vital activity, role emotional functioning and mental health. Every second woman-teacher noted high personal uneasiness; every third woman-teacher noted moderate situational uneasiness, confirming the existence of a professional stress. Revealed the interrelation between alarming conditions and a decrease in a mental component of health. Moreover, there was revealed exhaustion signs at low activity values that indicate a high tension of labor process.Keywords: expressiveness of uneasiness, quality of life, psychophysiological status, component of health
Procedia PDF Downloads 2946804 Electronic States at SnO/SnO2 Heterointerfaces
Authors: A. Albar, U. Schwingenschlogel
Abstract:
Device applications of transparent conducting oxides require a thorough understanding of the physical and chemical properties of the involved interfaces. We use ab-initio calculations within density functional theory to investigate the electronic states at the SnO/SnO2 hetero-interface. Tin dioxide and monoxide are transparent materials with high n-type and p-type mobilities, respectively. This work aims at exploring the modifications of the electronic states, in particular the charge transfer, in the vicinity of the hetero-interface. The (110) interface is modeled by a super-cell approach in order to minimize the mismatch between the lattice parameters of the two compounds. We discuss the electronic density of states as a function of the distance to the interface.Keywords: density of states, ab-initio calculations, interface states, charge transfer
Procedia PDF Downloads 4186803 The Effect of an Electric Field on the Falling Film Evaporation
Authors: Abdelaziz Nasr
Abstract:
This work conducted numerical simulations to examine the impact of the static electric field on a falling-film evaporation system. A constant electric field can alter the dynamics of a liquid film by modifying the heat and mass transfer properties of the system. The geometry problem consists of two parallel plates in a vertical channel, with the left plate experiencing a constant heat flux and the liquid flowing downward over it, while the right plate remains dry and maintains a constant temperature. The gaseous component consists of dry air and water vapor, whilst the liquid component comprises a thin coating of water. The results suggest that the electric field's impact on heat and mass transport, as well as the evaporation of the liquid sheet, is minimal. Experimental evidence demonstrates that the electric field exerts a minor influence on heat, mass transport, and liquid film evaporation at elevated electric field intensities.Keywords: electric field, evaporation, liquid film, heat and mass transfer
Procedia PDF Downloads 12