Search results for: software modelling
5907 Comparison of Tensile Strength and Folding Endurance of (FDM Process) 3D Printed ABS and PLA Materials
Authors: R. Devicharan
Abstract:
In a short span 3D Printing is expected to play a vital role in our life. The possibility of creativity and speed in manufacturing through various 3D printing processes is infinite. This study is performed on the FDM (Fused Deposition Modelling) method of 3D printing, which is one of the pre-dominant methods of 3D printing technologies. This study focuses on physical properties of the objects produced by 3D printing which determine the applications of the 3D printed objects. This paper specifically aims at the study of the tensile strength and the folding endurance of the 3D printed objects through the FDM (Fused Deposition Modelling) method using the ABS (Acronitirile Butadiene Styrene) and PLA (Poly Lactic Acid) plastic materials. The study is performed on a controlled environment and the specific machine settings. Appropriate tables, graphs are plotted and research analysis techniques will be utilized to analyse, verify and validate the experiment results.Keywords: FDM process, 3D printing, ABS for 3D printing, PLA for 3D printing, rapid prototyping
Procedia PDF Downloads 5995906 Digital Preservation: A Need of Tomorrow
Authors: Gaurav Kumar
Abstract:
Digital libraries have been established all over the world to create, maintain and to preserve the digital materials. This paper exhibits the importance and objectives of digital preservation. The necessities of preservation are hardware and software technology to interpret the digital documents and discuss various aspects of digital preservation.Keywords: preservation, digital preservation, conservation, archive, repository, document, information technology, hardware, software, organization, machine readable format
Procedia PDF Downloads 5885905 A Review Paper for Detecting Zero-Day Vulnerabilities
Authors: Tshegofatso Rambau, Tonderai Muchenje
Abstract:
Zero-day attacks (ZDA) are increasing day by day; there are many vulnerabilities in systems and software that date back decades. Companies keep discovering vulnerabilities in their systems and software and work to release patches and updates. A zero-day vulnerability is a software fault that is not widely known and is unknown to the vendor; attackers work very quickly to exploit these vulnerabilities. These are major security threats with a high success rate because businesses lack the essential safeguards to detect and prevent them. This study focuses on the factors and techniques that can help us detect zero-day attacks. There are various methods and techniques for detecting vulnerabilities. Various companies like edges can offer penetration testing and smart vulnerability management solutions. We will undertake literature studies on zero-day attacks and detection methods, as well as modeling approaches and simulations, as part of the study process.Keywords: zero-day attacks, exploitation, vulnerabilities
Procedia PDF Downloads 1025904 Effective Use of X-Box Kinect in Rehabilitation Centers of Riyadh
Authors: Reem Alshiha, Tanzila Saba
Abstract:
Physical rehabilitation is the process of helping people to recover and be able to go back to their former activities that have been delayed due to external factors such as car accidents, old age and victims of strokes (chronic diseases and accidents, and those related to sport activities).The cost of hiring a personal nurse or driving the patient to and from the hospital could be costly and time-consuming. Also, there are other factors to take into account such as forgetfulness, boredom and lack of motivation. In order to solve this dilemma, some experts came up with rehabilitation software to be used with Microsoft Kinect to help the patients and their families for in-home rehabilitation. In home rehabilitation software is becoming more and more popular, since it is more convenient for all parties affiliated with the patient. In contrast to the other costly market-based systems that have no portability, Microsoft’s Kinect is a portable motion sensor that reads body movements and interprets it. New software development has made rehabilitation games available to be used at home for the convenience of the patient. The game will benefit its users (rehabilitation patients) in saving time and money. There are many software's that are used with the Kinect for rehabilitation, but the software that is chosen in this research is Kinectotherapy. Kinectotherapy software is used for rehabilitation patients in Riyadh clinics to test its acceptance by patients and their physicians. In this study, we used Kinect because it was affordable, portable and easy to access in contrast to expensive market-based motion sensors. This paper explores the importance of in-home rehabilitation by using Kinect with Kinectotherapy software. The software targets both upper and lower limbs, but in this research, the main focus is on upper-limb functionality. However, the in-home rehabilitation is applicable to be used by all patients with motor disability, since the patient must have some self-reliance. The targeted subjects are patients with minor motor impairment that are somewhat independent in their mobility. The presented work is the first to consider the implementation of in-home rehabilitation with real-time feedback to the patient and physician. This research proposes the implementation of in-home rehabilitation in Riyadh, Saudi Arabia. The findings show that most of the patients are interested and motivated in using the in-home rehabilitation system in the future. The main value of the software application is due to these factors: improve patient engagement through stimulating rehabilitation, be a low cost rehabilitation tool and reduce the need for expensive one-to-one clinical contact. Rehabilitation is a crucial treatment that can improve the quality of life and confidence of the patient as well as their self-esteem.Keywords: x-box, rehabilitation, physical therapy, rehabilitation software, kinect
Procedia PDF Downloads 3445903 Quality Assurance in Software Design Patterns
Authors: Rabbia Tariq, Hannan Sajjad, Mehreen Sirshar
Abstract:
Design patterns are widely used to make the process of development easier as they greatly help the developers to develop the software. Different design patterns have been introduced till now but the behavior of same design pattern may differ in different domains that can lead to the wrong selection of the design pattern. The paper aims to discover the design patterns that suits best with respect to their domain thereby helping the developers to choose an effective design pattern. It presents the comprehensive analysis of design patterns based on different methodologies that include simulation, case study and comparison of various algorithms. Due to the difference of the domain the methodology used in one domain may be inapplicable to the other domain. The paper draws a conclusion based on strength and limitation of each design pattern in their respective domain.Keywords: design patterns, evaluation, quality assurance, software domains
Procedia PDF Downloads 5225902 In situ Real-Time Multivariate Analysis of Methanolysis Monitoring of Sunflower Oil Using FTIR
Authors: Pascal Mwenge, Tumisang Seodigeng
Abstract:
The combination of world population and the third industrial revolution led to high demand for fuels. On the other hand, the decrease of global fossil 8fuels deposits and the environmental air pollution caused by these fuels has compounded the challenges the world faces due to its need for energy. Therefore, new forms of environmentally friendly and renewable fuels such as biodiesel are needed. The primary analytical techniques for methanolysis yield monitoring have been chromatography and spectroscopy, these methods have been proven reliable but are more demanding, costly and do not provide real-time monitoring. In this work, the in situ monitoring of biodiesel from sunflower oil using FTIR (Fourier Transform Infrared) has been studied; the study was performed using EasyMax Mettler Toledo reactor equipped with a DiComp (Diamond) probe. The quantitative monitoring of methanolysis was performed by building a quantitative model with multivariate calibration using iC Quant module from iC IR 7.0 software. 15 samples of known concentrations were used for the modelling which were taken in duplicate for model calibration and cross-validation, data were pre-processed using mean centering and variance scale, spectrum math square root and solvent subtraction. These pre-processing methods improved the performance indexes from 7.98 to 0.0096, 11.2 to 3.41, 6.32 to 2.72, 0.9416 to 0.9999, RMSEC, RMSECV, RMSEP and R2Cum, respectively. The R2 value of 1 (training), 0.9918 (test), 0.9946 (cross-validation) indicated the fitness of the model built. The model was tested against univariate model; small discrepancies were observed at low concentration due to unmodelled intermediates but were quite close at concentrations above 18%. The software eliminated the complexity of the Partial Least Square (PLS) chemometrics. It was concluded that the model obtained could be used to monitor methanol of sunflower oil at industrial and lab scale.Keywords: biodiesel, calibration, chemometrics, methanolysis, multivariate analysis, transesterification, FTIR
Procedia PDF Downloads 1495901 Coupling Concept of Two Parallel Research Codes for Two and Three Dimensional Fluid Structure Interaction Analysis
Authors: Luciano Garelli, Marco Schauer, Jorge D’Elia, Mario A. Storti, Sabine C. Langer
Abstract:
This paper discuss a coupling strategy of two different software packages to provide fluid structure interaction (FSI) analysis. The basic idea is to combine the advantages of the two codes to create a powerful FSI solver for two and three dimensional analysis. The fluid part is computed by a program called PETSc-FEM, a software developed at Centro de Investigación de Métodos Computacionales (CIMEC). The structural part of the coupled process is computed by the research code elementary Parallel Solver (elPaSo) of the Technische Universität Braunschweig, Institut für Konstruktionstechnik (IK).Keywords: computational fluid dynamics (CFD), fluid structure interaction (FSI), finite element method (FEM), software
Procedia PDF Downloads 5535900 All-or-None Principle and Weakness of Hodgkin-Huxley Mathematical Model
Authors: S. A. Sadegh Zadeh, C. Kambhampati
Abstract:
Mathematical and computational modellings are the necessary tools for reviewing, analysing, and predicting processes and events in the wide spectrum range of scientific fields. Therefore, in a field as rapidly developing as neuroscience, the combination of these two modellings can have a significant role in helping to guide the direction the field takes. The paper combined mathematical and computational modelling to prove a weakness in a very precious model in neuroscience. This paper is intended to analyse all-or-none principle in Hodgkin-Huxley mathematical model. By implementation the computational model of Hodgkin-Huxley model and applying the concept of all-or-none principle, an investigation on this mathematical model has been performed. The results clearly showed that the mathematical model of Hodgkin-Huxley does not observe this fundamental law in neurophysiology to generating action potentials. This study shows that further mathematical studies on the Hodgkin-Huxley model are needed in order to create a model without this weakness.Keywords: all-or-none, computational modelling, mathematical model, transmembrane voltage, action potential
Procedia PDF Downloads 6175899 Experiences of Timing Analysis of Parallel Embedded Software
Authors: Muhammad Waqar Aziz, Syed Abdul Baqi Shah
Abstract:
The execution time analysis is fundamental to the successful design and execution of real-time embedded software. In such analysis, the Worst-Case Execution Time (WCET) of a program is a key measure, on the basis of which system tasks are scheduled. The WCET analysis of embedded software is also needed for system understanding and to guarantee its behavior. WCET analysis can be performed statically (without executing the program) or dynamically (through measurement). Traditionally, research on the WCET analysis assumes sequential code running on single-core platforms. However, as computation is steadily moving towards using a combination of parallel programs and multi-core hardware, new challenges in WCET analysis need to be addressed. In this article, we report our experiences of performing the WCET analysis of Parallel Embedded Software (PES) running on multi-core platform. The primary purpose was to investigate how WCET estimates of PES can be computed statically, and how they can be derived dynamically. Our experiences, as reported in this article, include the challenges we faced, possible suggestions to these challenges and the workarounds that were developed. This article also provides observations on the benefits and drawbacks of deriving the WCET estimates using the said methods and provides useful recommendations for further research in this area.Keywords: embedded software, worst-case execution-time analysis, static flow analysis, measurement-based analysis, parallel computing
Procedia PDF Downloads 3245898 Genetic Algorithm Optimization of Microcantilever Based Resonator
Authors: Manjula Sutagundar, B. G. Sheeparamatti, D. S. Jangamshetti
Abstract:
Micro Electro Mechanical Systems (MEMS) resonators have shown the potential of replacing quartz crystal technology for sensing and high frequency signal processing applications because of inherent advantages like small size, high quality factor, low cost, compatibility with integrated circuit chips. This paper presents the optimization and modelling and simulation of the optimized micro cantilever resonator. The objective of the work is to optimize the dimensions of a micro cantilever resonator for a specified range of resonant frequency and specific quality factor. Optimization is carried out using genetic algorithm. The genetic algorithm is implemented using MATLAB. The micro cantilever resonator is modelled in CoventorWare using the optimized dimensions obtained from genetic algorithm. The modeled cantilever is analysed for resonance frequency.Keywords: MEMS resonator, genetic algorithm, modelling and simulation, optimization
Procedia PDF Downloads 5505897 Risk Assessment of Building Information Modelling Adoption in Construction Projects
Authors: Amirhossein Karamoozian, Desheng Wu, Behzad Abbasnejad
Abstract:
Building information modelling (BIM) is a new technology to enhance the efficiency of project management in the construction industry. In addition to the potential benefits of this useful technology, there are various risks and obstacles to applying it in construction projects. In this study, a decision making approach is presented for risk assessment in BIM adoption in construction projects. Various risk factors of exerting BIM during different phases of the project lifecycle are identified with the help of Delphi method, experts’ opinions and related literature. Afterward, Shannon’s entropy and Fuzzy TOPSIS (Technique for Order Preference by Similarity to Ideal Situation) are applied to derive priorities of the identified risk factors. Results indicated that lack of knowledge between professional engineers about workflows in BIM and conflict of opinions between different stakeholders are the risk factors with the highest priority.Keywords: risk, BIM, fuzzy TOPSIS, construction projects
Procedia PDF Downloads 2305896 R Software for Parameter Estimation of Spatio-Temporal Model
Authors: Budi Nurani Ruchjana, Atje Setiawan Abdullah, I. Gede Nyoman Mindra Jaya, Eddy Hermawan
Abstract:
In this paper, we propose the application package to estimate parameters of spatiotemporal model based on the multivariate time series analysis using the R open-source software. We build packages mainly to estimate the parameters of the Generalized Space Time Autoregressive (GSTAR) model. GSTAR is a combination of time series and spatial models that have parameters vary per location. We use the method of Ordinary Least Squares (OLS) and use the Mean Average Percentage Error (MAPE) to fit the model to spatiotemporal real phenomenon. For case study, we use oil production data from volcanic layer at Jatibarang Indonesia or climate data such as rainfall in Indonesia. Software R is very user-friendly and it is making calculation easier, processing the data is accurate and faster. Limitations R script for the estimation of model parameters spatiotemporal GSTAR built is still limited to a stationary time series model. Therefore, the R program under windows can be developed either for theoretical studies and application.Keywords: GSTAR Model, MAPE, OLS method, oil production, R software
Procedia PDF Downloads 2435895 Automated Java Testing: JUnit versus AspectJ
Authors: Manish Jain, Dinesh Gopalani
Abstract:
Growing dependency of mankind on software technology increases the need for thorough testing of the software applications and automated testing techniques that support testing activities. We have outlined our testing strategy for performing various types of automated testing of Java applications using AspectJ which has become the de-facto standard for Aspect Oriented Programming (AOP). Likewise JUnit, a unit testing framework is the most popular Java testing tool. In this paper, we have evaluated our proposed AOP approach for automated testing and JUnit on various parameters. First we have provided the similarity between the two approaches and then we have done a detailed comparison of the two testing techniques on factors like lines of testing code, learning curve, testing of private members etc. We established that our AOP testing approach using AspectJ has got several advantages and is thus particularly more effective than JUnit.Keywords: aspect oriented programming, AspectJ, aspects, JU-nit, software testing
Procedia PDF Downloads 3315894 Drying and Transport Processes in Distributed Hydrological Modelling Based on Finite Volume Schemes (Iber Model)
Authors: Carlos Caro, Ernest Bladé, Pedro Acosta, Camilo Lesmes
Abstract:
The drying-wet process is one of the topics to be more careful in distributed hydrological modeling using finite volume schemes as a means of solving the equations of Saint Venant. In a hydrologic and hydraulic computer model, surface flow phenomena depend mainly on the different flow accumulation and subsequent runoff generation. These accumulations are generated by routing, cell by cell, from the heights of water, which begin to appear due to the rain at each instant of time. Determine when it is considered a dry cell and when considered wet to include in the full calculation is an issue that directly affects the quantification of direct runoff or generation of flow at the end of a zone of contribution by accumulations flow generated from cells or finite volume.Keywords: hydrology, transport processes, hydrological modelling, finite volume schemes
Procedia PDF Downloads 3865893 Introducing Design Principles for Clinical Decision Support Systems
Authors: Luca Martignoni
Abstract:
The increasing usage of clinical decision support systems in healthcare and the demand for software that enables doctors to take informed decisions is changing everyday clinical practice. However, as technology advances not only are the benefits of technology growing, but so are the potential risks. A growing danger is the doctors’ over-reliance on the proposed decision of the clinical decision support system, leading towards deskilling and rash decisions by doctors. In that regard, identifying doctors' requirements for software and developing approaches to prevent technological over-reliance is of utmost importance. In this paper, we report the results of a design science research study, focusing on the requirements and design principles of ultrasound software. We conducted a total of 15 interviews with experts about poten-tial ultrasound software functions. Subsequently, we developed meta-requirements and design principles to design future clinical decision support systems efficiently and as free from the occur-rence of technological over-reliance as possible.Keywords: clinical decision support systems, technological over-reliance, design principles, design science research
Procedia PDF Downloads 1015892 Dynamic Analysis of Turbo Machinery Foundation for Different Rotating Speed
Authors: Sungyani Tripathy, Atul Desai
Abstract:
Turbo machinery Frame Foundation is very important for power generation, gas, steam, hydro, geothermal and nuclear power plants. The Turbo machinery Foundation system was simulated in SAP: 2000 software and dynamic response of foundation was analysed. In this paper, the detailed study of turbo machinery foundation with different running speed has considered. The different revolution per minute considered in this study is 4000 rpm, 6000 rpm, 8000 rpm, 1000 rpm and 12000 rpm. The above analysis has been carried out considering Winkler spring soil model, solid finite element modelling and dynamic analysis of Turbo machinery foundations. The comparison of frequency and time periods at various mode shapes are addressed in this study. Current work investigates the effect of damping on the response spectra curve at the foundation top deck, considering the dynamic machine load. It has been found that turbo generator foundation with haunches remains more elastic during seismic action for different running speeds.Keywords: turbo machinery, SAP: 2000, response spectra, running speeds
Procedia PDF Downloads 2565891 Finite Element Analysis of Rom Silo Subjected to 5000 Tons Monotic Loads at an Anonymous Mine in Zimbabwe
Authors: T. Mushiri, K. Tengende, C. Mbohwa, T. Garikayi
Abstract:
This paper introduces finite element analysis of Run off Mine (ROM) silo subjected to dynamic loading. The proposed procedure is based on the use of theoretical equations to come up with pressure and forces exerted by Platinum Group Metals (PGMs) ore to the silo wall. Finite Element Analysis of the silo involves the use of CAD software (AutoCAD) for3D creation and CAE software (T-FLEX) for the simulation work with an optimization routine to minimize the mass and also ensure structural stiffness and stability. In this research an efficient way to design and analysis of a silo in 3D T-FLEX (CAD) program was created the silo to stay within the constrains and so as to know the points of failure due dynamic loading.Keywords: reinforced concrete silo, finite element analysis, T-FLEX software, AutoCAD
Procedia PDF Downloads 4825890 Implementation and Demonstration of Software-Defined Traffic Grooming
Authors: Lei Guo, Xu Zhang, Weigang Hou
Abstract:
Since the traditional network is closed and it has no architecture to create applications, it has been unable to evolve with changing demands under the rapid innovation in services. Additionally, due to the lack of the whole network profile, the quality of service cannot be well guaranteed in the traditional network. The Software Defined Network (SDN) utilizes global resources to support on-demand applications/services via open, standardized and programmable interfaces. In this paper, we implement the traffic grooming application under a real SDN environment, and the corresponding analysis is made. In our SDN: 1) we use OpenFlow protocol to control the entire network by using software applications running on the network operating system; 2) several virtual switches are combined into the data forwarding plane through Open vSwitch; 3) An OpenFlow controller, NOX, is involved as a logically centralized control plane that dynamically configures the data forwarding plane; 4) The traffic grooming based on SDN is demonstrated through dynamically modifying the idle time of flow entries. The experimental results demonstrate that the SDN-based traffic grooming effectively reduces the end-to-end delay, and the improvement ratio arrives to 99%.Keywords: NOX, OpenFlow, Software Defined Network (SDN), traffic grooming
Procedia PDF Downloads 2515889 Simplifying the Migration of Architectures in Embedded Applications Introducing a Pattern Language to Support the Workforce
Authors: Farha Lakhani, Michael J. Pont
Abstract:
There are two main architectures used to develop software for modern embedded systems: these can be labelled as “event-triggered” (ET) and “time-triggered” (TT). The research presented in this paper is concerned with the issues involved in migration between these two architectures. Although TT architectures are widely used in safety-critical applications they are less familiar to developers of mainstream embedded systems. The research presented in this paper began from the premise that–for a broad class of systems that have been implemented using an ET architecture–migration to a TT architecture would improve reliability. It may be tempting to assume that conversion between ET and TT designs will simply involve converting all event-handling software routines into periodic activities. However, the required changes to the software architecture are, in many cases rather more profound. The main contribution of the work presented in this paper is to identify ways in which the significant effort involved in migrating between existing ET architectures and “equivalent” (and effective) TT architectures could be reduced. The research described in this paper has taken an innovative step in this regard by introducing the use of ‘Design patterns’ for this purpose for the first time.Keywords: embedded applications, software architectures, reliability, pattern
Procedia PDF Downloads 3295888 The Impact of Introspective Models on Software Engineering
Authors: Rajneekant Bachan, Dhanush Vijay
Abstract:
The visualization of operating systems has refined the Turing machine, and current trends suggest that the emulation of 32 bit architectures will soon emerge. After years of technical research into Web services, we demonstrate the synthesis of gigabit switches, which embodies the robust principles of theory. Loam, our new algorithm for forward-error correction, is the solution to all of these challenges.Keywords: software engineering, architectures, introspective models, operating systems
Procedia PDF Downloads 5385887 Modelling the Effect of Distancing and Wearing of Face Masks on Transmission of COVID-19 Infection Dynamics
Authors: Nurudeen Oluwasola Lasisi
Abstract:
The COVID-19 is an infection caused by coronavirus, which has been designated as a pandemic in the world. In this paper, we proposed a model to study the effect of distancing and wearing masks on the transmission of COVID-19 infection dynamics. The invariant region of the model is established. The COVID-19 free equilibrium and the reproduction number of the model were obtained. The local and global stability of the model is determined using the linearization technique method and Lyapunov method. It was found that COVID-19 free equilibrium state is locally asymptotically stable in feasible region Ω if R₀ < 1 and globally asymptomatically stable if R₀ < 1, otherwise unstable if R₀ > 1. More so, numerical analysis and simulations of the dynamics of the COVID-19 infection are presented.Keywords: distancing, reproduction number, wearing of mask, local and global stability, modelling, transmission
Procedia PDF Downloads 1385886 Business Intelligence for Profiling of Telecommunication Customer
Authors: Rokhmatul Insani, Hira Laksmiwati Soemitro
Abstract:
Business Intelligence is a methodology that exploits the data to produce information and knowledge systematically, business intelligence can support the decision-making process. Some methods in business intelligence are data warehouse and data mining. A data warehouse can store historical data from transactional data. For data modelling in data warehouse, we apply dimensional modelling by Kimball. While data mining is used to extracting patterns from the data and get insight from the data. Data mining has many techniques, one of which is segmentation. For profiling of telecommunication customer, we use customer segmentation according to customer’s usage of services, customer invoice and customer payment. Customers can be grouped according to their characteristics and can be identified the profitable customers. We apply K-Means Clustering Algorithm for segmentation. The input variable for that algorithm we use RFM (Recency, Frequency and Monetary) model. All process in data mining, we use tools IBM SPSS modeller.Keywords: business intelligence, customer segmentation, data warehouse, data mining
Procedia PDF Downloads 4855885 The Use of the Matlab Software as the Best Way to Recognize Penumbra Region in Radiotherapy
Authors: Alireza Shayegan, Morteza Amirabadi
Abstract:
The y tool was developed to quantitatively compare dose distributions, either measured or calculated. Before computing ɣ, the dose and distance scales of the two distributions, referred to as evaluated and reference, are re-normalized by dose and distance criteria, respectively. The re-normalization allows the dose distribution comparison to be conducted simultaneously along dose and distance axes. Several two-dimensional images were acquired using a Scanning Liquid Ionization Chamber EPID and Extended Dose Range (EDR2) films for regular and irregular radiation fields. The raw images were then converted into two-dimensional dose maps. Transitional and rotational manipulations were performed for images using Matlab software. As evaluated dose distribution maps, they were then compared with the corresponding original dose maps as the reference dose maps.Keywords: energetic electron, gamma function, penumbra, Matlab software
Procedia PDF Downloads 3015884 Seismic Retrofit of Existing Bridge Foundations with Micropiles: 3D Finite Element Analysis
Authors: Mohanad Talal Alfach
Abstract:
This paper concerns the seismic behaviour of soil-piles-bridge reinforced by additional micropiles. The analysis carried out by three-dimensional finite element modelling using the FE software ABAQUS. The soil behaviour is assumed to be elastic with Rayleigh damping, while the micropiles are modeled as 3D elastic beam elements. The bridge deck slab was represented by a concentrated mass at the top of the pier column. The interaction between the added micropiles and the existing piles as well as the performance of the retrofitted soil-pile-superstructure system were investigated for different configurations of additional micropiles (number, position, inclination). Numerical simulation results show that additional micropiles constitute an efficient retrofitting solution. Analysis of results also shows that spacing between existing piles and retrofitting micropiles has little effect; while it is observed a substantial improvement (in case of weak piles/micropiles - soil interface) with reducing the inclination angle of retrofitting micropiles.Keywords: retrofitting, seismic, finite element, micropiles, elastic
Procedia PDF Downloads 1465883 Analysis of the Impact of NVivo and EndNote on Academic Research Productivity
Authors: Sujit K. Basak
Abstract:
The aim of this paper is to analyze the impact of literature review software on researchers. The aim of this study was achieved by analyzing models in terms of perceived usefulness, perceived ease of use, and acceptance level. Collected data was analyzed using WarpPLS 4.0 software. This study used two theoretical frameworks namely Technology Acceptance Model and the Training Needs Assessment Model. The study was experimental and was conducted at a public university in South Africa. The results of the study showed that acceptance level has a high impact on research workload and productivity followed by perceived usefulness and perceived ease of use.Keywords: technology acceptance model, training needs assessment model, literature review software, research productivity
Procedia PDF Downloads 5045882 Discover a New Technique for Cancer Recognition by Analysis and Determination of Fractal Dimension Images in Matlab Software
Authors: Saeedeh Shahbazkhany
Abstract:
Cancer is a terrible disease that, if not diagnosed early, therapy can be difficult while it is easily medicable if it is diagnosed in early stages. So it is very important for cancer diagnosis that medical procedures are performed. In this paper we introduce a new method. In this method, we only need pictures of healthy cells and cancer cells. In fact, where we suspect cancer, we take a picture of cells or tissue in that area, and then take some pictures of the surrounding tissues. Then, fractal dimension of images are calculated and compared. Cancer can be easily detected by comparing the fractal dimension of images. In this method, we use Matlab software.Keywords: Matlab software, fractal dimension, cancer, surrounding tissues, cells or tissue, new method
Procedia PDF Downloads 3545881 Modelling Conceptual Quantities Using Support Vector Machines
Authors: Ka C. Lam, Oluwafunmibi S. Idowu
Abstract:
Uncertainty in cost is a major factor affecting performance of construction projects. To our knowledge, several conceptual cost models have been developed with varying degrees of accuracy. Incorporating conceptual quantities into conceptual cost models could improve the accuracy of early predesign cost estimates. Hence, the development of quantity models for estimating conceptual quantities of framed reinforced concrete structures using supervised machine learning is the aim of the current research. Using measured quantities of structural elements and design variables such as live loads and soil bearing pressures, response and predictor variables were defined and used for constructing conceptual quantities models. Twenty-four models were developed for comparison using a combination of non-parametric support vector regression, linear regression, and bootstrap resampling techniques. R programming language was used for data analysis and model implementation. Gross soil bearing pressure and gross floor loading were discovered to have a major influence on the quantities of concrete and reinforcement used for foundations. Building footprint and gross floor loading had a similar influence on beams and slabs. Future research could explore the modelling of other conceptual quantities for walls, finishes, and services using machine learning techniques. Estimation of conceptual quantities would assist construction planners in early resource planning and enable detailed performance evaluation of early cost predictions.Keywords: bootstrapping, conceptual quantities, modelling, reinforced concrete, support vector regression
Procedia PDF Downloads 2065880 Evaluation of the Matching Optimization of Human-Machine Interface Matching in the Cab
Authors: Yanhua Ma, Lu Zhai, Xinchen Wang, Hongyu Liang
Abstract:
In this paper, by understanding the development status of the human-machine interface in today's automobile cab, a subjective and objective evaluation system for evaluating the optimization of human-machine interface matching in automobile cab was established. The man-machine interface of the car cab was divided into a software interface and a hard interface. Objective evaluation method of software human factor analysis is used to evaluate the hard interface matching; The analytic hierarchy process is used to establish the evaluation index system for the software interface matching optimization, and the multi-level fuzzy comprehensive evaluation method is used to evaluate hard interface machine. This article takes Dongfeng Sokon (DFSK) C37 model automobile as an example. The evaluation method given in the paper is used to carry out relevant analysis and evaluation, and corresponding optimization suggestions are given, which have certain reference value for designers.Keywords: analytic hierarchy process, fuzzy comprehension evaluation method, human-machine interface, matching optimization, software human factor analysis
Procedia PDF Downloads 1585879 Empirical Study From Final Exams of Graduate Courses in Computer Science to Demystify the Notion of an Average Software Engineer and Offer a Direction to Address Diversity of Professional Backgrounds of a Student Body
Authors: Alex Elentukh
Abstract:
The paper is based on data collected from final exams administered during five years of teaching the graduate course in software engineering. The visualization instrument with four distinct personas has been used to improve the effectiveness of each class. The study offers a plethora of clues toward students' behavioral preferences. Diversity among students (professional background, physical proximity) is too significant to assume a single face of a learner. This is particularly true for a body of online graduate students in computer science. Conclusions of the study (each learner is unique, and each class is unique) are extrapolated to demystify the notion of an 'average software engineer.' An immediate direction for an educator is to ensure a course applies to a wide audience of very different individuals. On the other hand, a student should be clear about his/her abilities and preferences - to follow the most effective learning path.Keywords: K.3.2 computer and information science education, learner profiling, adaptive learning, software engineering
Procedia PDF Downloads 1035878 Mixtures of Length-Biased Weibull Distributions for Loss Severity Modelling
Authors: Taehan Bae
Abstract:
In this paper, a class of length-biased Weibull mixtures is presented to model loss severity data. The proposed model generalizes the Erlang mixtures with the common scale parameter, and it shares many important modelling features, such as flexibility to fit various data distribution shapes and weak-denseness in the class of positive continuous distributions, with the Erlang mixtures. We show that the asymptotic tail estimate of the length-biased Weibull mixture is Weibull-type, which makes the model effective to fit loss severity data with heavy-tailed observations. A method of statistical estimation is discussed with applications on real catastrophic loss data sets.Keywords: Erlang mixture, length-biased distribution, transformed gamma distribution, asymptotic tail estimate, EM algorithm, expectation-maximization algorithm
Procedia PDF Downloads 224