Search results for: LCA tools and data
27250 Exploring Electroactive Polymers for Dynamic Data Physicalization
Authors: Joanna Dauner, Jan Friedrich, Linda Elsner, Kora Kimpel
Abstract:
Active materials such as Electroactive Polymers (EAPs) are promising for the development of novel shape-changing interfaces. This paper explores the potential of EAPs in a multilayer unimorph structure from a design perspective to investigate the visual qualities of the material for dynamic data visualization and data physicalization. We discuss various concepts of how the material can be used for this purpose. Multilayer unimorph EAPs are of particular interest to designers because they can be easily prototyped using everyday materials and tools. By changing the structure and geometry of the EAPs, their movement and behavior can be modified. We present the results of our preliminary user testing, where we evaluated different movement patterns. As a result, we introduce a prototype display built with EAPs for dynamic data physicalization. Finally, we discuss the potentials and drawbacks and identify further open research questions for the design discipline.Keywords: electroactive polymer, shape-changing interfaces, smart material interfaces, data physicalization
Procedia PDF Downloads 9827249 Application of the Building Information Modeling Planning Approach to the Factory Planning
Authors: Peggy Näser
Abstract:
Factory planning is a systematic, objective-oriented process for planning a factory, structured into a sequence of phases, each of which is dependent on the preceding phase and makes use of particular methods and tools, and extending from the setting of objectives to the start of production. The digital factory, on the other hand, is the generic term for a comprehensive network of digital models, methods, and tools – including simulation and 3D visualisation – integrated by a continuous data management system. Its aim is the holistic planning, evaluation and ongoing improvement of all the main structures, processes and resources of the real factory in conjunction with the product. Digital factory planning has already become established in factory planning. The application of Building Information Modeling has not yet been established in factory planning but has been used predominantly in the planning of public buildings. Furthermore, this concept is limited to the planning of the buildings and does not include the planning of equipment of the factory (machines, technical equipment) and their interfaces to the building. BIM is a cooperative method of working, in which the information and data relevant to its lifecycle are consistently recorded, managed and exchanged in a transparent communication between the involved parties on the basis of digital models of a building. Both approaches, the planning approach of Building Information Modeling and the methodical approach of the Digital Factory, are based on the use of a comprehensive data model. Therefore it is necessary to examine how the approach of Building Information Modeling can be extended in the context of factory planning in such a way that an integration of the equipment planning, as well as the building planning, can take place in a common digital model. For this, a number of different perspectives have to be investigated: the equipment perspective including the tools used to implement a comprehensive digital planning process, the communication perspective between the planners of different fields, the legal perspective, that the legal certainty in each country and the quality perspective, on which the quality criteria are defined and the planning will be evaluated. The individual perspectives are examined and illustrated in the article. An approach model for the integration of factory planning into the BIM approach, in particular for the integrated planning of equipment and buildings and the continuous digital planning is developed. For this purpose, the individual factory planning phases are detailed in the sense of the integration of the BIM approach. A comprehensive software concept is shown on the tool. In addition, the prerequisites required for this integrated planning are presented. With the help of the newly developed approach, a better coordination between equipment and buildings is to be achieved, the continuity of the digital factory planning is improved, the data quality is improved and expensive implementation errors are avoided in the implementation.Keywords: building information modeling, digital factory, digital planning, factory planning
Procedia PDF Downloads 26527248 A New Intelligent, Dynamic and Real Time Management System of Sewerage
Authors: R. Tlili Yaakoubi, H.Nakouri, O. Blanpain, S. Lallahem
Abstract:
The current tools for real time management of sewer systems are based on two software tools: the software of weather forecast and the software of hydraulic simulation. The use of the first ones is an important cause of imprecision and uncertainty, the use of the second requires temporal important steps of decision because of their need in times of calculation. This way of proceeding fact that the obtained results are generally different from those waited. The major idea of this project is to change the basic paradigm by approaching the problem by the "automatic" face rather than by that "hydrology". The objective is to make possible the realization of a large number of simulations at very short times (a few seconds) allowing to take place weather forecasts by using directly the real time meditative pluviometric data. The aim is to reach a system where the decision-making is realized from reliable data and where the correction of the error is permanent. A first model of control laws was realized and tested with different return-period rainfalls. The gains obtained in rejecting volume vary from 19 to 100 %. The development of a new algorithm was then used to optimize calculation time and thus to overcome the subsequent combinatorial problem in our first approach. Finally, this new algorithm was tested with 16- year-rainfall series. The obtained gains are 40 % of total volume rejected to the natural environment and of 65 % in the number of discharges.Keywords: automation, optimization, paradigm, RTC
Procedia PDF Downloads 29927247 Data Transformations in Data Envelopment Analysis
Authors: Mansour Mohammadpour
Abstract:
Data transformation refers to the modification of any point in a data set by a mathematical function. When applying transformations, the measurement scale of the data is modified. Data transformations are commonly employed to turn data into the appropriate form, which can serve various functions in the quantitative analysis of the data. This study addresses the investigation of the use of data transformations in Data Envelopment Analysis (DEA). Although data transformations are important options for analysis, they do fundamentally alter the nature of the variable, making the interpretation of the results somewhat more complex.Keywords: data transformation, data envelopment analysis, undesirable data, negative data
Procedia PDF Downloads 2027246 Solar Photovoltaic Pumping and Water Treatment Tools: A Case Study in Ethiopian Village
Authors: Corinna Barraco, Ornella Salimbene
Abstract:
This research involves the Ethiopian locality of Jeldi (North Africa), an area particularly affected by water shortage and in which the pumping and treatment of drinking water are extremely sensitive issues. The study aims to develop and apply low-cost tools for the design of solar water pumping and water purification systems in a not developed country. Consequently, two technical tools have been implemented in Excel i) Solar photovoltaic Pumping (Spv-P) ii) Water treatment (Wt). The Spv-P tool was applied to the existing well (depth 110 [m], dynamic water level 90 [m], static water level 53 [m], well yield 0.1728 [m³h⁻¹]) in the Jeldi area, where estimated water demand is about 50 [m3d-1]. Through the application of the tool, it was designed the water extraction system of the well, obtaining the number of pumps and solar panels necessary for water pumping from the well of Jeldi. Instead, the second tool Wt has been applied in the subsequent phase of extracted water treatment. According to the chemical-physical parameters of the water, Wt returns as output the type of purification treatment(s) necessary to potable the extracted water. In the case of the well of Jeldi, the tool identified a high criticality regarding the turbidity parameter (12 [NTU] vs 5 [NTU]), and a medium criticality regarding the exceeding limits of sodium concentration (234 [mg/L Na⁺] vs 200 [mg/L Na⁺]) and ammonia (0.64 [mg/L NH³-N] vs 0.5 [mg/L NH³-N]). To complete these tools, two specific manuals are provided for the users. The joint use of the two tools would help reduce problems related to access to water resources compared to the current situation and represents a simplified solution for the design of pumping systems and analysis of purification treatments to be performed in undeveloped countries.Keywords: drinking water, Ethiopia, treatments, water pumping
Procedia PDF Downloads 15627245 Conception of a Regulated, Dynamic and Intelligent Sewerage in Ostrevent
Authors: Rabaa Tlili Yaakoubi, Hind Nakouri, Olivier Blanpain
Abstract:
The current tools for real time management of sewer systems are based on two software tools: the software of weather forecast and the software of hydraulic simulation. The use of the first ones is an important cause of imprecision and uncertainty, the use of the second requires temporal important steps of decision because of their need in times of calculation. This way of proceeding fact that the obtained results are generally different from those waited. The major idea of the CARDIO project is to change the basic paradigm by approaching the problem by the "automatic" face rather than by that "hydrology". The objective is to make possible the realization of a large number of simulations at very short times (a few seconds) allowing to take place weather forecasts by using directly the real time meditative pluviometric data. The aim is to reach a system where the decision-making is realized from reliable data and where the correction of the error is permanent. A first model of control laws was realized and tested with different return-period rainfalls. The gains obtained in rejecting volume vary from 40 to 100%. The development of a new algorithm was then used to optimize calculation time and thus to overcome the subsequent combinatorial problem in our first approach. Finally, this new algorithm was tested with 16- year-rainfall series. The obtained gains are 60% of total volume rejected to the natural environment and of 80 % in the number of discharges.Keywords: RTC, paradigm, optimization, automation
Procedia PDF Downloads 28427244 Arabic Light Word Analyser: Roles with Deep Learning Approach
Authors: Mohammed Abu Shquier
Abstract:
This paper introduces a word segmentation method using the novel BP-LSTM-CRF architecture for processing semantic output training. The objective of web morphological analysis tools is to link a formal morpho-syntactic description to a lemma, along with morpho-syntactic information, a vocalized form, a vocalized analysis with morpho-syntactic information, and a list of paradigms. A key objective is to continuously enhance the proposed system through an inductive learning approach that considers semantic influences. The system is currently under construction and development based on data-driven learning. To evaluate the tool, an experiment on homograph analysis was conducted. The tool also encompasses the assumption of deep binary segmentation hypotheses, the arbitrary choice of trigram or n-gram continuation probabilities, language limitations, and morphology for both Modern Standard Arabic (MSA) and Dialectal Arabic (DA), which provide justification for updating this system. Most Arabic word analysis systems are based on the phonotactic morpho-syntactic analysis of a word transmitted using lexical rules, which are mainly used in MENA language technology tools, without taking into account contextual or semantic morphological implications. Therefore, it is necessary to have an automatic analysis tool taking into account the word sense and not only the morpho-syntactic category. Moreover, they are also based on statistical/stochastic models. These stochastic models, such as HMMs, have shown their effectiveness in different NLP applications: part-of-speech tagging, machine translation, speech recognition, etc. As an extension, we focus on language modeling using Recurrent Neural Network (RNN); given that morphological analysis coverage was very low in dialectal Arabic, it is significantly important to investigate deeply how the dialect data influence the accuracy of these approaches by developing dialectal morphological processing tools to show that dialectal variability can support to improve analysis.Keywords: NLP, DL, ML, analyser, MSA, RNN, CNN
Procedia PDF Downloads 4227243 Visual Analytics in K 12 Education: Emerging Dimensions of Complexity
Authors: Linnea Stenliden
Abstract:
The aim of this paper is to understand emerging learning conditions, when a visual analytics is implemented and used in K 12 (education). To date, little attention has been paid to the role visual analytics (digital media and technology that highlight visual data communication in order to support analytical tasks) can play in education, and to the extent to which these tools can process actionable data for young students. This study was conducted in three public K 12 schools, in four social science classes with students aged 10 to 13 years, over a period of two to four weeks at each school. Empirical data were generated using video observations and analyzed with help of metaphors by Latour. The learning conditions are found to be distinguished by broad complexity characterized by four dimensions. These emerge from the actors’ deeply intertwined relations in the activities. The paper argues in relation to the found dimensions that novel approaches to teaching and learning could benefit students’ knowledge building as they work with visual analytics, analyzing visualized data.Keywords: analytical reasoning, complexity, data use, problem space, visual analytics, visual storytelling, translation
Procedia PDF Downloads 37627242 Commericializing Fashion Goods in the Digital Age
Authors: Jianli Hu
Abstract:
The internet has fundamentally revolutionized access to data and ushered new ways of creating and accessing information for commercializing products. Digital media such as computer programs, software, apps, websites, and social media have allowed the proliferation of information and ideas to grow exponentially. In recent years, a new wave of innovative e-commerce formats has begun to emerge in the fashion marketplace, responding to the ever-greater need for transparency and connectivity. For example, many fashion wholesalers and retailers have modified their operations using software systems that enable brands to cost, track, and analyze products and client orders, sales tools that connect buyers and brands to create a more dynamic market-place, as well as retailer-end apps designed to drive traffic back to brick-and-mortar stores. In this paper, we review the recently developed tools and applications of commercializing fashion, and present results of several field studies and interviews with fashion producers and buyers.Keywords: fashion, digital media, commercializing products, internet
Procedia PDF Downloads 37727241 Sustainability of High-Rise Affordable Housing: Critical Issues in Applying Green Building Rating Tools
Authors: Poh Im. Lim, Hillary Yee Qin. Tan
Abstract:
Nowadays, going green has become a trend, and being emphasized in the construction industry. In Malaysia, there are several green rating tools available in the industry and among these, GBI and GreenRE are considered as the most common tools adopted for residential buildings. However, being green is not equal to or making something sustainable. Being sustainable is to take economic, environmental and social aspects into consideration. This is particularly essential in the affordable housing sector as the end-users belong to lower-income and places importance on many socio-economic needs beyond the environmental criteria. This paper discusses the arguments in proposing a sustainability framework that is tailor-made for high-rise affordable housing. In-depth interviews and observation mapping methods were used in gathering inputs from the end-users, non-governmental organisations (NGOs) as well as the professionals. ‘Bottom-up’ approach was applied in this research to show the significance of participation from the local community in the decision-making process. The proposed sustainability framework illustrates the discrepancies between user priorities and what the industry is providing. The outcome of this research suggests that integrating sustainability into high-rise affordable housing is achievable and beneficial to the industry, society, and the environment.Keywords: green building rating tools, high-rise affordable housing, sustainability framework, sustainable development
Procedia PDF Downloads 13827240 On the Development of Medical Additive Manufacturing in Egypt
Authors: Khalid Abdelghany
Abstract:
Additive Manufacturing (AM) is the manufacturing technology that is used to fabricate fast products direct from CAD models in very short time and with minimum operation steps. Jointly with the advancement in medical computer modeling, AM proved to be a very efficient tool to help physicians, orthopedic surgeons and dentists design and fabricate patient-tailored surgical guides, templates and customized implants from the patient’s CT / MRI images. AM jointly with computer-assisted designing/computer-assisted manufacturing (CAD/CAM) technology have enabled medical practitioners to tailor physical models in a patient-and purpose-specific fashion and helped to design and manufacture of templates, appliances and devices with a high range of accuracy using biocompatible materials. In developing countries, there are some technical and financial limitations of implementing such advanced tools as an essential portion of medical applications. CMRDI institute in Egypt has been working in the field of Medical Additive Manufacturing since 2003 and has assisted in the recovery of hundreds of poor patients using these advanced tools. This paper focuses on the surgical and dental use of 3D printing technology in Egypt as a developing country. The presented case studies have been designed and processed using the software tools and additive manufacturing machines in CMRDI through cooperative engineering and medical works. Results showed that the implementation of the additive manufacturing tools in developed countries is successful and could be economical comparing to long treatment plans.Keywords: additive manufacturing, dental and orthopeadic stents, patient specific surgical tools, titanium implants
Procedia PDF Downloads 31527239 Classification of Manufacturing Data for Efficient Processing on an Edge-Cloud Network
Authors: Onyedikachi Ulelu, Andrew P. Longstaff, Simon Fletcher, Simon Parkinson
Abstract:
The widespread interest in 'Industry 4.0' or 'digital manufacturing' has led to significant research requiring the acquisition of data from sensors, instruments, and machine signals. In-depth research then identifies methods of analysis of the massive amounts of data generated before and during manufacture to solve a particular problem. The ultimate goal is for industrial Internet of Things (IIoT) data to be processed automatically to assist with either visualisation or autonomous system decision-making. However, the collection and processing of data in an industrial environment come with a cost. Little research has been undertaken on how to specify optimally what data to capture, transmit, process, and store at various levels of an edge-cloud network. The first step in this specification is to categorise IIoT data for efficient and effective use. This paper proposes the required attributes and classification to take manufacturing digital data from various sources to determine the most suitable location for data processing on the edge-cloud network. The proposed classification framework will minimise overhead in terms of network bandwidth/cost and processing time of machine tool data via efficient decision making on which dataset should be processed at the ‘edge’ and what to send to a remote server (cloud). A fast-and-frugal heuristic method is implemented for this decision-making. The framework is tested using case studies from industrial machine tools for machine productivity and maintenance.Keywords: data classification, decision making, edge computing, industrial IoT, industry 4.0
Procedia PDF Downloads 18027238 Investigating the Effective Factors on Product Performance and Prioritizing Them: Case Study of Pars-Khazar Company
Authors: Ebrahim Sabermaash Eshghi, Donna Sandsmark
Abstract:
Nowadays, successful companies try to create a reliable and unique competitive position in the market. It is important to consider that only choosing and codifying a competitive strategy appropriate with the market conditions does not have any influence on the final performance of the company by itself, but it is the connection and interaction between upstream level strategies and functional level strategies which leads to development of company performance in its operating environment. Given the importance of the subject, this study tries to investigate effective factors on product performance and prioritize them. This study was done with quantitative-qualitative approach (interview and questionnaire). In sum, 103 informed managers and experts of Pars-Khazar Company were investigated in a census. Validity of measure tools was approved through experts’ judgments. Reliability of the tools was also gained through Cronbach's Alpha Coefficient as 0.930 and in sum, validity and reliability of the tools was approved generally. Analysis of collected data was done through Spearman Correlation Test and Friedman Test using SPSS software. The results showed that management of distribution and demand process (0.675), management of Product Pre-test (0.636) and Manufacturing and inventory management(0.628) had the highest correlation with product performance. Prioritization of factors of structure of launching new products based on the average showed that management of volume of launched products and Manufacturing and inventory management had the most importance.Keywords: product performance, home appliances, market, case study
Procedia PDF Downloads 22427237 Improvement of Students’ Active Experience through the Provision of Foundational Architecture Pedagogy by Virtual Reality Tools
Authors: Mehdi Khakzand, Flora Fakourian
Abstract:
It has been seen in recent years that architects are using virtual modeling to help them visualize their projects. Research has indicated that virtual media, particularly virtual reality, enhances architects' comprehension of design and spatial perception. Creating a communal experience for active learning is an essential component of the design process in architecture pedagogy. It has been particularly challenging to replicate design principles as a critical teaching function, and this is a complex issue that demands comprehension. Nonetheless, the usage of simulation should be studied and limited as appropriate. In conjunction with extensive technology, 3D geometric illustration can bridge the gap between the real and virtual worlds. This research intends to deliver a pedagogical experience in the architecture basics course to improve the architectural design process utilizing virtual reality tools. This tool seeks to tackle current challenges in current ways of architectural illustration by offering building geometry illustration, building information (data from the building information model), and simulation results. These tools were tested over three days in a design workshop with 12 architectural students. This article provided an architectural VR-based course and explored its application in boosting students' active experiences. According to the research, this technology can improve students' cognitive skills from challenging simulations by boosting visual understanding.Keywords: active experience, architecture pedagogy, virtual reality, spatial perception
Procedia PDF Downloads 8727236 Presenting a Model for Predicting the State of Being Accident-Prone of Passages According to Neural Network and Spatial Data Analysis
Authors: Hamd Rezaeifar, Hamid Reza Sahriari
Abstract:
Accidents are considered to be one of the challenges of modern life. Due to the fact that the victims of this problem and also internal transportations are getting increased day by day in Iran, studying effective factors of accidents and identifying suitable models and parameters about this issue are absolutely essential. The main purpose of this research has been studying the factors and spatial data affecting accidents of Mashhad during 2007- 2008. In this paper it has been attempted to – through matching spatial layers on each other and finally by elaborating them with the place of accident – at the first step by adding landmarks of the accident and through adding especial fields regarding the existence or non-existence of effective phenomenon on accident, existing information banks of the accidents be completed and in the next step by means of data mining tools and analyzing by neural network, the relationship between these data be evaluated and a logical model be designed for predicting accident-prone spots with minimum error. The model of this article has a very accurate prediction in low-accident spots; yet it has more errors in accident-prone regions due to lack of primary data.Keywords: accident, data mining, neural network, GIS
Procedia PDF Downloads 4727235 Efficient Pre-Processing of Single-Cell Assay for Transposase Accessible Chromatin with High-Throughput Sequencing Data
Authors: Fan Gao, Lior Pachter
Abstract:
The primary tool currently used to pre-process 10X Chromium single-cell ATAC-seq data is Cell Ranger, which can take very long to run on standard datasets. To facilitate rapid pre-processing that enables reproducible workflows, we present a suite of tools called scATAK for pre-processing single-cell ATAC-seq data that is 15 to 18 times faster than Cell Ranger on mouse and human samples. Our tool can also calculate chromatin interaction potential matrices, and generate open chromatin signal and interaction traces for cell groups. We use scATAK tool to explore the chromatin regulatory landscape of a healthy adult human brain and unveil cell-type specific features, and show that it provides a convenient and computational efficient approach for pre-processing single-cell ATAC-seq data.Keywords: single-cell, ATAC-seq, bioinformatics, open chromatin landscape, chromatin interactome
Procedia PDF Downloads 15527234 A Simulated Scenario of WikiGIS to Support the Iteration and Traceability Management of the Geodesign Process
Authors: Wided Batita, Stéphane Roche, Claude Caron
Abstract:
Geodesign is an emergent term related to a new and complex process. Hence, it needs to rethink tools, technologies and platforms in order to efficiently achieve its goals. A few tools have emerged since 2010 such as CommunityViz, GeoPlanner, etc. In the era of Web 2.0 and collaboration, WikiGIS has been proposed as a new category of tools. In this paper, we present WikiGIS functionalities dealing mainly with the iteration and traceability management to support the collaboration of the Geodesign process. Actually, WikiGIS is built on GeoWeb 2.0 technologies —and primarily on wiki— and aims at managing the tracking of participants’ editing. This paper focuses on a simplified simulation to illustrate the strength of WikiGIS in the management of traceability and in the access to history in a Geodesign process. Indeed, a cartographic user interface has been implemented, and then a hypothetical use case has been imagined as proof of concept.Keywords: geodesign, history, traceability, tracking of participants’ editing, WikiGIS
Procedia PDF Downloads 24527233 Performance Analysis of Proprietary and Non-Proprietary Tools for Regression Testing Using Genetic Algorithm
Authors: K. Hema Shankari, R. Thirumalaiselvi, N. V. Balasubramanian
Abstract:
The present paper addresses to the research in the area of regression testing with emphasis on automated tools as well as prioritization of test cases. The uniqueness of regression testing and its cyclic nature is pointed out. The difference in approach between industry, with business model as basis, and academia, with focus on data mining, is highlighted. Test Metrics are discussed as a prelude to our formula for prioritization; a case study is further discussed to illustrate this methodology. An industrial case study is also described in the paper, where the number of test cases is so large that they have to be grouped as Test Suites. In such situations, a genetic algorithm proposed by us can be used to reconfigure these Test Suites in each cycle of regression testing. The comparison is made between a proprietary tool and an open source tool using the above-mentioned metrics. Our approach is clarified through several tables.Keywords: APFD metric, genetic algorithm, regression testing, RFT tool, test case prioritization, selenium tool
Procedia PDF Downloads 43527232 Analyzing Information Management in Science and Technology Institute Libraries in India
Authors: P. M. Naushad Ali
Abstract:
India’s strength in basic research is recognized internationally. Science and Technology research in India has been performed by six distinct bodies or organizations such as Cooperative Research Associations, Autonomous Research Council, Institute under Ministries, Industrial R&D Establishment, Universities, Private Institutions. All most all these institutions are having a well-established library/information center to cater the information needs of their users like scientists and technologists. Information Management (IM) comprises disciplines concerned with the study and the effective and efficient management of information and resources, products and services as well as the understanding of the involved technologies and the people engaged in this activity. It is also observed that the libraries and information centers in India are also using modern technologies for the management of various activities and services to serve their users in a better way. Science and Technology libraries in the country are usually better equipped because the investment in Science and Technology in the country are much larger than those in other fields. Thus, most of the Science and Technology libraries are equipped with modern IT-based tools for handling and management of library services. In spite of these facts Science and Technology libraries are having all the characteristics of a model organization where computer application is found most successful, however, the adoption of this IT based management tool is not uniform in these libraries. The present study will help to know about the level use of IT-based management tools for the information management of Science and Technology libraries in India. The questionnaire, interview, observation and document review techniques have been used in data collection. Finally, the author discusses findings of the study and put forward some suggestions to improve the quality of Science and Technology institute library services in India.Keywords: information management, science and technology libraries, India, IT-based tools
Procedia PDF Downloads 39327231 A Novel Epitope Prediction for Vaccine Designing against Ebola Viral Envelope Proteins
Authors: Manju Kanu, Subrata Sinha, Surabhi Johari
Abstract:
Viral proteins of Ebola viruses belong to one of the best studied viruses; however no effective prevention against EBOV has been developed. Epitope-based vaccines provide a new strategy for prophylactic and therapeutic application of pathogen-specific immunity. A critical requirement of this strategy is the identification and selection of T-cell epitopes that act as vaccine targets. This study describes current methodologies for the selection process, with Ebola virus as a model system. Hence great challenge in the field of ebola virus research is to design universal vaccine. A combination of publicly available bioinformatics algorithms and computational tools are used to screen and select antigen sequences as potential T-cell epitopes of supertypes Human Leukocyte Antigen (HLA) alleles. MUSCLE and MOTIF tools were used to find out most conserved peptide sequences of viral proteins. Immunoinformatics tools were used for prediction of immunogenic peptides of viral proteins in zaire strains of Ebola virus. Putative epitopes for viral proteins (VP) were predicted from conserved peptide sequences of VP. Three tools NetCTL 1.2, BIMAS and Syfpeithi were used to predict the Class I putative epitopes while three tools, ProPred, IEDB-SMM-align and NetMHCII 2.2 were used to predict the Class II putative epitopes. B cell epitopes were predicted by BCPREDS 1.0. Immunogenic peptides were identified and selected manually by putative epitopes predicted from online tools individually for both MHC classes. Finally sequences of predicted peptides for both MHC classes were looked for common region which was selected as common immunogenic peptide. The immunogenic peptides were found for viral proteins of Ebola virus: epitopes FLESGAVKY, SSLAKHGEY. These predicted peptides could be promising candidates to be used as target for vaccine design.Keywords: epitope, b cell, immunogenicity, ebola
Procedia PDF Downloads 31427230 Hybrid Quasi-Steady Thermal Lattice Boltzmann Model for Studying the Behavior of Oil in Water Emulsions Used in Machining Tool Cooling and Lubrication
Authors: W. Hasan, H. Farhat, A. Alhilo, L. Tamimi
Abstract:
Oil in water (O/W) emulsions are utilized extensively for cooling and lubricating cutting tools during parts machining. A robust Lattice Boltzmann (LBM) thermal-surfactants model, which provides a useful platform for exploring complex emulsions’ characteristics under variety of flow conditions, is used here for the study of the fluid behavior during conventional tools cooling. The transient thermal capabilities of the model are employed for simulating the effects of the flow conditions of O/W emulsions on the cooling of cutting tools. The model results show that the temperature outcome is slightly affected by reversing the direction of upper plate (workpiece). On the other hand, an important increase in effective viscosity is seen which supports better lubrication during the work.Keywords: hybrid lattice Boltzmann method, Gunstensen model, thermal, surfactant-covered droplet, Marangoni stress
Procedia PDF Downloads 30327229 Leveraging Power BI for Advanced Geotechnical Data Analysis and Visualization in Mining Projects
Authors: Elaheh Talebi, Fariba Yavari, Lucy Philip, Lesley Town
Abstract:
The mining industry generates vast amounts of data, necessitating robust data management systems and advanced analytics tools to achieve better decision-making processes in the development of mining production and maintaining safety. This paper highlights the advantages of Power BI, a powerful intelligence tool, over traditional Excel-based approaches for effectively managing and harnessing mining data. Power BI enables professionals to connect and integrate multiple data sources, ensuring real-time access to up-to-date information. Its interactive visualizations and dashboards offer an intuitive interface for exploring and analyzing geotechnical data. Advanced analytics is a collection of data analysis techniques to improve decision-making. Leveraging some of the most complex techniques in data science, advanced analytics is used to do everything from detecting data errors and ensuring data accuracy to directing the development of future project phases. However, while Power BI is a robust tool, specific visualizations required by geotechnical engineers may have limitations. This paper studies the capability to use Python or R programming within the Power BI dashboard to enable advanced analytics, additional functionalities, and customized visualizations. This dashboard provides comprehensive tools for analyzing and visualizing key geotechnical data metrics, including spatial representation on maps, field and lab test results, and subsurface rock and soil characteristics. Advanced visualizations like borehole logs and Stereonet were implemented using Python programming within the Power BI dashboard, enhancing the understanding and communication of geotechnical information. Moreover, the dashboard's flexibility allows for the incorporation of additional data and visualizations based on the project scope and available data, such as pit design, rock fall analyses, rock mass characterization, and drone data. This further enhances the dashboard's usefulness in future projects, including operation, development, closure, and rehabilitation phases. Additionally, this helps in minimizing the necessity of utilizing multiple software programs in projects. This geotechnical dashboard in Power BI serves as a user-friendly solution for analyzing, visualizing, and communicating both new and historical geotechnical data, aiding in informed decision-making and efficient project management throughout various project stages. Its ability to generate dynamic reports and share them with clients in a collaborative manner further enhances decision-making processes and facilitates effective communication within geotechnical projects in the mining industry.Keywords: geotechnical data analysis, power BI, visualization, decision-making, mining industry
Procedia PDF Downloads 9227228 Comparison of Statistical Methods for Estimating Missing Precipitation Data in the River Subbasin Lenguazaque, Colombia
Authors: Miguel Cañon, Darwin Mena, Ivan Cabeza
Abstract:
In this work was compared and evaluated the applicability of statistical methods for the estimation of missing precipitations data in the basin of the river Lenguazaque located in the departments of Cundinamarca and Boyacá, Colombia. The methods used were the method of simple linear regression, distance rate, local averages, mean rates, correlation with nearly stations and multiple regression method. The analysis used to determine the effectiveness of the methods is performed by using three statistical tools, the correlation coefficient (r2), standard error of estimation and the test of agreement of Bland and Altmant. The analysis was performed using real rainfall values removed randomly in each of the seasons and then estimated using the methodologies mentioned to complete the missing data values. So it was determined that the methods with the highest performance and accuracy in the estimation of data according to conditions that were counted are the method of multiple regressions with three nearby stations and a random application scheme supported in the precipitation behavior of related data sets.Keywords: statistical comparison, precipitation data, river subbasin, Bland and Altmant
Procedia PDF Downloads 46727227 DeepOmics: Deep Learning for Understanding Genome Functioning and the Underlying Genetic Causes of Disease
Authors: Vishnu Pratap Singh Kirar, Madhuri Saxena
Abstract:
Advancement in sequence data generation technologies is churning out voluminous omics data and posing a massive challenge to annotate the biological functional features. With so much data available, the use of machine learning methods and tools to make novel inferences has become obvious. Machine learning methods have been successfully applied to a lot of disciplines, including computational biology and bioinformatics. Researchers in computational biology are interested to develop novel machine learning frameworks to classify the huge amounts of biological data. In this proposal, it plan to employ novel machine learning approaches to aid the understanding of how apparently innocuous mutations (in intergenic DNA and at synonymous sites) cause diseases. We are also interested in discovering novel functional sites in the genome and mutations in which can affect a phenotype of interest.Keywords: genome wide association studies (GWAS), next generation sequencing (NGS), deep learning, omics
Procedia PDF Downloads 9727226 Streamlining the Fuzzy Front-End and Improving the Usability of the Tools Involved
Authors: Michael N. O'Sullivan, Con Sheahan
Abstract:
Researchers have spent decades developing tools and techniques to aid teams in the new product development (NPD) process. Despite this, it is evident that there is a huge gap between their academic prevalence and their industry adoption. For the fuzzy front-end, in particular, there is a wide range of tools to choose from, including the Kano Model, the House of Quality, and many others. In fact, there are so many tools that it can often be difficult for teams to know which ones to use and how they interact with one another. Moreover, while the benefits of using these tools are obvious to industrialists, they are rarely used as they carry a learning curve that is too steep and they become too complex to manage over time. In essence, it is commonly believed that they are simply not worth the effort required to learn and use them. This research explores a streamlined process for the fuzzy front-end, assembling the most effective tools and making them accessible to everyone. The process was developed iteratively over the course of 3 years, following over 80 final year NPD teams from engineering, design, technology, and construction as they carried a product from concept through to production specification. Questionnaires, focus groups, and observations were used to understand the usability issues with the tools involved, and a human-centred design approach was adopted to produce a solution to these issues. The solution takes the form of physical toolkit, similar to a board game, which allows the team to play through an example of a new product development in order to understand the process and the tools, before using it for their own product development efforts. A complimentary website is used to enhance the physical toolkit, and it provides more examples of the tools being used, as well as deeper discussions on each of the topics, allowing teams to adapt the process to their skills, preferences and product type. Teams found the solution very useful and intuitive and experienced significantly less confusion and mistakes with the process than teams who did not use it. Those with a design background found it especially useful for the engineering principles like Quality Function Deployment, while those with an engineering or technology background found it especially useful for design and customer requirements acquisition principles, like Voice of the Customer. Products developed using the toolkit are added to the website as more examples of how it can be used, creating a loop which helps future teams understand how the toolkit can be adapted to their project, whether it be a small consumer product or a large B2B service. The toolkit unlocks the potential of these beneficial tools to those in industry, both for large, experienced teams and for inexperienced start-ups. It allows users to assess the market potential of their product concept faster and more effectively, arriving at the product design stage with technical requirements prioritized according to their customers’ needs and wants.Keywords: new product development, fuzzy front-end, usability, Kano model, quality function deployment, voice of customer
Procedia PDF Downloads 10827225 Implementation of Lean Manufacturing in Some Companies in Colombia: A Case Study
Authors: Natalia Marulanda, Henry González, Gonzalo León, Alejandro Hincapié
Abstract:
Continuous improvement tools are the result of a set of studies that developed theories and methodologies. These methodologies enable organizations to increase their levels of efficiency, effectiveness, and productivity. Based on these methodologies, lean manufacturing philosophy, which is based on the optimization of resources, waste disposal, and generation of value to products and services, was developed. Lean application has been massive globally, but Colombian companies have been made it incipiently. Therefore, the purpose of this article is to identify the impacts generated by the implementation of lean manufacturing tools in five companies located in Colombia and Medellín metropolitan area. It also seeks to make a comparison of the results obtained from the implementation of lean philosophy and Theory of Constraints. The methodology is qualitative and quantitative, is based on the case study interview from dialogue with the leaders of the processes that used lean tools. The most used tools by research companies are 5's with 100% and TPM with 80%. The less used tool is the synchronous production with 20%. The main reason for the implementation of lean was supply chain management with 83.3%. For the application of lean and TOC, we did not find significant differences between the impact, in terms of methodology, areas of application, staff initiatives, supply chain management, planning, and training.Keywords: business strategy, lean manufacturing, theory of constraints, supply chain
Procedia PDF Downloads 35427224 Improving Overall Equipment Effectiveness of CNC-VMC by Implementing Kobetsu Kaizen
Authors: Nakul Agrawal, Y. M. Puri
Abstract:
TPM methodology is a proven approach to increase Overall Equipment Effectiveness (OEE) of machine. OEE is an established method to monitor and improve the effectiveness of manufacturing process. OEE is a product of equipment availability, performance efficiency and quality performance of manufacturing operations. The paper presents a project work for improving OEE of CNC-VMC in a manufacturing industry with the help of TPM tools Kaizen and Autonomous Maintenance. The aim of paper is to enhance OEE by minimizing the breakdown and re-work, increase availability, performance and quality. The calculated OEE of bottle necking machines for 4 months is lower of 53.3%. Root Cause Analysis RCA tools like fishbone diagram, Pareto chart are used for determining the reasons behind low OEE. While Tool like Why-Why analysis is use for determining the basis reasons for low OEE. Tools like Kaizen and Autonomous Maintenance are effectively implemented on CNC-VMC which eliminate the causes of breakdown and prevent from reoccurring. The result obtains from approach shows that OEE of CNC-VMC improved from 53.3% to 73.7% which saves an average sum of Rs.3, 19,000.Keywords: OEE, TPM, Kaizen, CNC-VMC, why-why analysis, RCA
Procedia PDF Downloads 39427223 Institutional Engineering and Party Politics in Nigeria’s Fourth Republic
Authors: Emmanuel Ayobami Adesiyan
Abstract:
Political theorists have identified ethnicity as an obstacle to democratic stability in deeply divided societies. Nigeria belongs to the categories of problematic states labeled divided or deeply divided societies, as such post-independence politics is characterized by ethnicity with its ruinous effect on democratic governance and development. Institutional Engineering, the purposive manipulation of the electoral rule relating to party organization and the electoral formula has been established in comparative political studies as a policy measure for managing ethnicity in order to stabilize politics in divided societies. This paper examines the use of electoral engineering tools in managing ethnic politics in Nigeria’s Fourth Republic. The study is guided by rational institutional theory. Secondary data on electoral rules and disaggregated results of presidential elections were collected from archival documents. Data were subjected to content analysis. Institutional changes in electoral rules have promoted the development of inter-ethnic bargaining and compromises within the party system. Presidential Electoral Formula aided the emergence of national rather parochial parties. Electoral engineering tools moved Nigerian Politics from ethnic parochialism to inclusion and accommodation. These innovations should be strengthened to enhance democratic stability.Keywords: Nigeria, presidential-elections, ethnic politics, institutional engineering
Procedia PDF Downloads 23627222 Usage of Visual Tools for Light Exploring with Children in the Geographical Istria Region Kindergartens in Republic of Croatia and Republic of Slovenia
Authors: Urianni Merlin, Đeni Zuliani Blašković
Abstract:
Inspired by the Reggio Pedagogy approach that explores light from physical, mathematical, artistic, and natural perspectives, emphasizes the value of visual tools in light exploring that opens up a wide area of experiential discovery and knowledge, especially if used in kindergartens with children. While there is some literature evidence of visual tool usage for light exploring in kindergartens in the Republic of Slovenia, in the Republic of Croatia there are few researches, and those published are focused at shadow exploring, exploring of physical characteristics and teatrical play of light and shadow. The objectives of this research are to assess how much visual tools are used for light exploring by preschool teachers from geographical Istria kindergartens as part of the activities offered to children and if the usage of the visual tool for light exploring it’s different regarding the work environment (Slovenian and Croatian Istria kindergartens; city vs. village kindergartens; preschool teachers age and length of service). One hundred one preschool teachers from Croatian Istria Region and 70 preschool teachers from Slovenian Istria Region responded to a self-made questionnaire regarding visual tool usage habits in their work. As predicted, results show significant differences in visual tool usage regarding preschool teachers' work environment, length of service, and age. Preschool teachers from Slovenian Istria that work in kindergartens located in the city that have from 15 to 19 years of service and are more than 30 years of age use significantly more visual tools for light exploring. The results highlight the differences in visual tools usage for light exploring in the small Istria peninsula that can be attributed to different University art curricula in Slovenia and Croatia or lifelong education offered in Slovenia that is more open to Italian reggio pedagogy influence and are further used by older preschool teachers with more service experience. Considering the small number of researches, this research significantly contributes to science and motivates preschool teachers and scientists to implement the use of light tools in the preschool and university curriculum, especially in Croatia.Keywords: activities with light, light exploring, preschool children, visual tools
Procedia PDF Downloads 7827221 The Development of an Automated Computational Workflow to Prioritize Potential Resistance Variants in HIV Integrase Subtype C
Authors: Keaghan Brown
Abstract:
The prioritization of drug resistance mutations impacting protein folding or protein-drug and protein-DNA interactions within macromolecular systems is critical to the success of treatment regimens. With a continual increase in computational tools to assess these impacts, the need for scalability and reproducibility became an essential component of computational analysis and experimental research. Here it introduce a bioinformatics pipeline that combines several structural analysis tools in a simplified workflow, by optimizing the present computational hardware and software to automatically ease the flow of data transformations. Utilizing preestablished software tools, it was possible to develop a pipeline with a set of pre-defined functions that will automate mutation introduction into the HIV-1 Integrase protein structure, calculate the gain and loss of polar interactions and calculate the change in energy of protein fold. Additionally, an automated molecular dynamics analysis was implemented which reduces the constant need for user input and output management. The resulting pipeline, Automated Mutation Introduction and Analysis (AMIA) is an open source set of scripts designed to introduce and analyse the effects of mutations on the static protein structure as well as the results of the multi-conformational states from molecular dynamic simulations. The workflow allows the user to visualize all outputs in a user friendly manner thereby successfully enabling the prioritization of variant systems for experimental validation.Keywords: automated workflow, variant prioritization, drug resistance, HIV Integrase
Procedia PDF Downloads 77