Search results for: missing data estimation
23414 Data Confidentiality in Public Cloud: A Method for Inclusion of ID-PKC Schemes in OpenStack Cloud
Authors: N. Nalini, Bhanu Prakash Gopularam
Abstract:
The term data security refers to the degree of resistance or protection given to information from unintended or unauthorized access. The core principles of information security are the confidentiality, integrity and availability, also referred as CIA triad. Cloud computing services are classified as SaaS, IaaS and PaaS services. With cloud adoption the confidential enterprise data are moved from organization premises to untrusted public network and due to this the attack surface has increased manifold. Several cloud computing platforms like OpenStack, Eucalyptus, Amazon EC2 offer users to build and configure public, hybrid and private clouds. While the traditional encryption based on PKI infrastructure still works in cloud scenario, the management of public-private keys and trust certificates is difficult. The Identity based Public Key Cryptography (also referred as ID-PKC) overcomes this problem by using publicly identifiable information for generating the keys and works well with decentralized systems. The users can exchange information securely without having to manage any trust information. Another advantage is that access control (role based access control policy) information can be embedded into data unlike in PKI where it is handled by separate component or system. In OpenStack cloud platform the keystone service acts as identity service for authentication and authorization and has support for public key infrastructure for auto services. In this paper, we explain OpenStack security architecture and evaluate the PKI infrastructure piece for data confidentiality. We provide method to integrate ID-PKC schemes for securing data while in transit and stored and explain the key measures for safe guarding data against security attacks. The proposed approach uses JPBC crypto library for key-pair generation based on IEEE P1636.3 standard and secure communication to other cloud services.Keywords: data confidentiality, identity based cryptography, secure communication, open stack key stone, token scoping
Procedia PDF Downloads 38523413 First Experimental Evidence on Feasibility of Molecular Magnetic Particle Imaging of Tumor Marker Alpha-1-Fetoprotein Using Antibody Conjugated Nanoparticles
Authors: Kolja Them, Priyal Chikhaliwala, Sudeshna Chandra
Abstract:
Purpose: The purpose of this work is to examine possibilities for noninvasive imaging and identification of tumor markers for cancer diagnosis. The proposed method uses antibody conjugated iron oxide nanoparticles and multicolor Magnetic Particle Imaging (mMPI). The method has the potential for radiation exposure free real-time estimation of local tumor marker concentrations in vivo. In this study, the method is applied to human Alpha-1-Fetoprotein. Materials and Methods: As tracer material AFP antibody-conjugated Dendrimer-Fe3O4 nanoparticles were used. The nanoparticle bioconjugates were then incubated with bovine serum albumin (BSA) to block any possible nonspecific binding sites. Parts of the resulting solution were then incubated with AFP antigen. MPI measurements were done using the preclinical MPI scanner (Bruker Biospin MRI GmbH) and the multicolor method was used for image reconstruction. Results: In multicolor MPI images the nanoparticles incubated only with BSA were clearly distinguished from nanoparticles incubated with BSA and AFP antigens. Conclusion: Tomographic imaging of human tumor marker Alpha-1-Fetoprotein is possible using AFP antibody conjugated iron oxide nanoparticles in presence of BSA. This opens interesting perspectives for cancer diagnosis.Keywords: noninvasive imaging, tumor antigens, antibody conjugated iron oxide nanoparticles, multicolor magnetic particle imaging, cancer diagnosis
Procedia PDF Downloads 30523412 Exploring RQ-EQ Relatons among Psychology Majors
Authors: Maria T. Mamba, Febe Marl G. Paat
Abstract:
The illustrious estimation that psychology majors, psychologists and allied psychology practitioners as expert behavior analysts, if not, “life enthusiasts” spurred two essentially linked endeavors. First, the reconsideration of the time-honored ingenuity and expectations from psychologists such as the ability to perceive ways to undertake a range of difficulties, the ability to apply psychology in order to self-regulate and to display personal integrity, and among others. Second, is to ascertain solid support to uphold aforesaid expectations. This study achieved its goals by having explored how two burgeoning constructs- RQ and EQ play parts in the lives of psychology people. Having involved the total population of psychology majors in Cagayan State University along with the use of Emotional Quotient Test and Resilience Assessment Questionnaire, the study provides a précis of how perceived “champions” of psychological well-being respond emotionally to different situations and deal effectively with and even thrive on the demands of frequently changing environmental circumstances. Significant findings about how the major variables correlated with the population’s demographic profile (e.g. age, sex, and year level) were also accounted. To realize a more academic concept with the present study, significant connections between RQ (self-assurance, personal vision, flexible and adaptable, organized, problem solver, interpersonal competence, socially connected, and active) and EQ (e.g. emotional maturity, emotional sensitivity, and emotional competency) dimensions were uncovered.Keywords: emotional quotient, resilience quotient, psychology majors, exploring
Procedia PDF Downloads 45123411 A User Identification Technique to Access Big Data Using Cloud Services
Authors: A. R. Manu, V. K. Agrawal, K. N. Balasubramanya Murthy
Abstract:
Authentication is required in stored database systems so that only authorized users can access the data and related cloud infrastructures. This paper proposes an authentication technique using multi-factor and multi-dimensional authentication system with multi-level security. The proposed technique is likely to be more robust as the probability of breaking the password is extremely low. This framework uses a multi-modal biometric approach and SMS to enforce additional security measures with the conventional Login/password system. The robustness of the technique is demonstrated mathematically using a statistical analysis. This work presents the authentication system along with the user authentication architecture diagram, activity diagrams, data flow diagrams, sequence diagrams, and algorithms.Keywords: design, implementation algorithms, performance, biometric approach
Procedia PDF Downloads 47823410 Recovery of the Demolition and Construction Waste, Casablanca (Morocco)
Authors: Morsli Mourad, Tahiri Mohamed, Samdi Azzeddine
Abstract:
Casablanca is the biggest city in Morocco. It concentrates more than 60% of the economic and industrial activity of the kingdom. Its building and public works (BTP) sector is the leading source of inert waste scattered in open areas. This inert waste is a major challenge for the city of Casablanca, as it is not properly managed, thus causing a significant nuisance for the environment and the health of the population. Hence the vision of our project is to recycle and valorize concrete waste. In this work, we present concrete results in the exploitation of this abundant and permanent deposit. Typical wastes are concrete, clay and concrete bricks, ceramic tiles, marble panels, gypsum, scrap metal, wood . The work performed included: geolocation with a combination of artificial intelligence and Google Earth, estimation of the amount of waste per site, sorting, crushing, grinding, and physicochemical characterization of the samples. Then, we proceeded to the exploitation of the types of substrates to be developed: light cement, coating, and glue for ceramics... The said products were tested and characterized by X-ray fluorescence, specific surface, resistance to bending and crushing, etc. We will present in detail the main results of our research work and also describe the specific properties of each material developed.Keywords: déchets de démolition et des chantiers de construction, logiciels de combinaison SIG, valorisation de déchets inertes, enduits, ciment leger, casablanca
Procedia PDF Downloads 11523409 Input Data Balancing in a Neural Network PM-10 Forecasting System
Authors: Suk-Hyun Yu, Heeyong Kwon
Abstract:
Recently PM-10 has become a social and global issue. It is one of major air pollutants which affect human health. Therefore, it needs to be forecasted rapidly and precisely. However, PM-10 comes from various emission sources, and its level of concentration is largely dependent on meteorological and geographical factors of local and global region, so the forecasting of PM-10 concentration is very difficult. Neural network model can be used in the case. But, there are few cases of high concentration PM-10. It makes the learning of the neural network model difficult. In this paper, we suggest a simple input balancing method when the data distribution is uneven. It is based on the probability of appearance of the data. Experimental results show that the input balancing makes the neural networks’ learning easy and improves the forecasting rates.Keywords: artificial intelligence, air quality prediction, neural networks, pattern recognition, PM-10
Procedia PDF Downloads 23323408 Metabolic Predictive Model for PMV Control Based on Deep Learning
Authors: Eunji Choi, Borang Park, Youngjae Choi, Jinwoo Moon
Abstract:
In this study, a predictive model for estimating the metabolism (MET) of human body was developed for the optimal control of indoor thermal environment. Human body images for indoor activities and human body joint coordinated values were collected as data sets, which are used in predictive model. A deep learning algorithm was used in an initial model, and its number of hidden layers and hidden neurons were optimized. Lastly, the model prediction performance was analyzed after the model being trained through collected data. In conclusion, the possibility of MET prediction was confirmed, and the direction of the future study was proposed as developing various data and the predictive model.Keywords: deep learning, indoor quality, metabolism, predictive model
Procedia PDF Downloads 25923407 Adsorption and Electrochemical Regeneration for Industrial Wastewater Treatment
Authors: H. M. Mohammad, A. Martin, N. Brown, N. Hodson, P. Hill, E. Roberts
Abstract:
Graphite intercalation compound (GIC) has been demonstrated to be a useful, low capacity and rapid adsorbent for the removal of organic micropollutants from water. The high electrical conductivity and low capacity of the material lends itself to electrochemical regeneration. Following electrochemical regeneration, equilibrium loading under similar conditions is reported to exceed that achieved by the fresh adsorbent. This behavior is reported in terms of the regeneration efficiency being greater than 100%. In this work, surface analysis techniques are employed to investigate the material in three states: ‘Fresh’, ‘Loaded’ and ‘Regenerated’. ‘Fresh’ GIC is shown to exhibit a hydrogen and oxygen rich surface layer approximately 150 nm thick. ‘Loaded’ GIC shows a similar but slightly thicker surface layer (approximately 370 nm thick) and significant enhancement in the hydrogen and oxygen abundance extending beyond 600 nm from the surface. 'Regenerated’ GIC shows an oxygen rich layer, slightly thicker than the fresh case at approximately 220 nm while showing a very much lower hydrogen enrichment at the surface. Results demonstrate that while the electrochemical regeneration effectively removes the phenol model pollutant, it also oxidizes the exposed carbon surface. These results may have a significant impact on the estimation of adsorbent life.Keywords: graphite, adsorbent, electrochemical, regeneration, phenol
Procedia PDF Downloads 14023406 Analysis of Brownfield Soil Contamination Using Local Government Planning Data
Authors: Emma E. Hellawell, Susan J. Hughes
Abstract:
BBrownfield sites are currently being redeveloped for residential use. Information on soil contamination on these former industrial sites is collected as part of the planning process by the local government. This research project analyses this untapped resource of environmental data, using site investigation data submitted to a local Borough Council, in Surrey, UK. Over 150 site investigation reports were collected and interrogated to extract relevant information. This study involved three phases. Phase 1 was the development of a database for soil contamination information from local government reports. This database contained information on the source, history, and quality of the data together with the chemical information on the soil that was sampled. Phase 2 involved obtaining site investigation reports for development within the study area and extracting the required information for the database. Phase 3 was the data analysis and interpretation of key contaminants to evaluate typical levels of contaminants, their distribution within the study area, and relating these results to current guideline levels of risk for future site users. Preliminary results for a pilot study using a sample of the dataset have been obtained. This pilot study showed there is some inconsistency in the quality of the reports and measured data, and careful interpretation of the data is required. Analysis of the information has found high levels of lead in shallow soil samples, with mean and median levels exceeding the current guidance for residential use. The data also showed elevated (but below guidance) levels of potentially carcinogenic polyaromatic hydrocarbons. Of particular concern from the data was the high detection rate for asbestos fibers. These were found at low concentrations in 25% of the soil samples tested (however, the sample set was small). Contamination levels of the remaining chemicals tested were all below the guidance level for residential site use. These preliminary pilot study results will be expanded, and results for the whole local government area will be presented at the conference. The pilot study has demonstrated the potential for this extensive dataset to provide greater information on local contamination levels. This can help inform regulators and developers and lead to more targeted site investigations, improving risk assessments, and brownfield development.Keywords: Brownfield development, contaminated land, local government planning data, site investigation
Procedia PDF Downloads 14123405 Carbon Footprint Assessment Initiative and Trees: Role in Reducing Emissions
Authors: Omar Alelweet
Abstract:
Carbon emissions are quantified in terms of carbon dioxide equivalents, generated through a specific activity or accumulated throughout the life stages of a product or service. Given the growing concern about climate change and the role of carbon dioxide emissions in global warming, this initiative aims to create awareness and understanding of the impact of human activities and identify potential areas for improvement regarding the management of the carbon footprint on campus. Given that trees play a vital role in reducing carbon emissions by absorbing CO₂ during the photosynthesis process, this paper evaluated the contribution of each tree to reducing those emissions. Collecting data over an extended period of time is essential to monitoring carbon dioxide levels. This will help capture changes at different times and identify any patterns or trends in the data. By linking the data to specific activities, events, or environmental factors, it is possible to identify sources of emissions and areas where carbon dioxide levels are rising. Analyzing the collected data can provide valuable insights into ways to reduce emissions and mitigate the impact of climate change.Keywords: sustainability, green building, environmental impact, CO₂
Procedia PDF Downloads 7223404 Detection of Change Points in Earthquakes Data: A Bayesian Approach
Authors: F. A. Al-Awadhi, D. Al-Hulail
Abstract:
In this study, we applied the Bayesian hierarchical model to detect single and multiple change points for daily earthquake body wave magnitude. The change point analysis is used in both backward (off-line) and forward (on-line) statistical research. In this study, it is used with the backward approach. Different types of change parameters are considered (mean, variance or both). The posterior model and the conditional distributions for single and multiple change points are derived and implemented using BUGS software. The model is applicable for any set of data. The sensitivity of the model is tested using different prior and likelihood functions. Using Mb data, we concluded that during January 2002 and December 2003, three changes occurred in the mean magnitude of Mb in Kuwait and its vicinity.Keywords: multiple change points, Markov Chain Monte Carlo, earthquake magnitude, hierarchical Bayesian mode
Procedia PDF Downloads 45923403 Blood Volume Pulse Extraction for Non-Contact Photoplethysmography Measurement from Facial Images
Authors: Ki Moo Lim, Iman R. Tayibnapis
Abstract:
According to WHO estimation, 38 out of 56 million (68%) global deaths in 2012, were due to noncommunicable diseases (NCDs). To avert NCD, one of the solutions is early detection of diseases. In order to do that, we developed 'U-Healthcare Mirror', which is able to measure vital sign such as heart rate (HR) and respiration rate without any physical contact and consciousness. To measure HR in the mirror, we utilized digital camera. The camera records red, green, and blue (RGB) discoloration from user's facial image sequences. We extracted blood volume pulse (BVP) from the RGB discoloration because the discoloration of the facial skin is accordance with BVP. We used blind source separation (BSS) to extract BVP from the RGB discoloration and adaptive filters for removing noises. We utilized singular value decomposition (SVD) method to implement the BSS and the adaptive filters. HR was estimated from the obtained BVP. We did experiment for HR measurement by using our method and previous method that used independent component analysis (ICA) method. We compared both of them with HR measurement from commercial oximeter. The experiment was conducted under various distance between 30~110 cm and light intensity between 5~2000 lux. For each condition, we did measurement 7 times. The estimated HR showed 2.25 bpm of mean error and 0.73 of pearson correlation coefficient. The accuracy has improved compared to previous work. The optimal distance between the mirror and user for HR measurement was 50 cm with medium light intensity, around 550 lux.Keywords: blood volume pulse, heart rate, photoplethysmography, independent component analysis
Procedia PDF Downloads 33023402 Non Interferometric Quantitative Phase Imaging of Yeast Cells
Authors: P. Praveen Kumar, P. Vimal Prabhu, Renu John
Abstract:
In biology most microscopy specimens, in particular living cells are transparent. In cell imaging, it is hard to create an image of a cell which is transparent with a very small refractive index change with respect to the surrounding media. Various techniques like addition of staining and contrast agents, markers have been applied in the past for creating contrast. Many of the staining agents or markers are not applicable to live cell imaging as they are toxic. In this paper, we report theoretical and experimental results from quantitative phase imaging of yeast cells with a commercial bright field microscope. We reconstruct the phase of cells non-interferometrically based on the transport of intensity equations (TIE). This technique estimates the axial derivative from positive through-focus intensity measurements. This technique allows phase imaging using a regular microscope with white light illumination. We demonstrate nano-metric depth sensitivity in imaging live yeast cells using this technique. Experimental results will be shown in the paper demonstrating the capability of the technique in 3-D volume estimation of living cells. This real-time imaging technique would be highly promising in real-time digital pathology applications, screening of pathogens and staging of diseases like malaria as it does not need any pre-processing of samples.Keywords: axial derivative, non-interferometric imaging, quantitative phase imaging, transport of intensity equation
Procedia PDF Downloads 38523401 Productivity and Structural Design of Manufacturing Systems
Authors: Ryspek Usubamatov, Tan San Chin, Sarken Kapaeva
Abstract:
Productivity of the manufacturing systems depends on technological processes, a technical data of machines and a structure of systems. Technology is presented by the machining mode and data, a technical data presents reliability parameters and auxiliary time for discrete production processes. The term structure of manufacturing systems includes the number of serial and parallel production machines and links between them. Structures of manufacturing systems depend on the complexity of technological processes. Mathematical models of productivity rate for manufacturing systems are important attributes that enable to define best structure by criterion of a productivity rate. These models are important tool in evaluation of the economical efficiency for production systems.Keywords: productivity, structure, manufacturing systems, structural design
Procedia PDF Downloads 58523400 The Effect of Tacit Knowledge for Intelligence Cycle
Authors: Bahadir Aydin
Abstract:
It is difficult to access accurate knowledge because of mass data. This huge data make environment more and more caotic. Data are main piller of intelligence. The affiliation between intelligence and knowledge is quite significant to understand underlying truths. The data gathered from different sources can be modified, interpreted and classified by using intelligence cycle process. This process is applied in order to progress to wisdom as well as intelligence. Within this process the effect of tacit knowledge is crucial. Knowledge which is classified as explicit and tacit knowledge is the key element for any purpose. Tacit knowledge can be seen as "the tip of the iceberg”. This tacit knowledge accounts for much more than we guess in all intelligence cycle. If the concept of intelligence cycle is scrutinized, it can be seen that it contains risks, threats as well as success. The main purpose of all organizations is to be successful by eliminating risks and threats. Therefore, there is a need to connect or fuse existing information and the processes which can be used to develop it. Thanks to this process the decision-makers can be presented with a clear holistic understanding, as early as possible in the decision making process. Altering from the current traditional reactive approach to a proactive intelligence cycle approach would reduce extensive duplication of work in the organization. Applying new result-oriented cycle and tacit knowledge intelligence can be procured and utilized more effectively and timely.Keywords: information, intelligence cycle, knowledge, tacit Knowledge
Procedia PDF Downloads 51523399 Evaluation of the CRISP-DM Business Understanding Step: An Approach for Assessing the Predictive Power of Regression versus Classification for the Quality Prediction of Hydraulic Test Results
Authors: Christian Neunzig, Simon Fahle, Jürgen Schulz, Matthias Möller, Bernd Kuhlenkötter
Abstract:
Digitalisation in production technology is a driver for the application of machine learning methods. Through the application of predictive quality, the great potential for saving necessary quality control can be exploited through the data-based prediction of product quality and states. However, the serial use of machine learning applications is often prevented by various problems. Fluctuations occur in real production data sets, which are reflected in trends and systematic shifts over time. To counteract these problems, data preprocessing includes rule-based data cleaning, the application of dimensionality reduction techniques, and the identification of comparable data subsets to extract stable features. Successful process control of the target variables aims to centre the measured values around a mean and minimise variance. Competitive leaders claim to have mastered their processes. As a result, much of the real data has a relatively low variance. For the training of prediction models, the highest possible generalisability is required, which is at least made more difficult by this data availability. The implementation of a machine learning application can be interpreted as a production process. The CRoss Industry Standard Process for Data Mining (CRISP-DM) is a process model with six phases that describes the life cycle of data science. As in any process, the costs to eliminate errors increase significantly with each advancing process phase. For the quality prediction of hydraulic test steps of directional control valves, the question arises in the initial phase whether a regression or a classification is more suitable. In the context of this work, the initial phase of the CRISP-DM, the business understanding, is critically compared for the use case at Bosch Rexroth with regard to regression and classification. The use of cross-process production data along the value chain of hydraulic valves is a promising approach to predict the quality characteristics of workpieces. Suitable methods for leakage volume flow regression and classification for inspection decision are applied. Impressively, classification is clearly superior to regression and achieves promising accuracies.Keywords: classification, CRISP-DM, machine learning, predictive quality, regression
Procedia PDF Downloads 14623398 Implementation Association Rule Method in Determining the Layout of Qita Supermarket as a Strategy in the Competitive Retail Industry in Indonesia
Authors: Dwipa Rizki Utama, Hanief Ibrahim
Abstract:
The development of industry retail in Indonesia is very fast, various strategy was undertaken to boost the customer satisfaction and the productivity purchases to boost the profit, one of which is implementing strategies layout. The purpose of this study is to determine the layout of Qita supermarket, a retail industry in Indonesia, in order to improve customer satisfaction and to maximize the rate of products’ sale as a whole, so as the infrequently purchased products will be purchased. This research uses a literature study method, and one of the data mining methods is association rule which applied in market basket analysis. Data were tested amounted 100 from 160 after pre-processing data, so then the distribution department and 26 departments corresponding to the data previous layout will be obtained. From those data, by the association rule method, customer behavior when purchasing items simultaneously can be studied, so then the layout of the supermarket based on customer behavior can be determined. Using the rapid miner software by the minimal support 25% and minimal confidence 30% showed that the 14th department purchased at the same time with department 10, 21st department purchased at the same time with department 13, 15th department purchased at the same time with department 12, 14th department purchased at the same time with department 12, and 10th department purchased at the same time with department 14. From those results, a better supermarket layout can be arranged than the previous layout.Keywords: industry retail, strategy, association rule, supermarket
Procedia PDF Downloads 19023397 Preparing Data for Calibration of Mechanistic-Empirical Pavement Design Guide in Central Saudi Arabia
Authors: Abdulraaof H. Alqaili, Hamad A. Alsoliman
Abstract:
Through progress in pavement design developments, a pavement design method was developed, which is titled the Mechanistic Empirical Pavement Design Guide (MEPDG). Nowadays, the evolution in roads network and highways is observed in Saudi Arabia as a result of increasing in traffic volume. Therefore, the MEPDG currently is implemented for flexible pavement design by the Saudi Ministry of Transportation. Implementation of MEPDG for local pavement design requires the calibration of distress models under the local conditions (traffic, climate, and materials). This paper aims to prepare data for calibration of MEPDG in Central Saudi Arabia. Thus, the first goal is data collection for the design of flexible pavement from the local conditions of the Riyadh region. Since, the modifying of collected data to input data is needed; the main goal of this paper is the analysis of collected data. The data analysis in this paper includes processing each: Trucks Classification, Traffic Growth Factor, Annual Average Daily Truck Traffic (AADTT), Monthly Adjustment Factors (MAFi), Vehicle Class Distribution (VCD), Truck Hourly Distribution Factors, Axle Load Distribution Factors (ALDF), Number of axle types (single, tandem, and tridem) per truck class, cloud cover percent, and road sections selected for the local calibration. Detailed descriptions of input parameters are explained in this paper, which leads to providing of an approach for successful implementation of MEPDG. Local calibration of MEPDG to the conditions of Riyadh region can be performed based on the findings in this paper.Keywords: mechanistic-empirical pavement design guide (MEPDG), traffic characteristics, materials properties, climate, Riyadh
Procedia PDF Downloads 22623396 The Impact of Temperature on the Threshold Capillary Pressure of Fine-Grained Shales
Authors: Talal Al-Bazali, S. Mohammad
Abstract:
The threshold capillary pressure of shale caprocks is an important parameter in CO₂ storage modeling. A correct estimation of the threshold capillary pressure is not only essential for CO₂ storage modeling but also important to assess the overall economical and environmental impact of the design process. A standard step by step approach has to be used to measure the threshold capillary pressure of shale and non-wetting fluids at different temperatures. The objective of this work is to assess the impact of high temperature on the threshold capillary pressure of four different shales as they interacted with four different oil based muds, air, CO₂, N₂, and methane. This study shows that the threshold capillary pressure of shale and non-wetting fluid is highly impacted by temperature. An empirical correlation for the dependence of threshold capillary pressure on temperature when different shales interacted with oil based muds and gasses has been developed. This correlation shows that the threshold capillary pressure decreases exponentially as the temperature increases. In this correlation, an experimental constant (α) appears, and this constant may depend on the properties of shale and non-wetting fluid. The value for α factor was found to be higher for gasses than for oil based muds. This is consistent with our intuition since the interfacial tension for gasses is higher than those for oil based muds. The author believes that measured threshold capillary pressure at ambient temperature is misleading and could yield higher values than those encountered at in situ conditions. Therefore one must correct for the impact of temperature when measuring threshold capillary pressure of shale at ambient temperature.Keywords: capillary pressure, shale, temperature, thresshold
Procedia PDF Downloads 37323395 Transforming Data Science Curriculum Through Design Thinking
Authors: Samar Swaid
Abstract:
Today, corporates are moving toward the adoption of Design-Thinking techniques to develop products and services, putting their consumer as the heart of the development process. One of the leading companies in Design-Thinking, IDEO (Innovation, Design, Engineering Organization), defines Design-Thinking as an approach to problem-solving that relies on a set of multi-layered skills, processes, and mindsets that help people generate novel solutions to problems. Design thinking may result in new ideas, narratives, objects or systems. It is about redesigning systems, organizations, infrastructures, processes, and solutions in an innovative fashion based on the users' feedback. Tim Brown, president and CEO of IDEO, sees design thinking as a human-centered approach that draws from the designer's toolkit to integrate people's needs, innovative technologies, and business requirements. The application of design thinking has been witnessed to be the road to developing innovative applications, interactive systems, scientific software, healthcare application, and even to utilizing Design-Thinking to re-think business operations, as in the case of Airbnb. Recently, there has been a movement to apply design thinking to machine learning and artificial intelligence to ensure creating the "wow" effect on consumers. The Association of Computing Machinery task force on Data Science program states that" Data scientists should be able to implement and understand algorithms for data collection and analysis. They should understand the time and space considerations of algorithms. They should follow good design principles developing software, understanding the importance of those principles for testability and maintainability" However, this definition hides the user behind the machine who works on data preparation, algorithm selection and model interpretation. Thus, the Data Science program includes design thinking to ensure meeting the user demands, generating more usable machine learning tools, and developing ways of framing computational thinking. Here, describe the fundamentals of Design-Thinking and teaching modules for data science programs.Keywords: data science, design thinking, AI, currculum, transformation
Procedia PDF Downloads 8323394 Economized Sensor Data Processing with Vehicle Platooning
Authors: Henry Hexmoor, Kailash Yelasani
Abstract:
We present vehicular platooning as a special case of crowd-sensing framework where sharing sensory information among a crowd is used for their collective benefit. After offering an abstract policy that governs processes involving a vehicular platoon, we review several common scenarios and components surrounding vehicular platooning. We then present a simulated prototype that illustrates efficiency of road usage and vehicle travel time derived from platooning. We have argued that one of the paramount benefits of platooning that is overlooked elsewhere, is the substantial computational savings (i.e., economizing benefits) in acquisition and processing of sensory data among vehicles sharing the road. The most capable vehicle can share data gathered from its sensors with nearby vehicles grouped into a platoon.Keywords: cloud network, collaboration, internet of things, social network
Procedia PDF Downloads 19523393 Exchange Rate Forecasting by Econometric Models
Authors: Zahid Ahmad, Nosheen Imran, Nauman Ali, Farah Amir
Abstract:
The objective of the study is to forecast the US Dollar and Pak Rupee exchange rate by using time series models. For this purpose, daily exchange rates of US and Pakistan for the period of January 01, 2007 - June 2, 2017, are employed. The data set is divided into in sample and out of sample data set where in-sample data are used to estimate as well as forecast the models, whereas out-of-sample data set is exercised to forecast the exchange rate. The ADF test and PP test are used to make the time series stationary. To forecast the exchange rate ARIMA model and GARCH model are applied. Among the different Autoregressive Integrated Moving Average (ARIMA) models best model is selected on the basis of selection criteria. Due to the volatility clustering and ARCH effect the GARCH (1, 1) is also applied. Results of analysis showed that ARIMA (0, 1, 1 ) and GARCH (1, 1) are the most suitable models to forecast the future exchange rate. Further the GARCH (1,1) model provided the volatility with non-constant conditional variance in the exchange rate with good forecasting performance. This study is very useful for researchers, policymakers, and businesses for making decisions through accurate and timely forecasting of the exchange rate and helps them in devising their policies.Keywords: exchange rate, ARIMA, GARCH, PAK/USD
Procedia PDF Downloads 56323392 Short Term Distribution Load Forecasting Using Wavelet Transform and Artificial Neural Networks
Authors: S. Neelima, P. S. Subramanyam
Abstract:
The major tool for distribution planning is load forecasting, which is the anticipation of the load in advance. Artificial neural networks have found wide applications in load forecasting to obtain an efficient strategy for planning and management. In this paper, the application of neural networks to study the design of short term load forecasting (STLF) Systems was explored. Our work presents a pragmatic methodology for short term load forecasting (STLF) using proposed two-stage model of wavelet transform (WT) and artificial neural network (ANN). It is a two-stage prediction system which involves wavelet decomposition of input data at the first stage and the decomposed data with another input is trained using a separate neural network to forecast the load. The forecasted load is obtained by reconstruction of the decomposed data. The hybrid model has been trained and validated using load data from Telangana State Electricity Board.Keywords: electrical distribution systems, wavelet transform (WT), short term load forecasting (STLF), artificial neural network (ANN)
Procedia PDF Downloads 43823391 The Best Prediction Data Mining Model for Breast Cancer Probability in Women Residents in Kabul
Authors: Mina Jafari, Kobra Hamraee, Saied Hossein Hosseini
Abstract:
The prediction of breast cancer disease is one of the challenges in medicine. In this paper we collected 528 records of women’s information who live in Kabul including demographic, life style, diet and pregnancy data. There are many classification algorithm in breast cancer prediction and tried to find the best model with most accurate result and lowest error rate. We evaluated some other common supervised algorithms in data mining to find the best model in prediction of breast cancer disease among afghan women living in Kabul regarding to momography result as target variable. For evaluating these algorithms we used Cross Validation which is an assured method for measuring the performance of models. After comparing error rate and accuracy of three models: Decision Tree, Naive Bays and Rule Induction, Decision Tree with accuracy of 94.06% and error rate of %15 is found the best model to predicting breast cancer disease based on the health care records.Keywords: decision tree, breast cancer, probability, data mining
Procedia PDF Downloads 14123390 Image Steganography Using Least Significant Bit Technique
Authors: Preeti Kumari, Ridhi Kapoor
Abstract:
In any communication, security is the most important issue in today’s world. In this paper, steganography is the process of hiding the important data into other data, such as text, audio, video, and image. The interest in this topic is to provide availability, confidentiality, integrity, and authenticity of data. The steganographic technique that embeds hides content with unremarkable cover media so as not to provoke eavesdropper’s suspicion or third party and hackers. In which many applications of compression, encryption, decryption, and embedding methods are used for digital image steganography. Due to compression, the nose produces in the image. To sustain noise in the image, the LSB insertion technique is used. The performance of the proposed embedding system with respect to providing security to secret message and robustness is discussed. We also demonstrate the maximum steganography capacity and visual distortion.Keywords: steganography, LSB, encoding, information hiding, color image
Procedia PDF Downloads 47623389 A Concept to Assess the Economic Importance of the On-Site Activities of ETICS
Authors: V. Sulakatko, F. U. Vogdt, I. Lill
Abstract:
Construction technology and on-site construction activities have a direct influence on the life cycle costs of energy efficiently renovated apartment buildings. The systematic inadequacies of the External Thermal Insulation Composite System (ETICS) which occur during the construction phase increase the risk for all stakeholders, reduce mechanical durability and increase the life cycle costs of the building. The economic effect of these shortcomings can be minimised if the risk of the most significant on-site activities is recognised. The objective of the presented ETICS economic assessment concept is to evaluate the economic influence of on-site shortcomings and reveal their significance to the foreseeable future repair costs. The model assembles repair techniques, discusses their direct cost calculation methods, argues over the proper usage of net present value over the life cycle of the building, and proposes a simulation tool to evaluate the risk of on-site activities. As the technique is dependent on the selected real interest rate, a sensitivity analysis is anticipated to determine the validity of the recommendations. After the verification of the model on the sample buildings by the industry, it is expected to increase economic rationality of resource allocation and reduce high-risk systematic shortcomings during the construction process of ETICS.Keywords: activity-based cost estimating, cost estimation, ETICS, life cycle costing
Procedia PDF Downloads 29923388 Investigating the Regulation System of the Synchronous Motor Excitation Mode Serving as a Reactive Power Source
Authors: Baghdasaryan Marinka, Ulikyan Azatuhi
Abstract:
The efficient usage of the compensation abilities of the electrical drive synchronous motors used in production processes can essentially improve the technical and economic indices of the process. Reducing the flows of the reactive electrical energy due to the compensation of reactive power allows to significantly reduce the load losses of power in the electrical networks. As a result of analyzing the scientific works devoted to the issues of regulating the excitation of the synchronous motors, the need for comprehensive investigation and estimation of the excitation mode has been substantiated. By means of the obtained transmission functions, in the Simulink environment of the software package MATLAB, the transition processes of the excitation mode have been studied. As a result of obtaining and estimating the graph of the Nyquist plot and the transient process, the necessity of developing the Proportional-Integral-Derivative (PID) regulator has been justified. The transient processes of the system of the PID regulator have been investigated, and the amplitude–phase characteristics of the system have been estimated. The analysis of the obtained results has shown that the regulation indices of the developed system have been improved. The developed system can be successfully applied for regulating the excitation voltage of different-power synchronous motors, operating with a changing load, ensuring a value of the power coefficient close to 1.Keywords: transition process, synchronous motor, excitation mode, regulator, reactive power
Procedia PDF Downloads 23623387 Optimal Hybrid Linear and Nonlinear Control for a Quadcopter Drone
Authors: Xinhuang Wu, Yousef Sardahi
Abstract:
A hybrid and optimal multi-loop control structure combining linear and nonlinear control algorithms are introduced in this paper to regulate the position of a quadcopter unmanned aerial vehicle (UAV) driven by four brushless DC motors. To this end, a nonlinear mathematical model of the UAV is derived and then linearized around one of its operating points. Using the nonlinear version of the model, a sliding mode control is used to derive the control laws of the motor thrust forces required to drive the UAV to a certain position. The linear model is used to design two controllers, XG-controller and YG-controller, responsible for calculating the required roll and pitch to maneuver the vehicle to the desired X and Y position. Three attitude controllers are designed to calculate the desired angular rates of rotors, assuming that the Euler angles are minimal. After that, a many-objective optimization problem involving 20 design parameters and ten objective functions is formulated and solved by HypE (Hypervolume estimation algorithm), one of the widely used many-objective optimization algorithms approaches. Both stability and performance constraints are imposed on the optimization problem. The optimization results in terms of Pareto sets and fronts are obtained and show that some of the design objectives are competing. That is, when one objective goes down, the other goes up. Also, Numerical simulations conducted on the nonlinear UAV model show that the proposed optimization method is quite effective.Keywords: optimal control, many-objective optimization, sliding mode control, linear control, cascade controllers, UAV, drones
Procedia PDF Downloads 7423386 Towards a Distributed Computation Platform Tailored for Educational Process Discovery and Analysis
Authors: Awatef Hicheur Cairns, Billel Gueni, Hind Hafdi, Christian Joubert, Nasser Khelifa
Abstract:
Given the ever changing needs of the job markets, education and training centers are increasingly held accountable for student success. Therefore, education and training centers have to focus on ways to streamline their offers and educational processes in order to achieve the highest level of quality in curriculum contents and managerial decisions. Educational process mining is an emerging field in the educational data mining (EDM) discipline, concerned with developing methods to discover, analyze and provide a visual representation of complete educational processes. In this paper, we present our distributed computation platform which allows different education centers and institutions to load their data and access to advanced data mining and process mining services. To achieve this, we present also a comparative study of the different clustering techniques developed in the context of process mining to partition efficiently educational traces. Our goal is to find the best strategy for distributing heavy analysis computations on many processing nodes of our platform.Keywords: educational process mining, distributed process mining, clustering, distributed platform, educational data mining, ProM
Procedia PDF Downloads 45523385 Enhancing Robustness in Federated Learning through Decentralized Oracle Consensus and Adaptive Evaluation
Authors: Peiming Li
Abstract:
This paper presents an innovative blockchain-based approach to enhance the reliability and efficiency of federated learning systems. By integrating a decentralized oracle consensus mechanism into the federated learning framework, we address key challenges of data and model integrity. Our approach utilizes a network of redundant oracles, functioning as independent validators within an epoch-based training system in the federated learning model. In federated learning, data is decentralized, residing on various participants' devices. This scenario often leads to concerns about data integrity and model quality. Our solution employs blockchain technology to establish a transparent and tamper-proof environment, ensuring secure data sharing and aggregation. The decentralized oracles, a concept borrowed from blockchain systems, act as unbiased validators. They assess the contributions of each participant using a Hidden Markov Model (HMM), which is crucial for evaluating the consistency of participant inputs and safeguarding against model poisoning and malicious activities. Our methodology's distinct feature is its epoch-based training. An epoch here refers to a specific training phase where data is updated and assessed for quality and relevance. The redundant oracles work in concert to validate data updates during these epochs, enhancing the system's resilience to security threats and data corruption. The effectiveness of this system was tested using the Mnist dataset, a standard in machine learning for benchmarking. Results demonstrate that our blockchain-oriented federated learning approach significantly boosts system resilience, addressing the common challenges of federated environments. This paper aims to make these advanced concepts accessible, even to those with a limited background in blockchain or federated learning. We provide a foundational understanding of how blockchain technology can revolutionize data integrity in decentralized systems and explain the role of oracles in maintaining model accuracy and reliability.Keywords: federated learning system, block chain, decentralized oracles, hidden markov model
Procedia PDF Downloads 66