Search results for: customer information process
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 23831

Search results for: customer information process

19151 Prediction of Compressive Strength of Concrete from Early Age Test Result Using Design of Experiments (Rsm)

Authors: Salem Alsanusi, Loubna Bentaher

Abstract:

Response Surface Methods (RSM) provide statistically validated predictive models that can then be manipulated for finding optimal process configurations. Variation transmitted to responses from poorly controlled process factors can be accounted for by the mathematical technique of propagation of error (POE), which facilitates ‘finding the flats’ on the surfaces generated by RSM. The dual response approach to RSM captures the standard deviation of the output as well as the average. It accounts for unknown sources of variation. Dual response plus propagation of error (POE) provides a more useful model of overall response variation. In our case, we implemented this technique in predicting compressive strength of concrete of 28 days in age. Since 28 days is quite time consuming, while it is important to ensure the quality control process. This paper investigates the potential of using design of experiments (DOE-RSM) to predict the compressive strength of concrete at 28th day. Data used for this study was carried out from experiment schemes at university of Benghazi, civil engineering department. A total of 114 sets of data were implemented. ACI mix design method was utilized for the mix design. No admixtures were used, only the main concrete mix constituents such as cement, coarse-aggregate, fine aggregate and water were utilized in all mixes. Different mix proportions of the ingredients and different water cement ratio were used. The proposed mathematical models are capable of predicting the required concrete compressive strength of concrete from early ages.

Keywords: mix proportioning, response surface methodology, compressive strength, optimal design

Procedia PDF Downloads 250
19150 Literature Review on the Barriers to Access Credit for Small Agricultural Producers and Policies to Mitigate Them in Developing Countries

Authors: Margarita Gáfaro, Karelys Guzmán, Paola Poveda

Abstract:

This paper establishes the theoretical aspects that explain the barriers to accessing credit for small agricultural producers in developing countries and identifies successful policy experiences to mitigate them. We will test two hypotheses. The first one is that information asymmetries, high transaction costs and high-risk exposure limit the supply of credit to small agricultural producers in developing countries. The second hypothesis is that low levels of financial education and productivity and high uncertainty about the returns of agricultural activity limit the demand for credit. To test these hypotheses, a review of the theoretical and empirical literature on access to rural credit in developing countries will be carried out. The first part of this review focuses on theoretical models that incorporate information asymmetries in the credit market and analyzes the interaction between these asymmetries and the characteristics of the agricultural sector in developing countries. Some of the characteristics we will focus on are the absence of collateral, the underdevelopment of the judicial systems and insurance markets, and the high dependence on climatic factors of production technologies. The second part of this review focuses on the determinants of credit demand by small agricultural producers, including the profitability of productive projects, security conditions, risk aversion or loss, financial education, and cognitive biases, among others. There are policies that focus on resolving these supply and demand constraints and managing to improve credit access. Therefore, another objective of this paper is to present a review of effective policies that have promoted access to credit for smallholders in the world. For this, information available in policy documents will be collected. This information will be complemented by interviews with officials in charge of the design and execution of these policies in a subset of selected countries. The information collected will be analyzed in light of the conceptual framework proposed in the first two parts of this section. The barriers to access to credit that each policy attempts to resolve and the factors that could explain its effectiveness will be identified.

Keywords: agricultural economics, credit access, smallholder, developing countries

Procedia PDF Downloads 52
19149 A Technique for Image Segmentation Using K-Means Clustering Classification

Authors: Sadia Basar, Naila Habib, Awais Adnan

Abstract:

The paper presents the Technique for Image Segmentation Using K-Means Clustering Classification. The presented algorithms were specific, however, missed the neighboring information and required high-speed computerized machines to run the segmentation algorithms. Clustering is the process of partitioning a group of data points into a small number of clusters. The proposed method is content-aware and feature extraction method which is able to run on low-end computerized machines, simple algorithm, required low-quality streaming, efficient and used for security purpose. It has the capability to highlight the boundary and the object. At first, the user enters the data in the representation of the input. Then in the next step, the digital image is converted into groups clusters. Clusters are divided into many regions. The same categories with same features of clusters are assembled within a group and different clusters are placed in other groups. Finally, the clusters are combined with respect to similar features and then represented in the form of segments. The clustered image depicts the clear representation of the digital image in order to highlight the regions and boundaries of the image. At last, the final image is presented in the form of segments. All colors of the image are separated in clusters.

Keywords: clustering, image segmentation, K-means function, local and global minimum, region

Procedia PDF Downloads 362
19148 Marine Natural Products: A Rich Source of Medicine in Ayurveda, the Ancient Indian Medical Science

Authors: Ashok D. Satpute

Abstract:

Ayurveda, the ancient Indian Medical system is practiced all over India and abroad, is rich in natural source of medicines, including marine products. The marine drugs which prominently used are pravala (coral), mukta (pearl), kapardika (cowry).Shukti (oyster shell), shankha (conch), agnijara (amber) etc. Except agnijara (amber) all are rich in calcium. Interestingly they are not used as supplements in calcium deficiency as done in conventional medical practice. They are used as medicines in the disease like fever, tuberculosis, bleeding disorders, eye problems, digestive complaints etc. Many scientific studies have shown their potent medicinal value. Each has its own properties and used therapeutically after subjecting them to various purificatory processes which are called shodhana in which several medicinal plants are used which also help in enhancing therapeutical activity. Then these purified marine products are subjected to marana (incineration) process and obtained in the form of Bhasma (a finest form of medicine). Agnijara, a derivative of whale is useful as aphrodisiac and prescribed in neuromuscular disorders and tetanus. The ancient scriptures written in Sanskrit language thousands of years back have rich information about all these natural marine products and their medicinal usage.

Keywords: Ayurveda, bhasma, marana, shodhana

Procedia PDF Downloads 262
19147 Scaling-Down an Agricultural Waste Biogas Plant Fermenter

Authors: Matheus Pessoa, Matthias Kraume

Abstract:

Scale-Down rules in process engineering help us to improve and develop Industrial scale parameters into lab scale. Several scale-down rules available in the literature like Impeller Power Number, Agitation device Power Input, Substrate Tip Speed, Reynolds Number and Cavern Development were investigated in order to stipulate the rotational speed to operate an 11 L working volume lab-scale bioreactor within industrial process parameters. Herein, xanthan gum was used as a fluid with a representative viscosity of a hypothetical biogas plant, with H/D = 1 and central agitation, fermentation broth using sewage sludge and sugar beet pulp as substrate. The results showed that the cavern development strategy was the best method for establishing a rotational speed for the bioreactor operation, while the other rules presented values out of reality for this article proposes.

Keywords: anaerobic digestion, cavern development, scale down rules, xanthan gum

Procedia PDF Downloads 472
19146 A Multi-criteria Decision Method For The Recruitment Of Academic Personnel Based On The Analytical Hierarchy Process And The Delphi Method In A Neutrosophic Environment (Full Text)

Authors: Antonios Paraskevas, Michael Madas

Abstract:

For a university to maintain its international competitiveness in education, it is essential to recruit qualitative academic staff as it constitutes its most valuable asset. This selection demonstrates a significant role in achieving strategic objectives, particularly by emphasizing a firm commitment to exceptional student experience and innovative teaching and learning practices of high quality. In this vein, the appropriate selection of academic staff establishes a very important factor of competitiveness, efficiency and reputation of an academic institute. Within this framework, our work demonstrates a comprehensive methodological concept that emphasizes on the multi-criteria nature of the problem and on how decision makers could utilize our approach in order to proceed to the appropriate judgment. The conceptual framework introduced in this paper is built upon a hybrid neutrosophic method based on the Neutrosophic Analytical Hierarchy Process (N-AHP), which uses the theory of neutrosophy sets and is considered suitable in terms of significant degree of ambiguity and indeterminacy observed in decision-making process. To this end, our framework extends the N-AHP by incorporating the Neutrosophic Delphi Method (N-DM). By applying the N-DM, we can take into consideration the importance of each decision-maker and their preferences per evaluation criterion. To the best of our knowledge, the proposed model is the first which applies Neutrosophic Delphi Method in the selection of academic staff. As a case study, it was decided to use our method to a real problem of academic personnel selection, having as main goal to enhance the algorithm proposed in previous scholars’ work, and thus taking care of the inherit ineffectiveness which becomes apparent in traditional multi-criteria decision-making methods when dealing with situations alike. As a further result, we prove that our method demonstrates greater applicability and reliability when compared to other decision models.

Keywords: analytical hierarchy process, delphi method, multi-criteria decision maiking method, neutrosophic set theory, personnel recruitment

Procedia PDF Downloads 172
19145 Deep Learning Based on Image Decomposition for Restoration of Intrinsic Representation

Authors: Hyohun Kim, Dongwha Shin, Yeonseok Kim, Ji-Su Ahn, Kensuke Nakamura, Dongeun Choi, Byung-Woo Hong

Abstract:

Artefacts are commonly encountered in the imaging process of clinical computed tomography (CT) where the artefact refers to any systematic discrepancy between the reconstructed observation and the true attenuation coefficient of the object. It is known that CT images are inherently more prone to artefacts due to its image formation process where a large number of independent detectors are involved, and they are assumed to yield consistent measurements. There are a number of different artefact types including noise, beam hardening, scatter, pseudo-enhancement, motion, helical, ring, and metal artefacts, which cause serious difficulties in reading images. Thus, it is desired to remove nuisance factors from the degraded image leaving the fundamental intrinsic information that can provide better interpretation of the anatomical and pathological characteristics. However, it is considered as a difficult task due to the high dimensionality and variability of data to be recovered, which naturally motivates the use of machine learning techniques. We propose an image restoration algorithm based on the deep neural network framework where the denoising auto-encoders are stacked building multiple layers. The denoising auto-encoder is a variant of a classical auto-encoder that takes an input data and maps it to a hidden representation through a deterministic mapping using a non-linear activation function. The latent representation is then mapped back into a reconstruction the size of which is the same as the size of the input data. The reconstruction error can be measured by the traditional squared error assuming the residual follows a normal distribution. In addition to the designed loss function, an effective regularization scheme using residual-driven dropout determined based on the gradient at each layer. The optimal weights are computed by the classical stochastic gradient descent algorithm combined with the back-propagation algorithm. In our algorithm, we initially decompose an input image into its intrinsic representation and the nuisance factors including artefacts based on the classical Total Variation problem that can be efficiently optimized by the convex optimization algorithm such as primal-dual method. The intrinsic forms of the input images are provided to the deep denosing auto-encoders with their original forms in the training phase. In the testing phase, a given image is first decomposed into the intrinsic form and then provided to the trained network to obtain its reconstruction. We apply our algorithm to the restoration of the corrupted CT images by the artefacts. It is shown that our algorithm improves the readability and enhances the anatomical and pathological properties of the object. The quantitative evaluation is performed in terms of the PSNR, and the qualitative evaluation provides significant improvement in reading images despite degrading artefacts. The experimental results indicate the potential of our algorithm as a prior solution to the image interpretation tasks in a variety of medical imaging applications. This work was supported by the MISP(Ministry of Science and ICT), Korea, under the National Program for Excellence in SW (20170001000011001) supervised by the IITP(Institute for Information and Communications Technology Promotion).

Keywords: auto-encoder neural network, CT image artefact, deep learning, intrinsic image representation, noise reduction, total variation

Procedia PDF Downloads 178
19144 Ideological Manipulations and Cultural-Norm Constraints

Authors: Masoud Hassanzade Novin, Bahloul Salmani

Abstract:

Translation cannot be considered as a simple linguistic act. Through the rise of descriptive approach in the late 1970s and 1980s, translation process managed to meet the requirements of social aspects as well as linguistic approaches. To have the translation considered as the cross-cultural communication through which various cultures communicate in ideological and cultural constraints, the contrastive analysis was conducted in this paper to reveal the distortions imposed in the translated texts. The corpus of the study involved the novel 1984 written by George Orwell and its Persian translated texts which were analyzed through the qualitative type of the research based on critical discourse analysis (CDA) and Toury's norms as well as Lefever's concepts of ideology. Results of the study revealed the point that ideology and the cultural constraints were considered as an important stimulus which can control the process of the translation.

Keywords: critical discourse analysis, ideology, norms, translated texts

Procedia PDF Downloads 323
19143 High Pressure Delignification Process for Nanocrystalline Cellulose Production from Agro-Waste Biomass

Authors: Sakinul Islam, Nhol Kao, Sati Bhattacharya, Rahul Gupta

Abstract:

Nanocrystalline cellulose (NCC) has been widely used for miscellaneous applications due to its superior properties over other nanomaterials. However, the major problems associated with the production of NCC are long reaction time, low production rate and inefficient process. The mass production of NCC within a short period of time is still a great challenge. The main objective of this study is to produce NCC from rice husk agro waste biomass from a high pressure delignification process (HPDP), followed by bleaching and hydrolysis processes. The HPDP has not been explored for NCC production from rice husk biomass (RHB) until now. In order to produce NCC, powder rice husk (PRH) was placed into a stainless steel reactor at 80 ˚C under 5 bars. Aqueous solution of NaOH (4M) was used for the dissolution of lignin and other amorphous impurities from PRH. After certain experimental times (1h, 3.5h and 6h), bleaching and hydrolysis were carried out on delignified samples. NaOCl (20%) and H2SO4 (4M) solutions were used for bleaching and hydrolysis processes, respectively. The NCC suspension from hydrolysis was sonicated and neutralized by buffer solution for various characterisations. Finally NCC suspension was dried and analyzed by FTIR, XRD, SEM, AFM and TEM. The chemical composition of NCC and PRH was estimated by TAPPI (Technical Association of Pulp and Paper Industry) standard methods to observe the product purity. It was found that, the 6h of the HPDP was more efficient to produce good quality NCC than that at 1h and 3.5h due to low separation of non-cellulosic components from RHB. The analyses indicated the crystallinity of NCC to be 71 %, particle size of 20-50 nm (diameter) and 100-200 nm in length.

Keywords: nanocrystalline cellulose, NCC, high pressure delignification, bleaching, hydrolysis, agro-waste biomass

Procedia PDF Downloads 249
19142 Improvement of Ground Truth Data for Eye Location on Infrared Driver Recordings

Authors: Sorin Valcan, Mihail Gaianu

Abstract:

Labeling is a very costly and time consuming process which aims to generate datasets for training neural networks in several functionalities and projects. For driver monitoring system projects, the need for labeled images has a significant impact on the budget and distribution of effort. This paper presents the modifications done to an algorithm used for the generation of ground truth data for 2D eyes location on infrared images with drivers in order to improve the quality of the data and performance of the trained neural networks. The algorithm restrictions become tougher, which makes it more accurate but also less constant. The resulting dataset becomes smaller and shall not be altered by any kind of manual label adjustment before being used in the neural networks training process. These changes resulted in a much better performance of the trained neural networks.

Keywords: labeling automation, infrared camera, driver monitoring, eye detection, convolutional neural networks

Procedia PDF Downloads 93
19141 Leaching of Copper from Copper Ore Using Sulphuric Acid in the Presence of Hydrogen Peroxide as an Oxidizing Agent: An Optimized Process

Authors: Hilary Rutto

Abstract:

Leaching with acids are the most commonly reagents used to remove copper ions from its copper ores. It is important that the process conditions are optimized to improve the leaching efficiency. In the present study the effects of pH, oxidizing agent (hydrogen peroxide), stirring speed, solid to liquid ratio and acid concentration on the leaching of copper ions from it ore were investigated using a pH Stat apparatus. Copper ions were analyzed at the end of each experiment using Atomic Absorption (AAS) machine. Results showed that leaching efficiency improved with an increase in acid concentration, stirring speed, oxidizing agent, pH and decreased with an increase in the solid to liquid ratio.

Keywords: leaching, copper, oxidizing agent, pH stat apparatus

Procedia PDF Downloads 359
19140 The Role of Marketing Information System on Decision-Making: An Applied Study on Algeria Telecoms Mobile "MOBILIS"

Authors: Benlakhdar Mohamed Larbi, Yagoub Asma

Abstract:

Purpose: This study aims at highlighting the significance and importance of utilizing marketing information system (MKIS) on decision-making, by clarifying the need for quick and efficient decision-making due to time saving and preventing of duplication of work. Design, methodology, approach: The study shows the roles of each part of MKIS for developing marketing strategy, which present a real challenge to individuals and institutions in an era characterized by uncertainty and clarifying the importance of each part separately, depending on decision type and the nature of the situation. The empirical research method was evaluated by specialized experts, conducted by means of questionnaires. Correlation analysis was employed to test the validity of the procedure. Results: The empirical study findings confirmed positive relationships between the level of utilizing and adopting ‘decision support system and marketing intelligence’ and the success of an organizational decision-making, and provide the organization with a competitive advantage as it allows the organization to solve problems. Originality/value: The study offer better understanding of performance- increasing market share as an organizational decision making based on marketing information system.

Keywords: database, marketing research, marketing intelligence, decision support system, decision-making

Procedia PDF Downloads 307
19139 Design of the Fiber Lay-Up for the Composite Wind Turbine Blade in VARTM

Authors: Tzai-Shiung Li, Wen-Bin Young

Abstract:

The wind turbine blade sustains various kinds of loadings during the operating and parking state. Due to the increasing size of the wind turbine blade, it is important to arrange the composite materials in a sufficient way to reach the optimal utilization of the material strength. In the fabrication process of the vacuum assisted resin transfer molding, the fiber content of the turbine blade depends on the vacuum pressure. In this study, a design of the fiber layup for the vacuum assisted resin transfer molding is conducted to achieve the efficient utilization the material strength. This design is for the wind turbine blade consisting of shell skins with or without the spar structure.

Keywords: resin film infiltration, vacuum assisted resin transfer molding process, wind turbine blade, composite materials

Procedia PDF Downloads 368
19138 Building Information Modelling-Based Diminished Reality Visualisation to Facilitate Building Renovation Projects

Authors: Roghieh Eskandari, Ali Motamedi

Abstract:

There is a significant demand for renovation as-built assets are aging. To plan for a desirable and comfortable indoor environment, stakeholders use simulation technics to assess potential renovation scenarios with the innovative designs. Diminished Reality (DR), which is a technique of visually removing unwanted objects from the real-world scene in real-time, can contribute to the renovation design visualization for stakeholders by removing existing structures and assets from the scene. Using DR, the objects to be demolished or changed will be visually removed from the scene for a better understanding of the intended design scenarios for stakeholders. This research proposes an integrated system for renovation plan visualization using Building Information Modelling (BIM) data and mixed reality (MR) technologies. It presents a BIM-based DR method that utilizes a textured BIM model of the environment to accurately register the virtual model of the occluded background to the physical world in real-time. This system can facilitate the simulation of the renovation plan by visually diminishing building elements in an indoor environment.

Keywords: diminished reality, building information modelling, mixed reality, stock renovation

Procedia PDF Downloads 101
19137 Attribute Analysis of Quick Response Code Payment Users Using Discriminant Non-negative Matrix Factorization

Authors: Hironori Karachi, Haruka Yamashita

Abstract:

Recently, the system of quick response (QR) code is getting popular. Many companies introduce new QR code payment services and the services are competing with each other to increase the number of users. For increasing the number of users, we should grasp the difference of feature of the demographic information, usage information, and value of users between services. In this study, we conduct an analysis of real-world data provided by Nomura Research Institute including the demographic data of users and information of users’ usages of two services; LINE Pay, and PayPay. For analyzing such data and interpret the feature of them, Nonnegative Matrix Factorization (NMF) is widely used; however, in case of the target data, there is a problem of the missing data. EM-algorithm NMF (EMNMF) to complete unknown values for understanding the feature of the given data presented by matrix shape. Moreover, for comparing the result of the NMF analysis of two matrices, there is Discriminant NMF (DNMF) shows the difference of users features between two matrices. In this study, we combine EMNMF and DNMF and also analyze the target data. As the interpretation, we show the difference of the features of users between LINE Pay and Paypay.

Keywords: data science, non-negative matrix factorization, missing data, quality of services

Procedia PDF Downloads 118
19136 Statistically Significant Differences of Carbon Dioxide and Carbon Monoxide Emission in Photocopying Process

Authors: Kiurski S. Jelena, Kecić S. Vesna, Oros B. Ivana

Abstract:

Experimental results confirmed the temporal variation of carbon dioxide and carbon monoxide concentration during the working shift of the photocopying process in a small photocopying shop in Novi Sad, Serbia. The statistically significant differences of target gases were examined with two-way analysis of variance without replication followed by Scheffe's post hoc test. The existence of statistically significant differences was obtained for carbon monoxide emission which is pointed out with F-values (12.37 and 31.88) greater than Fcrit (6.94) in contrary to carbon dioxide emission (F-values of 1.23 and 3.12 were less than Fcrit).  Scheffe's post hoc test indicated that sampling point A (near the photocopier machine) and second time interval contribute the most on carbon monoxide emission.

Keywords: analysis of variance, carbon dioxide, carbon monoxide, photocopying indoor, Scheffe's test

Procedia PDF Downloads 310
19135 Applying Kinect on the Development of a Customized 3D Mannequin

Authors: Shih-Wen Hsiao, Rong-Qi Chen

Abstract:

In the field of fashion design, 3D Mannequin is a kind of assisting tool which could rapidly realize the design concepts. While the concept of 3D Mannequin is applied to the computer added fashion design, it will connect with the development and the application of design platform and system. Thus, the situation mentioned above revealed a truth that it is very critical to develop a module of 3D Mannequin which would correspond with the necessity of fashion design. This research proposes a concrete plan that developing and constructing a system of 3D Mannequin with Kinect. In the content, ergonomic measurements of objective human features could be attained real-time through the implement with depth camera of Kinect, and then the mesh morphing can be implemented through transformed the locations of the control-points on the model by inputting those ergonomic data to get an exclusive 3D mannequin model. In the proposed methodology, after the scanned points from the Kinect are revised for accuracy and smoothening, a complete human feature would be reconstructed by the ICP algorithm with the method of image processing. Also, the objective human feature could be recognized to analyze and get real measurements. Furthermore, the data of ergonomic measurements could be applied to shape morphing for the division of 3D Mannequin reconstructed by feature curves. Due to a standardized and customer-oriented 3D Mannequin would be generated by the implement of subdivision, the research could be applied to the fashion design or the presentation and display of 3D virtual clothes. In order to examine the practicality of research structure, a system of 3D Mannequin would be constructed with JAVA program in this study. Through the revision of experiments the practicability-contained research result would come out.

Keywords: 3D mannequin, kinect scanner, interactive closest point, shape morphing, subdivision

Procedia PDF Downloads 294
19134 Design of a Tool for Generating Test Cases from BPMN

Authors: Prat Yotyawilai, Taratip Suwannasart

Abstract:

Business Process Model and Notation (BPMN) is more important in the business process and creating functional models, and is a standard for OMG, which becomes popular in various organizations and in education. Researches related to software testing based on models are prominent. Although most researches use the UML model in software testing, not many researches use the BPMN Model in creating test cases. Therefore, this research proposes a design of a tool for generating test cases from the BPMN. The model is analyzed and the details of the various components are extracted before creating a flow graph. Both details of components and the flow graph are used in generating test cases.

Keywords: software testing, test case, BPMN, flow graph

Procedia PDF Downloads 540
19133 The Effects of Cost-Sharing Contracts on the Costs and Operations of E-Commerce Supply Chains

Authors: Sahani Rathnasiri, Pritee Ray, Sardar M. N. Isalm, Carlos A. Vega-Mejia

Abstract:

This study develops a cooperative game theory-based cost-sharing contract model for a business to consumer (B2C) e-commerce supply chain to minimize the overall supply chain costs and the individual costs within an information asymmetry scenario. The objective of this study is to address the issues of strategic interactions among the key players of the e-commerce supply chain operation, which impedes the optimal operational outcomes. Game theory has been included in the field of supply chain management to resolve strategic decision-making issues; however, most of the studies are limited only to two-echelons of the supply chains. Multi-echelon supply chain optimizations based on game-theoretic models are less explored in the previous literature. This study adopts a cooperative game model to focus on the common payoff of operations and addresses the issues of information asymmetry and coordination of a three-echelon e-commerce supply chain. The cost-sharing contract model integrates operational features such as production, inventory management and distribution with the contract related constraints. The outcomes of the model highlight the importance of maintaining lower operational costs by all players to obtain benefits from the cost-sharing contract. Further, the cost-sharing contract ensures true cost revelation, and hence eliminates the information asymmetry issues among the players. Comparing the results of the contract model with the de-centralized e-commerce supply chain operation further emphasizes that the cost-sharing contract derives Pareto-improved outcomes and minimizes the costs of overall e-commerce supply chain operation.

Keywords: cooperative game theory, cost-sharing contract, e-commerce supply chain, information asymmetry

Procedia PDF Downloads 110
19132 Efficient Position Based Operation Code Authentication

Authors: Hashim Ali, Sheheryar Khan

Abstract:

Security for applications is always been a keen issue of concern. In general, security is to allow access of grant to legal user or to deny non-authorized access to the system. Shoulder surfing is an observation technique to hack an account or to enter into a system. When a malicious observer is capturing or recording the fingers of a user while he is entering sensitive inputs (PIN, Passwords etc.) and may be able to observe user’s password credential. It is very rigorous for a novice user to prevent himself from shoulder surfing or unaided observer in a public place while accessing his account. In order to secure the user account, there are five factors of authentication; they are: “(i) something you have, (ii) something you are, (iii) something you know, (iv) somebody you know, (v) something you process”. A technique has been developed of fifth-factor authentication “something you process” to provide novel approach to the user. In this paper, we have applied position based operational code authentication in such a way to more easy and user friendly to the user.

Keywords: shoulder surfing, malicious observer, sensitive inputs, authentication

Procedia PDF Downloads 254
19131 The Analysis of Gizmos Online Program as Mathematics Diagnostic Program: A Story from an Indonesian Private School

Authors: Shofiayuningtyas Luftiani

Abstract:

Some private schools in Indonesia started integrating the online program Gizmos in the teaching-learning process. Gizmos was developed to supplement the existing curriculum by integrating it into the instructional programs. The program has some features using an inquiry-based simulation, in which students conduct exploration by using a worksheet while teachers use the teacher guidelines to direct and assess students’ performance In this study, the discussion about Gizmos highlights its features as the assessment media of mathematics learning for secondary school students. The discussion is based on the case study and literature review from the Indonesian context. The purpose of applying Gizmos as an assessment media refers to the diagnostic assessment. As a part of the diagnostic assessment, the teachers review the student exploration sheet, analyze particularly in the students’ difficulties and consider findings in planning future learning process. This assessment becomes important since the teacher needs the data about students’ persistent weaknesses. Additionally, this program also helps to build student’ understanding by its interactive simulation. Currently, the assessment over-emphasizes the students’ answers in the worksheet based on the provided answer keys while students perform their skill in translating the question, doing the simulation and answering the question. Whereas, the assessment should involve the multiple perspectives and sources of students’ performance since teacher should adjust the instructional programs with the complexity of students’ learning needs and styles. Consequently, the approach to improving the assessment components is selected to challenge the current assessment. The purpose of this challenge is to involve not only the cognitive diagnosis but also the analysis of skills and error. Concerning the selected setting for this diagnostic assessment that develops the combination of cognitive diagnosis, skills analysis and error analysis, the teachers should create an assessment rubric. The rubric plays the important role as the guide to provide a set of criteria for the assessment. Without the precise rubric, the teacher potentially ineffectively documents and follows up the data about students at risk of failure. Furthermore, the teachers who employ the program of Gizmos as the diagnostic assessment might encounter some obstacles. Based on the condition of assessment in the selected setting, the obstacles involve the time constrain, the reluctance of higher teaching burden and the students’ behavior. Consequently, the teacher who chooses the Gizmos with those approaches has to plan, implement and evaluate the assessment. The main point of this assessment is not in the result of students’ worksheet. However, the diagnostic assessment has the two-stage process; the process to prompt and effectively follow-up both individual weaknesses and those of the learning process. Ultimately, the discussion of Gizmos as the media of the diagnostic assessment refers to the effort to improve the mathematical learning process.

Keywords: diagnostic assessment, error analysis, Gizmos online program, skills analysis

Procedia PDF Downloads 166
19130 A Review of Current Knowledge on Assessment of Precast Structures Using Fragility Curves

Authors: E. Akpinar, A. Erol, M.F. Cakir

Abstract:

Precast reinforced concrete (RC) structures are excellent alternatives for construction world all over the globe, thanks to their rapid erection phase, ease mounting process, better quality and reasonable prices. Such structures are rather popular for industrial buildings. For the sake of economic importance of such industrial buildings as well as significance of safety, like every other type of structures, performance assessment and structural risk analysis are important. Fragility curves are powerful tools for damage projection and assessment for any sort of building as well as precast structures. In this study, a comparative review of current knowledge on fragility analysis of industrial precast RC structures were presented and findings in previous studies were compiled. Effects of different structural variables, parameters and building geometries as well as soil conditions on fragility analysis of precast structures are reviewed. It was aimed to briefly present the information in the literature about the procedure of damage probability prediction including fragility curves for such industrial facilities. It is found that determination of the aforementioned structural parameters as well as selecting analysis procedure are critically important for damage prediction of industrial precast RC structures using fragility curves.

Keywords: damage prediction, fragility curve, industrial buildings, precast reinforced concrete structures

Procedia PDF Downloads 176
19129 Optimizing Approach for Sifting Process to Solve a Common Type of Empirical Mode Decomposition Mode Mixing

Authors: Saad Al-Baddai, Karema Al-Subari, Elmar Lang, Bernd Ludwig

Abstract:

Empirical mode decomposition (EMD), a new data-driven of time-series decomposition, has the advantage of supposing that a time series is non-linear or non-stationary, as is implicitly achieved in Fourier decomposition. However, the EMD suffers of mode mixing problem in some cases. The aim of this paper is to present a solution for a common type of signals causing of EMD mode mixing problem, in case a signal suffers of an intermittency. By an artificial example, the solution shows superior performance in terms of cope EMD mode mixing problem comparing with the conventional EMD and Ensemble Empirical Mode decomposition (EEMD). Furthermore, the over-sifting problem is also completely avoided; and computation load is reduced roughly six times compared with EEMD, an ensemble number of 50.

Keywords: empirical mode decomposition (EMD), mode mixing, sifting process, over-sifting

Procedia PDF Downloads 376
19128 Transesterification of Jojoba Oil Wax Using Microwave Technique

Authors: Moataz Elsawy, Hala F. Naguib, Hilda A. Aziz, Eid A. Ismail, Labiba I. Hussein, Maher Z. Elsabee

Abstract:

Jojoba oil-wax is extracted from the seeds of the jojoba (Simmondsia chinensis Link Schneider), a perennial shrub that grows in semi-desert areas in Egypt and in some parts of the world. The main uses of jojoba oil wax are in the cosmetics and pharmaceutical industry, but new uses could arise related to the search of new energetic crops. This paper summarizes a process to convert the jojoba oil wax to biodiesel by transesterification with ethanol and a series of aliphatic alcohols using a more economic and energy saving method in a domestic microwave. The effect of time and power of the microwave on the extent of the transesterification using ethanol and other aliphatic alcohols has been studied. The separation of the alkyl esters from the fatty alcohols rich fraction has been done in a single crystallization step at low temperature (−18°C) from low boiling point petroleum ether. Gas chromatography has been used to follow up the transesterification process. All products have been characterized by spectral analysis.

Keywords: jojoba oil, transesterification, microwave, gas chromatography jojoba esters, jojoba alcohol

Procedia PDF Downloads 448
19127 Study of Biofuel Produced by Babassu Oil Fatty Acids Esterification

Authors: F. A. F. da Ponte, J. Q. Malveira, I. A. Maciel, M. C. G. Albuquerque

Abstract:

In this work aviation, biofuel production was studied by fatty acids (C6 to C16) esterification. The process variables in heterogeneous catalysis were evaluated using an experimental design. Temperature and reaction time were the studied parameters, and the methyl esters content was the response of the experimental design. An ion exchange resin was used as a heterogeneous catalyst. The process optimization was carried out using response surface methodology (RSM) and polynomial model of second order. Results show that the most influential variables on the linear coefficient of each effect studied were temperature and reaction time. The best result of methyl esters conversion in the experimental design was under the conditions: 10% wt of catalyst; 100 °C and 4 hours of reaction. The best-achieved conversion was 96.5% wt of biofuel.

Keywords: esterification, ion-exchange resins, response surface methodology, biofuel

Procedia PDF Downloads 479
19126 Development of a Triangular Evaluation Protocol in a Multidisciplinary Design Process of an Ergometric Step

Authors: M. B. Ricardo De Oliveira, A. Borghi-Silva, E. Paravizo, F. Lizarelli, L. Di Thomazzo, D. Braatz

Abstract:

Prototypes are a critical feature in the product development process, as they help the project team visualize early concept flaws, communicate ideas and introduce an initial product testing. Involving stakeholders, such as consumers and users, in prototype tests allows the gathering of valuable feedback, contributing for a better product and making the design process more participatory. Even though recent studies have shown that user evaluation of prototypes is valuable, few articles provide a method or protocol on how designers should conduct it. This multidisciplinary study (involving the areas of physiotherapy, engineering and computer science) aims to develop an evaluation protocol, using an ergometric step prototype as the product prototype to be assessed. The protocol consisted of performing two tests (the 2 Minute Step Test and the Portability Test) to allow users (patients) and consumers (physiotherapists) to have an experience with the prototype. Furthermore, the protocol contained four Likert-Scale questionnaires (one for users and three for consumers), that inquired participants about how they perceived the design characteristics of the product (performance, safety, materials, maintenance, portability, usability and ergonomics), in their use of the prototype. Additionally, the protocol indicated the need to conduct interviews with the product designers, in order to link their feedback to the ones from the consumers and users. Both tests and interviews were recorded for further analysis. The participation criteria for the study was gender and age for patients, gender and experience with 2 Minute Step Test for physiotherapists and involvement level in the product development project for designers. The questionnaire's reliability was validated using Cronbach's Alpha and the quantitative data of the questionnaires were analyzed using non-parametric hypothesis tests with a significance level of 0.05 (p <0.05) and descriptive statistics. As a result, this study provides a concise evaluation protocol which can assist designers in their development process, collecting quantitative feedback from consumer and users, and qualitative feedback from designers.

Keywords: Product Design, Product Evaluation, Prototypes, Step

Procedia PDF Downloads 103
19125 A Social Identity Analysis of Ottoman and Safavid Architects in the Historical Documents of the 16th to 17th Centuries

Authors: Farzaneh Farrokhfar, Mohammad Khazaie

Abstract:

The 16th and 17th centuries coincide with the classical age of Ottoman art history. Simultaneously with this age and in the eastern neighborhood of the Ottoman state, the Safavid Shiite state emerged, which, despite political and religious differences with the Ottomans, played an important role in cultural and artistic exchanges with Anatolia. The harmony of arts, including architecture, is one of the most important manifestations of cultural exchange between the two regions, which shows the intellectual commonalities of the two regions. In parallel with the production of works of art, the registration of information and identities of Ottoman and Safavid artists and craftsmen has been done by many historians and biographers, some of whom, fortunately, are available to us today and can be evaluated. This research first intends to read historical documents and reports related to the architects of the two Ottoman states in Anatolia and Safavid states in Iran in the 16th and 17th centuries and then examines the status of architects' information records and their location in the two regions. The results reveal the names and identities of some Ottoman and Safavid architects in the 16th and 17th centuries and show the method of recording information in the documents of the two regions. This research is done in a comparative historical method, and the method of collecting its resources is a documentary library.

Keywords: classical era, Ottoman architecture, Safavid architecture, Central Asian historical documents

Procedia PDF Downloads 116
19124 Through Seligman’s Lenses: Creating a Culture of Well-Being in Higher-Education

Authors: Neeru Deep, Kimberly McAlister

Abstract:

Mental health issues have been increasing worldwide for many decades, but the COVID-19 pandemic has brought mental health issues into the spotlight. Within higher education, promoting the well-being of students has dramatically increased in focus. The Northwestern State University of Louisiana opened the Center for Positivity, Well-being, and Hope using the action research process of reflecting, planning, acting, and observing. The study’s purpose is two-fold: First, it highlights how to create a collaborative team to reflect, plan, and act to develop a well-being culture in higher education institutions. Second, it investigates the efficacy of the center through Seligman’s lenses. The researchers shared their experience in the first three phases of the action research process and then applied an identical concurrent mixed methods design. A purposive sample evaluated the efficacy of the center through Seligman’s lenses. The researcher administered PERMA-Profiler Measure, the PERMA-Profiler Measure overview, the CoPWH Evaluation I, and the CoPWH Evaluation II questionnaires to collect qualitative and quantitative data. The thematic analysis for qualitative and descriptive statistics for quantitative data concluded that the center creates a well-being culture and promotes well-being in college students. In conclusion, this action research shares the successful implementation of the cyclic process of research in promoting a well-being culture in higher education with the implications for promoting a well-being culture in various educational settings, workplaces, and communities.

Keywords: action research, mixed methods research design, Seligman, well-being.

Procedia PDF Downloads 115
19123 Serious Gaming for Behaviour Change: A Review

Authors: Ramy Hammady, Sylvester Arnab

Abstract:

Significant attention has been directed to adopt game interventions practically to change certain behaviours in many disciplines such as health, education, psychology through many years. That’s due to the intrinsic motivation that games can cause and the substantial impact the games can leave on the player. Many review papers were induced to highlight and measure the effectiveness of the game’s interventions on changing behaviours; however, most of these studies neglected the game design process itself and the game features and elements that can stimuli changing behaviours. Therefore, this paper aims to identify the most game design mechanics and features that are the most influencing on changing behaviour during or after games interventions. This paper also sheds light on the theories of changing behaviours that clearly can led the game design process. This study gives directions to game designers to spot the most influential game features and mechanics for changing behaviour games in order to exploit it on the same manner.

Keywords: behaviour change, game design, serious gaming, gamification, review

Procedia PDF Downloads 195
19122 Correlation between Funding and Publications: A Pre-Step towards Future Research Prediction

Authors: Ning Kang, Marius Doornenbal

Abstract:

Funding is a very important – if not crucial – resource for research projects. Usually, funding organizations will publish a description of the funded research to describe the scope of the funding award. Logically, we would expect research outcomes to align with this funding award. For that reason, we might be able to predict future research topics based on present funding award data. That said, it remains to be shown if and how future research topics can be predicted by using the funding information. In this paper, we extract funding project information and their generated paper abstracts from the Gateway to Research database as a group, and use the papers from the same domains and publication years in the Scopus database as a baseline comparison group. We annotate both the project awards and the papers resulting from the funded projects with linguistic features (noun phrases), and then calculate tf-idf and cosine similarity between these two set of features. We show that the cosine similarity between the project-generated papers group is bigger than the project-baseline group, and also that these two groups of similarities are significantly different. Based on this result, we conclude that the funding information actually correlates with the content of future research output for the funded project on the topical level. How funding really changes the course of science or of scientific careers remains an elusive question.

Keywords: natural language processing, noun phrase, tf-idf, cosine similarity

Procedia PDF Downloads 232