Search results for: data mining technique
26636 Data Integrity: Challenges in Health Information Systems in South Africa
Authors: T. Thulare, M. Herselman, A. Botha
Abstract:
Poor system use, including inappropriate design of health information systems, causes difficulties in communication with patients and increased time spent by healthcare professionals in recording the necessary health information for medical records. System features like pop-up reminders, complex menus, and poor user interfaces can make medical records far more time consuming than paper cards as well as affect decision-making processes. Although errors associated with health information and their real and likely effect on the quality of care and patient safety have been documented for many years, more research is needed to measure the occurrence of these errors and determine the causes to implement solutions. Therefore, the purpose of this paper is to identify data integrity challenges in hospital information systems through a scoping review and based on the results provide recommendations on how to manage these. Only 34 papers were found to be most suitable out of 297 publications initially identified in the field. The results indicated that human and computerized systems are the most common challenges associated with data integrity and factors such as policy, environment, health workforce, and lack of awareness attribute to these challenges but if measures are taken the data integrity challenges can be managed.Keywords: data integrity, data integrity challenges, hospital information systems, South Africa
Procedia PDF Downloads 18226635 Mixed Effects Models for Short-Term Load Forecasting for the Spanish Regions: Castilla-Leon, Castilla-La Mancha and Andalucia
Authors: C. Senabre, S. Valero, M. Lopez, E. Velasco, M. Sanchez
Abstract:
This paper focuses on an application of linear mixed models to short-term load forecasting. The challenge of this research is to improve a currently working model at the Spanish Transport System Operator, programmed by us, and based on linear autoregressive techniques and neural networks. The forecasting system currently forecasts each of the regions within the Spanish grid separately, even though the behavior of the load in each region is affected by the same factors in a similar way. A load forecasting system has been verified in this work by using the real data from a utility. In this research it has been used an integration of several regions into a linear mixed model as starting point to obtain the information from other regions. Firstly, the systems to learn general behaviors present in all regions, and secondly, it is identified individual deviation in each regions. The technique can be especially useful when modeling the effect of special days with scarce information from the past. The three most relevant regions of the system have been used to test the model, focusing on special day and improving the performance of both currently working models used as benchmark. A range of comparisons with different forecasting models has been conducted. The forecasting results demonstrate the superiority of the proposed methodology.Keywords: short-term load forecasting, mixed effects models, neural networks, mixed effects models
Procedia PDF Downloads 19326634 Design and Comparative Analysis of Grid-Connected Bipv System with Monocrystalline Silicon and Polycrystalline Silicon in Kandahar Climate
Authors: Ahmad Shah Irshad, Naqibullah Kargar, Wais Samadi
Abstract:
Building an integrated photovoltaic (BIPV) system is a new and modern technique for solar energy production in Kandahar. Due to its location, Kandahar has abundant sources of solar energy. People use both monocrystalline and polycrystalline silicon solar PV modules for the grid-connected solar PV system, and they don’t know which technology performs better for the BIPV system. This paper analyses the parameters described by IEC61724, “Photovoltaic System Performance Monitoring Guidelines for Measurement, Data Exchange and Analysis,” to evaluate which technology shows better performance for the BIPV system. The monocrystalline silicon BIPV system has a 3.1% higher array yield than the polycrystalline silicon BIPV system. The final yield is 0.2%, somewhat higher for monocrystalline silicon than polycrystalline silicon. Monocrystalline silicon has 0.2% and 4.5% greater yearly yield factor and capacity factors than polycrystalline silicon, respectively. Monocrystalline silicon shows 0.3% better performance than polycrystalline silicon. With 1.7% reduction and 0.4% addition in collection losses and useful energy produced, respectively, monocrystalline silicon solar PV system shows good performance than polycrystalline silicon solar PV system. But system losses are the same for both technologies. The monocrystalline silicon BIPV system injects 0.2% more energy into the grid than the polycrystalline silicon BIPV system.Keywords: photovoltaic technologies, performance analysis, solar energy, solar irradiance, performance ratio
Procedia PDF Downloads 37426633 Investigating the Editing's Effect of Advertising Photos on the Virtual Purchase Decision Based on the Quantitative Electroencephalogram (EEG) Parameters
Authors: Parya Tabei, Maryam Habibifar
Abstract:
Decision-making is an important cognitive function that can be defined as the process of choosing an option among available options to achieve a specific goal. Consumer ‘need’ is the main reason for purchasing decisions. Human decision-making while buying products online is subject to various factors, one of which is the quality and effect of advertising photos. Advertising photo editing can have a significant impact on people's virtual purchase decisions. This technique helps improve the quality and overall appearance of photos by adjusting various aspects such as brightness, contrast, colors, cropping, resizing, and adding filters. This study, by examining the effect of editing advertising photos on the virtual purchase decision using EEG data, tries to investigate the effect of edited images on the decision-making of customers. A group of 30 participants were asked to react to 24 edited and unedited images while their EEG was recorded. Analysis of the EEG data revealed increased alpha wave activity in the occipital regions (O1, O2) for both edited and unedited images, which is related to visual processing and attention. Additionally, there was an increase in beta wave activity in the frontal regions (FP1, FP2, F4, F8) when participants viewed edited images, suggesting involvement in cognitive processes such as decision-making and evaluating advertising content. Gamma wave activity also increased in various regions, especially the frontal and parietal regions, which are associated with higher cognitive functions, such as attention, memory, and perception, when viewing the edited images. While the visual processing reflected by alpha waves remained consistent across different visual conditions, editing advertising photos appeared to boost neural activity in frontal and parietal regions associated with decision-making processes. These Findings suggest that photo editing could potentially influence consumer perceptions during virtual shopping experiences by modulating brain activity related to product assessment and purchase decisions.Keywords: virtual purchase decision, advertising photo, EEG parameters, decision Making
Procedia PDF Downloads 5626632 Finite Element Analysis of Human Tarsals, Meta Tarsals and Phalanges for Predicting probable location of Fractures
Authors: Irfan Anjum Manarvi, Fawzi Aljassir
Abstract:
Human bones have been a keen area of research over a long time in the field of biomechanical engineering. Medical professionals, as well as engineering academics and researchers, have investigated various bones by using medical, mechanical, and materials approaches to discover the available body of knowledge. Their major focus has been to establish properties of these and ultimately develop processes and tools either to prevent fracture or recover its damage. Literature shows that mechanical professionals conducted a variety of tests for hardness, deformation, and strain field measurement to arrive at their findings. However, they considered these results accuracy to be insufficient due to various limitations of tools, test equipment, difficulties in the availability of human bones. They proposed the need for further studies to first overcome inaccuracies in measurement methods, testing machines, and experimental errors and then carry out experimental or theoretical studies. Finite Element analysis is a technique which was developed for the aerospace industry due to the complexity of design and materials. But over a period of time, it has found its applications in many other industries due to accuracy and flexibility in selection of materials and types of loading that could be theoretically applied to an object under study. In the past few decades, the field of biomechanical engineering has also started to see its applicability. However, the work done in the area of Tarsals, metatarsals and phalanges using this technique is very limited. Therefore, present research has been focused on using this technique for analysis of these critical bones of the human body. This technique requires a 3-dimensional geometric computer model of the object to be analyzed. In the present research, a 3d laser scanner was used for accurate geometric scans of individual tarsals, metatarsals, and phalanges from a typical human foot to make these computer geometric models. These were then imported into a Finite Element Analysis software and a length refining process was carried out prior to analysis to ensure the computer models were true representatives of actual bone. This was followed by analysis of each bone individually. A number of constraints and load conditions were applied to observe the stress and strain distributions in these bones under the conditions of compression and tensile loads or their combination. Results were collected for deformations in various axis, and stress and strain distributions were observed to identify critical locations where fracture could occur. A comparative analysis of failure properties of all the three types of bones was carried out to establish which of these could fail earlier which is presented in this research. Results of this investigation could be used for further experimental studies by the academics and researchers, as well as industrial engineers, for development of various foot protection devices or tools for surgical operations and recovery treatment of these bones. Researchers could build up on these models to carryout analysis of a complete human foot through Finite Element analysis under various loading conditions such as walking, marching, running, and landing after a jump etc.Keywords: tarsals, metatarsals, phalanges, 3D scanning, finite element analysis
Procedia PDF Downloads 33226631 Composing Method of Decision-Making Function for Construction Management Using Active 4D/5D/6D Objects
Authors: Hyeon-Seung Kim, Sang-Mi Park, Sun-Ju Han, Leen-Seok Kang
Abstract:
As BIM (Building Information Modeling) application continually expands, the visual simulation techniques used for facility design and construction process information are becoming increasingly advanced and diverse. For building structures, BIM application is design - oriented to utilize 3D objects for conflict management, whereas for civil engineering structures, the usability of nD object - oriented construction stage simulation is important in construction management. Simulations of 5D and 6D objects, for which cost and resources are linked along with process simulation in 4D objects, are commonly used, but they do not provide a decision - making function for process management problems that occur on site because they mostly focus on the visual representation of current status for process information. In this study, an nD CAD system is constructed that facilitates an optimized schedule simulation that minimizes process conflict, a construction duration reduction simulation according to execution progress status, optimized process plan simulation according to project cost change by year, and optimized resource simulation for field resource mobilization capability. Through this system, the usability of conventional simple simulation objects is expanded to the usability of active simulation objects with which decision - making is possible. Furthermore, to close the gap between field process situations and planned 4D process objects, a technique is developed to facilitate a comparative simulation through the coordinated synchronization of an actual video object acquired by an on - site web camera and VR concept 4D object. This synchronization and simulation technique can also be applied to smartphone video objects captured in the field in order to increase the usability of the 4D object. Because yearly project costs change frequently for civil engineering construction, an annual process plan should be recomposed appropriately according to project cost decreases/increases compared with the plan. In the 5D CAD system provided in this study, an active 5D object utilization concept is introduced to perform a simulation in an optimized process planning state by finding a process optimized for the changed project cost without changing the construction duration through a technique such as genetic algorithm. Furthermore, in resource management, an active 6D object utilization function is introduced that can analyze and simulate an optimized process plan within a possible scope of moving resources by considering those resources that can be moved under a given field condition, instead of using a simple resource change simulation by schedule. The introduction of an active BIM function is expected to increase the field utilization of conventional nD objects.Keywords: 4D, 5D, 6D, active BIM
Procedia PDF Downloads 27926630 Detection of Keypoint in Press-Fit Curve Based on Convolutional Neural Network
Authors: Shoujia Fang, Guoqing Ding, Xin Chen
Abstract:
The quality of press-fit assembly is closely related to reliability and safety of product. The paper proposed a keypoint detection method based on convolutional neural network to improve the accuracy of keypoint detection in press-fit curve. It would provide an auxiliary basis for judging quality of press-fit assembly. The press-fit curve is a curve of press-fit force and displacement. Both force data and distance data are time-series data. Therefore, one-dimensional convolutional neural network is used to process the press-fit curve. After the obtained press-fit data is filtered, the multi-layer one-dimensional convolutional neural network is used to perform the automatic learning of press-fit curve features, and then sent to the multi-layer perceptron to finally output keypoint of the curve. We used the data of press-fit assembly equipment in the actual production process to train CNN model, and we used different data from the same equipment to evaluate the performance of detection. Compared with the existing research result, the performance of detection was significantly improved. This method can provide a reliable basis for the judgment of press-fit quality.Keywords: keypoint detection, curve feature, convolutional neural network, press-fit assembly
Procedia PDF Downloads 23326629 Estimation of the Mean of the Selected Population
Authors: Kalu Ram Meena, Aditi Kar Gangopadhyay, Satrajit Mandal
Abstract:
Two normal populations with different means and same variance are considered, where the variances are known. The population with the smaller sample mean is selected. Various estimators are constructed for the mean of the selected normal population. Finally, they are compared with respect to the bias and MSE risks by the method of Monte-Carlo simulation and their performances are analysed with the help of graphs.Keywords: estimation after selection, Brewster-Zidek technique, estimators, selected populations
Procedia PDF Downloads 51326628 Optimization Technique for the Contractor’s Portfolio in the Bidding Process
Authors: Taha Anjamrooz, Sareh Rajabi, Salwa Bheiry
Abstract:
Selection between the available projects in bidding processes for the contractor is one of the essential areas to concentrate on. It is important for the contractor to choose the right projects within its portfolio during the tendering stage based on certain criteria. It should align the bidding process with its origination strategies and goals as a screening process to have the right portfolio pool to start with. Secondly, it should set the proper framework and use a suitable technique in order to optimize its selection process for concertation purpose and higher efforts during the tender stage with goals of success and winning. In this research paper, a two steps framework proposed to increase the efficiency of the contractor’s bidding process and the winning chance of getting the new projects awarded. In this framework, initially, all the projects pass through the first stage screening process, in which the portfolio basket will be evaluated and adjusted in accordance with the organization strategies to the reduced version of the portfolio pool, which is in line with organization activities. In the second stage, the contractor uses linear programming to optimize the portfolio pool based on available resources such as manpower, light equipment, heavy equipment, financial capability, return on investment, and success rate of winning the bid. Therefore, this optimization model will assist the contractor in utilizing its internal resource to its maximum and increase its winning chance for the new project considering past experience with clients, built-relation between two parties, and complexity in the exertion of the projects. The objective of this research will be to increase the contractor's winning chance in the bidding process based on the success rate and expected return on investment.Keywords: bidding process, internal resources, optimization, contracting portfolio management
Procedia PDF Downloads 14426627 The Fabrication of Stress Sensing Based on Artificial Antibodies to Cortisol by Molecular Imprinted Polymer
Authors: Supannika Klangphukhiew, Roongnapa Srichana, Rina Patramanon
Abstract:
Cortisol has been used as a well-known commercial stress biomarker. A homeostasis response to psychological stress is indicated by an increased level of cortisol produced in hypothalamus-pituitary-adrenal (HPA) axis. Chronic psychological stress contributing to the high level of cortisol relates to several health problems. In this study, the cortisol biosensor was fabricated that mimicked the natural receptors. The artificial antibodies were prepared using molecular imprinted polymer technique that can imitate the performance of natural anti-cortisol antibody with high stability. Cortisol-molecular imprinted polymer (cortisol-MIP) was obtained using the multi-step swelling and polymerization protocol with cortisol as a target molecule combining methacrylic acid:acrylamide (2:1) with bisacryloyl-1,2-dihydroxy-1,2-ethylenediamine and ethylenedioxy-N-methylamphetamine as cross-linkers. Cortisol-MIP was integrated to the sensor. It was coated on the disposable screen-printed carbon electrode (SPCE) for portable electrochemical analysis. The physical properties of Cortisol-MIP were characterized by means of electron microscope techniques. The binding characteristics were evaluated via covalent patterns changing in FTIR spectra which were related to voltammetry response. The performance of cortisol-MIP modified SPCE was investigated in terms of detection range, high selectivity with a detection limit of 1.28 ng/ml. The disposable cortisol biosensor represented an application of MIP technique to recognize steroids according to their structures with feasibility and cost-effectiveness that can be developed to use in point-of-care.Keywords: stress biomarker, cortisol, molecular imprinted polymer, screen-printed carbon electrode
Procedia PDF Downloads 27426626 GIS for Simulating Air Traffic by Applying Different Multi-radar Positioning Techniques
Authors: Amara Rafik, Bougherara Maamar, Belhadj Aissa Mostefa
Abstract:
Radar data is one of the many data sources used by ATM Air Traffic Management systems. These data come from air navigation radar antennas. These radars intercept signals emitted by the various aircraft crossing the controlled airspace and calculate the position of these aircraft and retransmit their positions to the Air Traffic Management System. For greater reliability, these radars are positioned in such a way as to allow their coverage areas to overlap. An aircraft will therefore be detected by at least one of these radars. However, the position coordinates of the same aircraft and sent by these different radars are not necessarily identical. Therefore, the ATM system must calculate a single position (radar track) which will ultimately be sent to the control position and displayed on the air traffic controller's monitor. There are several techniques for calculating the radar track. Furthermore, the geographical nature of the problem requires the use of a Geographic Information System (GIS), i.e. a geographical database on the one hand and geographical processing. The objective of this work is to propose a GIS for traffic simulation which reconstructs the evolution over time of aircraft positions from a multi-source radar data set and by applying these different techniques.Keywords: ATM, GIS, radar data, air traffic simulation
Procedia PDF Downloads 8926625 Using Open Source Data and GIS Techniques to Overcome Data Deficiency and Accuracy Issues in the Construction and Validation of Transportation Network: Case of Kinshasa City
Authors: Christian Kapuku, Seung-Young Kho
Abstract:
An accurate representation of the transportation system serving the region is one of the important aspects of transportation modeling. Such representation often requires developing an abstract model of the system elements, which also requires important amount of data, surveys and time. However, in some cases such as in developing countries, data deficiencies, time and budget constraints do not always allow such accurate representation, leaving opportunities to assumptions that may negatively affect the quality of the analysis. With the emergence of Internet open source data especially in the mapping technologies as well as the advances in Geography Information System, opportunities to tackle these issues have raised. Therefore, the objective of this paper is to demonstrate such application through a practical case of the development of the transportation network for the city of Kinshasa. The GIS geo-referencing was used to construct the digitized map of Transportation Analysis Zones using available scanned images. Centroids were then dynamically placed at the center of activities using an activities density map. Next, the road network with its characteristics was built using OpenStreet data and other official road inventory data by intersecting their layers and cleaning up unnecessary links such as residential streets. The accuracy of the final network was then checked, comparing it with satellite images from Google and Bing. For the validation, the final network was exported into Emme3 to check for potential network coding issues. Results show a high accuracy between the built network and satellite images, which can mostly be attributed to the use of open source data.Keywords: geographic information system (GIS), network construction, transportation database, open source data
Procedia PDF Downloads 16926624 Academic Staff’s Perception and Willingness to Participate in Collaborative Research: Implication for Development in Sub-Saharan Africa
Authors: Ademola Ibukunolu Atanda
Abstract:
Research undertakings are meant to proffer solutions to issues and challenges in society. This justifies the need for research in ivory towers. Multinational and non-governmental organisations, as well as foundations, commit financial resources to support research endeavours. In recent times, the direction and dimension of research undertaking encourage collaborations, whereby experts from different disciplines or specializations would bring their expertise in addressing any identified problem, whether in humanities or sciences. However, the extent to which collaborative research undertakings are perceived and embraced by academic staff would determine the impact collaborative research would have on society. To this end, this study investigated academic staff’s perception and willingness to be involved in collaborative research for the purpose of proffering solutions to societal problems. The study adopted a descriptive research design. The population comprised academic staff in southern Nigeria. The sample was drawn through a convenient sampling technique. The data were collected using a questionnaire titled “Perception and Willingness to Participate in Collaborative Research Questionnaire (PWPCRQ)’ using Google Forms. Data collected were analyzed using descriptive statistics of simple percentages, mean and charts. The findings showed that Academic Staff’s readiness to participate in collaborative research is to a great extent (89%) and they participate in collaborative research very often (51%). The Academic Staff was involved more in collaboration research among their colleagues within their universities (1.98) than participation in inter-disciplines collaboration (1.47) with their colleagues outside Nigeria. Collaborative research was perceived to impact on development (2.5). Collaborative research offers the following benefits to members’ aggregation of views, the building of an extensive network of contacts, enhancement of sharing of skills, facilitation of tackling complex problems, increased visibility of research network and citations and promotion of funding opportunities. The study concluded that Academic staff in universities in the South-West of Nigeria participate in collaborative research but with their colleagues within Nigeria rather than outside the country. Based on the findings, it was recommended that the management of universities in South-West Nigeria should encourage collaborative research with some incentives.Keywords: collaboration, research, development, participation
Procedia PDF Downloads 6626623 Analysis of Business Intelligence Tools in Healthcare
Authors: Avishkar Gawade, Omkar Bansode, Ketan Bhambure, Bhargav Deore
Abstract:
In recent year wide range of business intelligence technology have been applied to different area in order to support decision making process BI enables extraction of knowledge from data store. BI tools usually used in public health field for financial and administrative purposes.BI uses a dashboard in presentation stage to deliver information to information to end users.In this paper,we intend to analyze some open source BI tools on the market and their applicability in the clinical sphere taking into consideration the general characteristics of the clinical environment.A pervasive BI platform was developed using a real case in order to prove the tool viability.Analysis of various BI Tools in done with the help of several parameters such as data security,data integration,data quality reporting and anlaytics,performance,scalability and cost effectivesness.Keywords: CDSS, EHR, business intelliegence, tools
Procedia PDF Downloads 14026622 Medical Ethics: Knowledge, Attitude and Practices among Young Healthcare Professionals – A Survey from Islamabad, Pakistan
Authors: Asima Mehaboob Khan, Rizwan Taj
Abstract:
Purpose: This study aims to estimate the knowledge, attitude and practices of medical ethics among young healthcare professionals. Method: A qualitative descriptive study was conducted among young healthcare professionals from both public and private sector medical institutions. Using the convenience sampling technique, 272 healthcare professionals participated in this study. A pre-structured modified questionnaire was used to collect the data. Descriptive analyses were executed for each variable. Result: About 76.47% of healthcare professional considers the importance of adequate knowledge of medical ethics, and 82.24% declared lecture, seminars and clinical discussion as the source of their medical knowledge of biomedical ethics. About 42.44% of healthcare professionals exhibited a negative attitude toward medical ethics, 57.72% showed a mildly positive attitude, whereas 1.10% and 0.74% indicated a moderately positive attitude and a highly positive attitude towards medical ethics. Similarly, the level of practice according to medical ethics is also very poor among young healthcare professionals. 34.56% of healthcare professionals deviated from medical ethics during their clinical practices, whereas 0.74% showed a good level of medical practice according to medical ethics. Conclusion: It is concluded in this research study that young healthcare professionals have adequate theoretical knowledge of medical ethics but are not properly trained to perform their clinical practices according to the guidelines of medical ethics. Furthermore, their professional attitude is poorly developed to maintain medical ethics during their clinical practices.Keywords: knowledge, attitude, practices, medical ethics
Procedia PDF Downloads 10726621 Comparative Analysis of Different Land Use Land Cover (LULC) Maps in WRF Modelling Over Indian Region
Authors: Sen Tanmoy, Jain Sarika, Panda Jagabandhu
Abstract:
The studies regarding the impact of urbanization using the WRF-ARW model rely heavily on the static geographical information selected, including domain configuration and land use land cover (LULC) data. Accurate representation of LULC data provides essential information for understanding urban growth and simulating meteorological parameters such as temperature, precipitation etc. Researchers are using different LULC data as per availability and their requirements. As far as India is concerned, we have very limited resources and data availability. So, it is important to understand how we can optimize our results using limited LULC data. In this review article, we explored how a LULC map is generated from different sources in the Indian context and what its significance is in WRF-ARW modeling to study urbanization/Climate change or any other meteorological parameters. Bibliometric analyses were also performed in this review article based on countries of study and indexed keywords. Finally, some key points are marked out for selecting the most suitable LULC map for any urbanization-related study.Keywords: LULC, LULC mapping, LANDSAT, WRF-ARW, ISRO, bibliometric Analysis.
Procedia PDF Downloads 3126620 Enhancement of 2, 4-Dichlorophenoxyacetic Acid Solubility via Solid Dispersion Technique
Authors: Tamer M. Shehata, Heba S. Elsewedy, Mashel Al Dosary, Alaa Elshehry, Mohamed A. Khedr, Maged E. Mohamed
Abstract:
Objective: 2,4-Dichlorophenoxy acetic acid (2,4-D) is a well-known herbicide widely used as a weed killer. Recently, 2,4-D was rediscovered as a new anti-inflammatory agent through in silico as well as in-vivo experiments. However, poor solubility of 2,4-D could represent a problems during pharmaceutical development in addition to lower bioavailability. Solid dispersion (SD) refers to a group of solid products consisting of at least two different components, usually a hydrophobic drug and hydrophilic matrix. It is well known technique for enhancing drug solubility. Therefore, selecting SD as a tool for enhancing 2,4-D could be of great interest to the formulator. Method: In our project, several polymers were investigated (such as PEG, HPMC, citric acid and others) in addition to drug polymer ratios and its effect on solubility. Evaluation of drug polymer interaction was investigated through both Fourier Transform Infrared (FTIR) and Differential Scanning Calorimetry (DSC). Finally, in-vivo evaluation was performed for the best selected preparation through inflammatory response of rat induce hind paw. Results: Results indicated that, citric acid 2,4-D and in ratio of 0.75 : 1 showed modified the dissolution profile of the drug. The FTIR resltes indicated no significant chemical interaction, however DSC showed shifting of the drug melting point. Finally, Carragenan induced rat hind paw edema showed significant reduction of the drug solid dispersion in comparison to the pure drug, indicating rapid and complete absorption of the drug in solid dispersion form. Conclusion: Solid dispersion technology can be utilized efficiently to enhance the solubility of 2,4-D.Keywords: solid dispersion, 2, 4-D solubility, carragenan induced edema
Procedia PDF Downloads 46026619 Efficient Implementation of Finite Volume Multi-Resolution Weno Scheme on Adaptive Cartesian Grids
Authors: Yuchen Yang, Zhenming Wang, Jun Zhu, Ning Zhao
Abstract:
An easy-to-implement and robust finite volume multi-resolution Weighted Essentially Non-Oscillatory (WENO) scheme is proposed on adaptive cartesian grids in this paper. Such a multi-resolution WENO scheme is combined with the ghost cell immersed boundary method (IBM) and wall-function technique to solve Navier-Stokes equations. Unlike the k-exact finite volume WENO schemes which involve large amounts of extra storage, repeatedly solving the matrix generated in a least-square method or the process of calculating optimal linear weights on adaptive cartesian grids, the present methodology only adds very small overhead and can be easily implemented in existing edge-based computational fluid dynamics (CFD) codes with minor modifications. Also, the linear weights of this adaptive finite volume multi-resolution WENO scheme can be any positive numbers on condition that their sum is one. It is a way of bypassing the calculation of the optimal linear weights and such a multi-resolution WENO scheme avoids dealing with the negative linear weights on adaptive cartesian grids. Some benchmark viscous problems are numerical solved to show the efficiency and good performance of this adaptive multi-resolution WENO scheme. Compared with a second-order edge-based method, the presented method can be implemented into an adaptive cartesian grid with slight modification for big Reynolds number problems.Keywords: adaptive mesh refinement method, finite volume multi-resolution WENO scheme, immersed boundary method, wall-function technique.
Procedia PDF Downloads 15226618 Entrepreneurship Education as a Pre-Requisite for Graduate Entrepreneurship: A Study of Graduate Entrepreneurs in Yenagoa City
Authors: Kurotimi M. Fems, Francis D. W. Poazi, Helen Opigo
Abstract:
The concepts of entrepreneurship education together with graduate entrepreneurship have taken centre stage in many countries as a 21st century strategy for economic growth and development. Entrepreneurship education has been viewed as a pre-requisite tool for a more effective and successful business operation. This paper seeks to verify if entrepreneurship education is pre-requisite to graduate entrepreneurship, and to ascertain if such other factors as the need for achievement, competence and experience etc. also play a foundational role in the choice of a graduate becoming an entrepreneur. The scope of the research study is entrepreneurs within Yenagoa metropolis in Bayelsa state, Nigeria. The sample target is graduates engaged in entrepreneurship activities (graduates who own and run businesses). Stratified sampling technique was used and 101 responses were obtained from a total of 300 questionnaires issued. Bar chart, tables and percentages were used to analyze the collected data. The findings revealed that personality traits, situational circumstance, need for achievement and experience/competence were the foundational factors stimulating graduate entrepreneurs to engage in entrepreneurial pursuits. Of all, personality trait showed the highest score with 73 (73%) out of 101 entrepreneurs agreeing. Experience/Competence and situational circumstances followed behind with 66 (65%) and 63 (62.4%), respectively. Entrepreneurship education revealed the least score with 33 (32.3%) out of 101 participating entrepreneurs. All hope, however, is not lost, as this shows that something can be done to increase the impact of entrepreneurship education on graduate entrepreneurship.Keywords: Creative destruction, entrepreneurs, entrepreneurship education, graduate entrepreneurship, pre-requisite
Procedia PDF Downloads 31126617 Social Networking Sites: A Platform for Communication and Collaboration for Visually Impaired
Authors: Sufia Khowaja, Nishat Fatima
Abstract:
Social networking sites are significant for visually impaired to overcome the unique challenges they face and access the resources they need to succeed in their education and beyond which might be difficult to obtain through traditional means. It provides them an opportunity to build relationships, stay connected with their support network as well as to develop social skills which give them emotional support to fell less isolated. In this connection the study is conducted with the aim to determine the use of social networking sites, purpose of using and activities performed by visually impaired at Delhi University, Delhi, Jawaharlal Nehru University, Delhi and Jamia Milia Islamia, Delhi. The study followed survey technique in which structured interview is followed to collect data from 137 visually impaired students and analysed using ‘SPSS ver23’. The findings of the study revealed that mostly used social networking sites are whatsapp by 89.23% students of DU, 95.12% of JNU, 87.09% of JMI, followed by e-mail by 78.46% of DU, 78.04% of JNU, 64.51%; youtube by 73.84% DU, 90.24% JNU, 80.64% JMI. Purpose for using these sites is for academics mentioned by 96.92% DU, 100% JNU, 93.54% JMI. Activities performed on sites are sending and receiving messaging 96.92% DU, 92.68% JNU, 93.55% JMI, communicating with friends and family as well as getting academic information. Findings of the study will be helpful for libraries to disseminate their services and resources as well as latest updates to their visually impaired users with the help of most used tools.Keywords: social networking sites, visually impaired, Delhi University, Jawaharlal Nehru University, Jamia Milia Islamia
Procedia PDF Downloads 9426616 Data Projects for “Social Good”: Challenges and Opportunities
Authors: Mikel Niño, Roberto V. Zicari, Todor Ivanov, Kim Hee, Naveed Mushtaq, Marten Rosselli, Concha Sánchez-Ocaña, Karsten Tolle, José Miguel Blanco, Arantza Illarramendi, Jörg Besier, Harry Underwood
Abstract:
One of the application fields for data analysis techniques and technologies gaining momentum is the area of social good or “common good”, covering cases related to humanitarian crises, global health care, or ecology and environmental issues, among others. The promotion of data-driven projects in this field aims at increasing the efficacy and efficiency of social initiatives, improving the way these actions help humanity in general and people in need in particular. This application field, however, poses its own barriers and challenges when developing data-driven projects, lagging behind in comparison with other scenarios. These challenges derive from aspects such as the scope and scale of the social issue to solve, cultural and political barriers, the skills of main stakeholders and the technological resources available, the motivation to be engaged in such projects, or the ethical and legal issues related to sensitive data. This paper analyzes the application of data projects in the field of social good, reviewing its current state and noteworthy initiatives, and presenting a framework covering the key aspects to analyze in such projects. The goal is to provide guidelines to understand the main challenges and opportunities for this type of data project, as well as identifying the main differential issues compared to “classical” data projects in general. A case study is presented on the initial steps and stakeholder analysis of a data project for the inclusion of refugees in the city of Frankfurt, Germany, in order to empirically confront the framework with a real example.Keywords: data-driven projects, humanitarian operations, personal and sensitive data, social good, stakeholders analysis
Procedia PDF Downloads 32926615 A Study on Green Building Certification Systems within the Context of Anticipatory Systems
Authors: Taner Izzet Acarer, Ece Ceylan Baba
Abstract:
This paper examines green building certification systems and their current processes in comparison with anticipatory systems. Rapid growth of human population and depletion of natural resources are causing irreparable damage to urban and natural environment. In this context, the concept of ‘sustainable architecture’ has emerged in the 20th century so as to establish and maintain standards for livable urban spaces, to improve quality of urban life, and to preserve natural resources for future generations. The construction industry is responsible for a large part of the resource consumption and it is believed that the ‘green building’ designs that emerge in construction industry can reduce environmental problems and contribute to sustainable development around the world. A building must meet a specific set of criteria, set forth through various certification systems, in order to be eligible for designation as a green building. It is disputable whether methods used by green building certification systems today truly serve the purposes of creating a sustainable world. Accordingly, this study will investigate the sets of rating systems used by the most popular green building certification programs, including LEED (Leadership in Energy and Environmental Design), BREEAM (Building Research Establishment's Environmental Assessment Methods), DGNB (Deutsche Gesellschaft für Nachhaltiges Bauen System), in terms of ‘Anticipatory Systems’ in accordance with the certification processes and their goals, while discussing their contribution to architecture. The basic methodology of the study is as follows. Firstly analyzes of brief historical and literature review of green buildings and certificate systems will be stated. Secondly, processes of green building certificate systems will be disputed by the help of anticipatory systems. Anticipatory Systems is a set of systems designed to generate action-oriented projections and to forecast potential side effects using the most current data. Anticipatory Systems pull the future into the present and take action based on future predictions. Although they do not have a claim to see into the future, they can provide foresight data. When shaping the foresight data, Anticipatory Systems use feedforward instead of feedback, enabling them to forecast the system’s behavior and potential side effects by establishing a correlation between the system’s present/past behavior and projected results. This study indicates the goals and current status of LEED, BREEAM and DGNB rating systems that created by using the feedback technique will be examined and presented in a chart. In addition, by examining these rating systems with the anticipatory system that using the feedforward method, the negative influences of the potential side effects on the purpose and current status of the rating systems will be shown in another chart. By comparing the two obtained data, the findings will be shown that rating systems are used for different goals than the purposes they are aiming for. In conclusion, the side effects of green building certification systems will be stated by using anticipatory system models.Keywords: anticipatory systems, BREEAM, certificate systems, DGNB, green buildings, LEED
Procedia PDF Downloads 22126614 Slugging Frequency Correlation for High Viscosity Oil-Gas Flow in Horizontal Pipeline
Authors: B. Y. Danjuma, A. Archibong-Eso, Aliyu M. Aliyu, H. Yeung
Abstract:
In this experimental investigation, a new data for slugging frequency for high viscosity oil-gas flow are reported. Scale experiments were carried out using a mixture of air and mineral oil as the liquid phase in a 17 m long horizontal pipe with 0.0762 ID. The data set was acquired using two high-speed Gamma Densitometers at a data acquisition frequency of 250 Hz over a time interval of 30 seconds. For the range of flow conditions investigated, increase in liquid oil viscosity was observed to strongly influence the slug frequency. A comparison of the present data with prediction models available in the literature revealed huge discrepancies. A new correlation incorporating the effect of viscosity on slug frequency has been proposed for the horizontal flow, which represents the main contribution of this work.Keywords: gamma densitometer, flow pattern, pressure gradient, slug frequency
Procedia PDF Downloads 41626613 Transferring Data from Glucometer to Mobile Device via Bluetooth with Arduino Technology
Authors: Tolga Hayit, Ucman Ergun, Ugur Fidan
Abstract:
Being healthy is undoubtedly an indispensable necessity for human life. With technological improvements, in the literature, various health monitoring and imaging systems have been developed to satisfy your health needs. In this context, the work of monitoring and recording the data of individual health monitoring data via wireless technology is also being part of these studies. Nowadays, mobile devices which are located in almost every house and which become indispensable of our life and have wireless technology infrastructure have an important place of making follow-up health everywhere and every time because these devices were using in the health monitoring systems. In this study, Arduino an open-source microcontroller card was used in which a sample sugar measuring device was connected in series. In this way, the glucose data (glucose ratio, time) obtained with the glucometer is transferred to the mobile device based on the Android operating system with the Bluetooth technology channel. A mobile application was developed using the Apache Cordova framework for listing data, presenting graphically and reading data over Arduino. Apache Cordova, HTML, Javascript and CSS are used in coding section. The data received from the glucometer is stored in the local database of the mobile device. It is intended that people can transfer their measurements to their mobile device by using wireless technology and access the graphical representations of their data. In this context, the aim of the study is to be able to perform health monitoring by using different wireless technologies in mobile devices that can respond to different wireless technologies at present. Thus, that will contribute the other works done in this area.Keywords: Arduino, Bluetooth, glucose measurement, mobile health monitoring
Procedia PDF Downloads 32626612 Analyzing the Risk Based Approach in General Data Protection Regulation: Basic Challenges Connected with Adapting the Regulation
Authors: Natalia Kalinowska
Abstract:
The adoption of the General Data Protection Regulation, (GDPR) finished the four-year work of the European Commission in this area in the European Union. Considering far-reaching changes, which will be applied by GDPR, the European legislator envisaged two-year transitional period. Member states and companies have to prepare for a new regulation until 25 of May 2018. The idea, which becomes a new look at an attitude to data protection in the European Union is risk-based approach. So far, as a result of implementation of Directive 95/46/WE, in many European countries (including Poland) there have been adopted very particular regulations, specifying technical and organisational security measures e.g. Polish implementing rules indicate even how long password should be. According to the new approach from May 2018, controllers and processors will be obliged to apply security measures adequate to level of risk associated with specific data processing. The risk in GDPR should be interpreted as the likelihood of a breach of the rights and freedoms of the data subject. According to Recital 76, the likelihood and severity of the risk to the rights and freedoms of the data subject should be determined by reference to the nature, scope, context and purposes of the processing. GDPR does not indicate security measures which should be applied – in recitals there are only examples such as anonymization or encryption. It depends on a controller’s decision what type of security measures controller considered as sufficient and he will be responsible if these measures are not sufficient or if his identification of risk level is incorrect. Data protection regulation indicates few levels of risk. Recital 76 indicates risk and high risk, but some lawyers think, that there is one more category – low risk/now risk. Low risk/now risk data processing is a situation when it is unlikely to result in a risk to the rights and freedoms of natural persons. GDPR mentions types of data processing when a controller does not have to evaluate level of risk because it has been classified as „high risk” processing e.g. processing on a large scale of special categories of data, processing with using new technologies. The methodology will include analysis of legal regulations e.g. GDPR, the Polish Act on the Protection of personal data. Moreover: ICO Guidelines and articles concerning risk based approach in GDPR. The main conclusion is that an appropriate risk assessment is a key to keeping data safe and avoiding financial penalties. On the one hand, this approach seems to be more equitable, not only for controllers or processors but also for data subjects, but on the other hand, it increases controllers’ uncertainties in the assessment which could have a direct impact on incorrect data protection and potential responsibility for infringement of regulation.Keywords: general data protection regulation, personal data protection, privacy protection, risk based approach
Procedia PDF Downloads 25426611 Monitoring Blood Pressure Using Regression Techniques
Authors: Qasem Qananwah, Ahmad Dagamseh, Hiam AlQuran, Khalid Shaker Ibrahim
Abstract:
Blood pressure helps the physicians greatly to have a deep insight into the cardiovascular system. The determination of individual blood pressure is a standard clinical procedure considered for cardiovascular system problems. The conventional techniques to measure blood pressure (e.g. cuff method) allows a limited number of readings for a certain period (e.g. every 5-10 minutes). Additionally, these systems cause turbulence to blood flow; impeding continuous blood pressure monitoring, especially in emergency cases or critically ill persons. In this paper, the most important statistical features in the photoplethysmogram (PPG) signals were extracted to estimate the blood pressure noninvasively. PPG signals from more than 40 subjects were measured and analyzed and 12 features were extracted. The features were fed to principal component analysis (PCA) to find the most important independent features that have the highest correlation with blood pressure. The results show that the stiffness index means and standard deviation for the beat-to-beat heart rate were the most important features. A model representing both features for Systolic Blood Pressure (SBP) and Diastolic Blood Pressure (DBP) was obtained using a statistical regression technique. Surface fitting is used to best fit the series of data and the results show that the error value in estimating the SBP is 4.95% and in estimating the DBP is 3.99%.Keywords: blood pressure, noninvasive optical system, principal component analysis, PCA, continuous monitoring
Procedia PDF Downloads 16226610 UAV’s Enhanced Data Collection for Heterogeneous Wireless Sensor Networks
Authors: Kamel Barka, Lyamine Guezouli, Assem Rezki
Abstract:
In this article, we propose a protocol called DataGA-DRF (a protocol for Data collection using a Genetic Algorithm through Dynamic Reference Points) that collects data from Heterogeneous wireless sensor networks. This protocol is based on DGA (Destination selection according to Genetic Algorithm) to control the movement of the UAV (Unmanned aerial vehicle) between dynamic reference points that virtually represent the sensor node deployment. The dynamics of these points ensure an even distribution of energy consumption among the sensors and also improve network performance. To determine the best points, DataGA-DRF uses a classification algorithm such as K-Means.Keywords: heterogeneous wireless networks, unmanned aerial vehicles, reference point, collect data, genetic algorithm
Procedia PDF Downloads 8626609 Flexural Behavior of Eco-Friendly Prefabricated Low Cost Bamboo Reinforced Wall Panels
Authors: Vishal Puri, Pradipta Chakrabortty, Swapan Majumdar
Abstract:
Precast concrete construction is the most commonly used technique for a rapid construction. This technique is very frequently used in the developed countries. Different guidelines required to utilize the potential of prefabricated construction are still not available in the developing countries. This causes over dependence on in-situ construction procedure which further affects the quality, scheduling, and duration of construction. Also with the ever increasing costs of building materials and their negative impact on the environment it has become imperative to look out for alternate construction materials which are cheap and sustainable. Bamboo and fly ash are alternate construction materials having great potential in the construction industry. Thus there is a great need to develop prefabricated components by utilizing the potential of these materials. Bamboo reinforced beams, bamboo reinforced columns and bamboo arches as researched previously have shown great prospects for prefabricated construction industry. But, many other prefabricated components still need to be studied and widely tested before their utilization in the prefabricated construction industry. In the present study, authors have showcased prefabricated bamboo reinforced wall panel for the prefabricated construction industry. It presents a detailed methodology for the development of such prefabricated panels. It also presents the flexural behavior of such panels as tested under flexural loads following ASTM guidelines. It was observed that these wall panels are much flexible and do not show brittle failure as observed in traditional brick walls. It was observed that prefabricated walls are about 42% cheaper as compared to conventional brick walls. It was also observed that prefabricated walls are considerably lighter in weight and are environment friendly. It was thus concluded that this type of wall panels are an excellent alternative for partition brick walls.Keywords: bamboo, prefabricated walls, reinforced structure, sustainable infrastructure
Procedia PDF Downloads 31326608 Antibacterial Wound Dressing Based on Metal Nanoparticles Containing Cellulose Nanofibers
Authors: Mohamed Gouda
Abstract:
Antibacterial wound dressings based on cellulose nanofibers containing different metal nanoparticles (CMC-MNPs) were synthesized using an electrospinning technique. First, the composite of carboxymethyl cellulose containing different metal nanoparticles (CMC/MNPs), such as copper nanoparticles (CuNPs), iron nanoparticles (FeNPs), zinc nanoparticles (ZnNPs), cadmium nanoparticles (CdNPs) and cobalt nanoparticles (CoNPs) were synthesized, and finally, these composites were transferred to the electrospinning process. Synthesized CMC-MNPs were characterized using scanning electron microscopy (SEM) coupled with high-energy dispersive X-ray (EDX) and UV-visible spectroscopy used to confirm nanoparticle formation. The SEM images clearly showed regular flat shapes with semi-porous surfaces. All MNPs were well distributed inside the backbone of the cellulose without aggregation. The average particle diameters were 29-39 nm for ZnNPs, 29-33 nm for CdNPs, 25-33 nm for CoNPs, 23-27 nm for CuNPs and 22-26 nm for FeNPs. Surface morphology, water uptake and release of MNPs from the nanofibers in water and antimicrobial efficacy were studied. SEM images revealed that electrospun CMC-MNPs nanofibers are smooth and uniformly distributed without bead formation with average fiber diameters in the range of 300 to 450 nm. Fiber diameters were not affected by the presence of MNPs. TEM images showed that MNPs are present in/on the electrospun CMC-MNPs nanofibers. The diameter of the electrospun nanofibers containing MNPs was in the range of 300–450 nm. The MNPs were observed to be spherical in shape. The CMC-MNPs nanofibers showed good hydrophilic properties and had excellent antibacterial activity against the Gram-negative bacteria Escherichia coli and the Gram-positive bacteria Staphylococcus aureus.Keywords: electrospinning technique, metal nanoparticles, cellulosic nanofibers, wound dressing
Procedia PDF Downloads 33226607 Product Features Extraction from Opinions According to Time
Authors: Kamal Amarouche, Houda Benbrahim, Ismail Kassou
Abstract:
Nowadays, e-commerce shopping websites have experienced noticeable growth. These websites have gained consumers’ trust. After purchasing a product, many consumers share comments where opinions are usually embedded about the given product. Research on the automatic management of opinions that gives suggestions to potential consumers and portrays an image of the product to manufactures has been growing recently. After launching the product in the market, the reviews generated around it do not usually contain helpful information or generic opinions about this product (e.g. telephone: great phone...); in the sense that the product is still in the launching phase in the market. Within time, the product becomes old. Therefore, consumers perceive the advantages/ disadvantages about each specific product feature. Therefore, they will generate comments that contain their sentiments about these features. In this paper, we present an unsupervised method to extract different product features hidden in the opinions which influence its purchase, and that combines Time Weighting (TW) which depends on the time opinions were expressed with Term Frequency-Inverse Document Frequency (TF-IDF). We conduct several experiments using two different datasets about cell phones and hotels. The results show the effectiveness of our automatic feature extraction, as well as its domain independent characteristic.Keywords: opinion mining, product feature extraction, sentiment analysis, SentiWordNet
Procedia PDF Downloads 417