Search results for: data driven decision making
24454 A Study on the Establishment of a 4-Joint Based Motion Capture System and Data Acquisition
Authors: Kyeong-Ri Ko, Seong Bong Bae, Jang Sik Choi, Sung Bum Pan
Abstract:
A simple method for testing the posture imbalance of the human body is to check for differences in the bilateral shoulder and pelvic height of the target. In this paper, to check for spinal disorders the authors have studied ways to establish a motion capture system to obtain and express motions of 4-joints, and to acquire data based on this system. The 4 sensors are attached to the both shoulders and pelvis. To verify the established system, the normal and abnormal postures of the targets listening to a lecture were obtained using the established 4-joint based motion capture system. From the results, it was confirmed that the motions taken by the target was identical to the 3-dimensional simulation.Keywords: inertial sensor, motion capture, motion data acquisition, posture imbalance
Procedia PDF Downloads 51924453 Analysis of Rural Roads in Developing Countries Using Principal Component Analysis and Simple Average Technique in the Development of a Road Safety Performance Index
Authors: Muhammad Tufail, Jawad Hussain, Hammad Hussain, Imran Hafeez, Naveed Ahmad
Abstract:
Road safety performance index is a composite index which combines various indicators of road safety into single number. Development of a road safety performance index using appropriate safety performance indicators is essential to enhance road safety. However, a road safety performance index in developing countries has not been given as much priority as needed. The primary objective of this research is to develop a general Road Safety Performance Index (RSPI) for developing countries based on the facility as well as behavior of road user. The secondary objectives include finding the critical inputs in the RSPI and finding the better method of making the index. In this study, the RSPI is developed by selecting four main safety performance indicators i.e., protective system (seat belt, helmet etc.), road (road width, signalized intersections, number of lanes, speed limit), number of pedestrians, and number of vehicles. Data on these four safety performance indicators were collected using observation survey on a 20 km road section of the National Highway N-125 road Taxila, Pakistan. For the development of this composite index, two methods are used: a) Principal Component Analysis (PCA) and b) Equal Weighting (EW) method. PCA is used for extraction, weighting, and linear aggregation of indicators to obtain a single value. An individual index score was calculated for each road section by multiplication of weights and standardized values of each safety performance indicator. However, Simple Average technique was used for weighting and linear aggregation of indicators to develop a RSPI. The road sections are ranked according to RSPI scores using both methods. The two weighting methods are compared, and the PCA method is found to be much more reliable than the Simple Average Technique.Keywords: indicators, aggregation, principle component analysis, weighting, index score
Procedia PDF Downloads 16724452 21st Century Biotechnological Research and Development Advancements for Industrial Development in India
Authors: Monisha Isaac
Abstract:
Biotechnology is a discipline which explains the use of living organisms and systems to construct a product, or we can define it as an application or technology developed to use biological systems and organisms processes for a specific use. Particularly, it includes cells and its components use for new technologies and inventions. The tools developed can be further used in diverse fields such as agriculture, industry, research and hospitals etc. The 21st century has seen a drastic development and advancement in biotechnology in India. Significant increase in Government of India’s outlays for biotechnology over the past decade has been observed. A sectoral break up of biotechnology-based companies in India shows that most of the companies are agriculture-based companies having interests ranging from tissue culture to biopesticides. Major attention has been given by the companies in health related activities and in environmental biotechnology. The biopharmaceutical, which comprises of vaccines, diagnostic, and recombinant products is the most reliable and largest segment of the Indian Biotech industry. India has developed its vaccine markets and supplies them to various countries. Then there are the bio-services, which mainly comprise of contract researches and manufacturing services. India has made noticeable developments in the field of bio industries including manufacturing of enzymes, biofuels and biopolymers. Biotechnology is also playing a crucial and significant role in the field of agriculture. Traditional methods have been replaced by new technologies that mainly focus on GM crops, marker assisted technologies and the use of biotechnological tools to improve the quality of fertilizers and soil. It may only be a small contributor but has shown to have huge potential for growth. Bioinformatics is a computational method which helps to store, manage, arrange and design tools to interpret the extensive data gathered through experimental trials, making it important in the design of drugs.Keywords: biotechnology, advancement, agriculture, bio-services, bio-industries, bio-pharmaceuticals
Procedia PDF Downloads 24224451 Predictive Analytics in Oil and Gas Industry
Authors: Suchitra Chnadrashekhar
Abstract:
Earlier looked as a support function in an organization information technology has now become a critical utility to manage their daily operations. Organizations are processing huge amount of data which was unimaginable few decades before. This has opened the opportunity for IT sector to help industries across domains to handle the data in the most intelligent manner. Presence of IT has been a leverage for the Oil & Gas industry to store, manage and process the data in most efficient way possible thus deriving the economic value in their day-to-day operations. Proper synchronization between Operational data system and Information Technology system is the need of the hour. Predictive analytics supports oil and gas companies by addressing the challenge of critical equipment performance, life cycle, integrity, security, and increase their utilization. Predictive analytics go beyond early warning by providing insights into the roots of problems. To reach their full potential, oil and gas companies need to take a holistic or systems approach towards asset optimization and thus have the functional information at all levels of the organization in order to make the right decisions. This paper discusses how the use of predictive analysis in oil and gas industry is redefining the dynamics of this sector. Also, the paper will be supported by real time data and evaluation of the data for a given oil production asset on an application tool, SAS. The reason for using SAS as an application for our analysis is that SAS provides an analytics-based framework to improve uptimes, performance and availability of crucial assets while reducing the amount of unscheduled maintenance, thus minimizing maintenance-related costs and operation disruptions. With state-of-the-art analytics and reporting, we can predict maintenance problems before they happen and determine root causes in order to update processes for future prevention.Keywords: hydrocarbon, information technology, SAS, predictive analytics
Procedia PDF Downloads 36724450 Urban Change Detection and Pattern Analysis Using Satellite Data
Authors: Shivani Jha, Klaus Baier, Rafiq Azzam, Ramakar Jha
Abstract:
In India, generally people migrate from rural area to the urban area for better infra-structural facilities, high standard of living, good job opportunities and advanced transport/communication availability. In fact, unplanned urban development due to migration of people causes seriou damage to the land use, water pollution and available water resources. In the present work, an attempt has been made to use satellite data of different years for urban change detection of Chennai metropolitan city along with pattern analysis to generate future scenario of urban development using buffer zoning in GIS environment. In the analysis, SRTM (30m) elevation data and IRS-1C satellite data for the years 1990, 2000, and 2014, are used. The flow accumulation, aspect, flow direction and slope maps developed using SRTM 30 m data are very useful for finding suitable urban locations for industrial setup and urban settlements. Normalized difference vegetation index (NDVI) and Principal Component Analysis (PCA) have been used in ERDAS imagine software for change detection in land use of Chennai metropolitan city. It has been observed that the urban area has increased exponentially in Chennai metropolitan city with significant decrease in agriculture and barren lands. However, the water bodies located in the study regions are protected and being used as freshwater for drinking purposes. Using buffer zone analysis in GIS environment, it has been observed that the development has taken place in south west direction significantly and will do so in future.Keywords: urban change, satellite data, the Chennai metropolis, change detection
Procedia PDF Downloads 41324449 Finding Optimal Solutions to Management Problems with the use of Econometric and Multiobjective Programming
Authors: M. Moradi Dalini, M. R. Talebi
Abstract:
This research revolves around a technical method according to combines econometric and multiobjective programming to select and obtain optimal solutions to management problems. It is taken for a generation that; it is important to analyze which combination of values of the explanatory variables -in an econometric method- would point to the simultaneous achievement of the best values of the response variables. In this case, if a certain degree of conflict is viewed among the response variables, we suggest a multiobjective method in order to the results obtained from a regression analysis. In fact, with the use of a multiobjective method, we will have the best decision about the conflicting relationship between the response variables and the optimal solution. The combined multiobjective programming and econometrics benefit is an assessment of a balanced “optimal” situation among them because a find of information can hardly be extracted just by econometric techniques.Keywords: econometrics, multiobjective optimization, management problem, optimization
Procedia PDF Downloads 8624448 The Effects of Green Logistics Management Practices on Sustainability Performance in Nigeria
Authors: Ozoemelam Ikechukwu Lazarus, Nizamuddin B. Zainuddi, Abdul Kafi
Abstract:
Numerous studies have been carried out on Green Logistics Management Practices (GLMPs) across the globe. The study on the practices and performance of green chain practices in Africa in particular has not gained enough scholarly attention. Again, the majority of supply chain sustainability research being conducted focus on environmental sustainability. Logistics has been a major cause of supply chain resource waste and environmental damage. Many sectors of the economy that engage in logistical operations significantly rely on vehicles, which emit pollutants into the environment. Due to urbanization and industrialization, the logistical operations of manufacturing companies represent a serious hazard to the society and human life, making the sector one of the fastest expanding in the world today. Logistics companies are faced with numerous difficulties when attempting to implement logistics practices along their supply chains. In Nigeria, manufacturing companies aspire to implement reverse logistics in response to stakeholders’ requirements to reduce negative environmental consequences. However, implementing this is impeded by a criteria framework, and necessitates the careful analysis of how such criteria interact with each other in the presence of uncertainty. This study integrates most of the green logistics management practices (GLMPs) into the Nigerian firms to improve generalizability, and credibility. It examines the effect of Green Logistics Management Practices on environmental performance, social performance, market performance, and financial performance in the logistics industries. It seeks to identify the critical success factors in order to develop a model that incorporates different factors from the perspectives of the technology, organization, human and environment to inform the adoption and use of technologies for logistics supply chain social sustainability in Nigeria. It uses exploratory research approach to collect and analyse the data.Keywords: logistics, management, sustainability, environment, operations
Procedia PDF Downloads 8824447 Balanced Score Card a Tool to Improve Naac Accreditation – a Case Study in Indian Higher Education
Authors: CA Kishore S. Peshori
Abstract:
Introduction: India, a country with vast diversity and huge population is going to have largest young population by 2020. Higher education has and will always be the basic requirement for making a developing nation to a developed nation. To improve any system it needs to be bench-marked. There have been various tools for bench-marking the systems. Education is delivered in India by universities which are mainly funded by government. This universities for delivering the education sets up colleges which are again funded mainly by government. Recently however there has also been autonomy given to universities and colleges. Moreover foreign universities are waiting to enter Indian boundaries. With a large number of universities and colleges it has become more and more necessary to measure this institutes for bench-marking. There have been various tools for measuring the institute. In India college assessments have been made compulsory by UGC. Naac has been offically recognised as the accrediation criteria. The Naac criteria has been based on seven criterias namely: 1. Curricular assessments, 2. Teaching learning and evaluation, 3. Research Consultancy and Extension, 4. Infrastructure and learning resources, 5. Student support and progression, 6. Governance leadership and management, 7. Innovation and best practices. The Naac tries to bench mark the institution for identification, sustainability, dissemination and adaption of best practices. It grades the institution according to this seven criteria and the funding of institution is based on these grades. Many of the colleges are struggling to get best of grades but they have not come across a systematic tool to achieve the results. Balanced Scorecard developed by Kaplan has been a successful tool for corporates to develop best of practices so as to increase their financial performance and also retain and increase their customers so as to grow the organization to next level.It is time to test this tool for an educational institute. Methodology: The paper tries to develop a prototype for college based on the secondary data. Once a prototype is developed the researcher based on questionnaire will try to test this tool for successful implementation. The success of this research will depend on its implementation of BSC on an institute and its grading improved due to this successful implementation. Limitation of time is a major constraint in this research as Naac cycle takes minimum 4 years for accreditation and reaccreditation the methodology will limit itself to secondary data and questionnaire to be circulated to colleges along with the prototype model of BSC. Conclusion: BSC is a successful tool for enhancing growth of an organization. Educational institutes are no exception to these. BSC will only have to be realigned to suit the Naac criteria. Once this prototype is developed the success will be tested only on its implementation but this research paper will be the first step towards developing this tool and will also initiate the success by developing a questionnaire and getting and evaluating the responses for moving to the next level of actual implementationKeywords: balanced scorecard, bench marking, Naac, UGC
Procedia PDF Downloads 27724446 Strategic Analysis of Hospitality Marketing Driven by Culture in Historical City: Comparable Case Studies in Tainan City
Authors: Tsung-Han Lin, Chia-Han Yang
Abstract:
Recently, it’s obvious that many hotels not only provide customized service, but offer local culture to give customers different experiences. Luxury decoration and fancy equipment are not enough anymore. Customers expect that hotels not just as an accommodation, but connecting to their trips. Therefore, culture becomes significant when developing marketing strategy for hotels. Tainan, located in south of Taiwan, is the city full of culture that some hospitality industries promote their hotels exactly base on culture. Historic Tainan (台南; táinán), is a city of ancient monuments, delicious food and, above all, temples: there are more gods worshipped and more festivals and rituals observed in Tainan than in any other place in Taiwan. The oldest and most absorbing parts of Tainan are historic Anping, on the west side of town by the sea, and the cultural zones in the heart of the old city; the latter were created specifically to make things easier for visitors, with city information, signs and maps tailored to each zone and well-marked in English. The Chihkan, Dong-an Fang, Five Canals and Confucius Temple cultural zones contain the richest concentration of sights – reckon on spending at least two days to do them justice. As a result, the study aims to analyze the significance of culture on marketing strategies and C and H two hotels in Tainan city as case studies conducting the comparison of cultural marketing and experience marketing to provide a framework for hotels to develop their marketing strategies.Keywords: cultural marketing, hospitality, historical city, Tainan city
Procedia PDF Downloads 45424445 Design and Implementation of the Embedded Control System for the Electrical Motor Based Cargo Vehicle
Authors: Syed M. Rizvi, Yiqing Meng, Simon Iwnicki
Abstract:
With an increased demand in the land cargo industry, it is predicted that the freight trade will rise to a record $1.1 trillion in revenue and volume in the following years to come. This increase is mainly driven by the e-commerce model ever so popular in the consumer market. Many innovative ideas have stemmed from this demand and change in lifestyle likes of which include e-bike cargo and drones. Rural and urban areas are facing air quality challenges to keep pollution levels in city centre to a minimum. For this purpose, this paper presents the design and implementation of a non-linear PID control system, employing a micro-controller and low cost sensing technique, for controlling an electrical motor based cargo vehicle with various loads, to follow a leading vehicle (bike). Within using this system, the cargo vehicle will have no load influence on the bike rider on different gradient conditions, such as hill climbing. The system is being integrated with a microcontroller to continuously measure several parameters such as relative displacement between bike and the cargo vehicle and gradient of the road, and process these measurements to create a portable controller capable of controlling the performance of electrical vehicle without the need of a PC. As a result, in the case of carrying 180kg of parcel weight, the cargo vehicle can maintain a reasonable spacing over a short length of sensor travel between the bike and itself.Keywords: cargo, e-bike, microcontroller, embedded system, nonlinear pid, self-adaptive, inertial measurement unit (IMU)
Procedia PDF Downloads 21324444 Optimum Dimensions of Hydraulic Structures Foundation and Protections Using Coupled Genetic Algorithm with Artificial Neural Network Model
Authors: Dheyaa W. Abbood, Rafa H. AL-Suhaili, May S. Saleh
Abstract:
A model using the artificial neural networks and genetic algorithm technique is developed for obtaining optimum dimensions of the foundation length and protections of small hydraulic structures. The procedure involves optimizing an objective function comprising a weighted summation of the state variables. The decision variables considered in the optimization are the upstream and downstream cutoffs length sand their angles of inclination, the foundation length, and the length of the downstream soil protection. These were obtained for a given maximum difference in head, depth of impervious layer and degree of anisotropy.The optimization carried out subjected to constraints that ensure a safe structure against the uplift pressure force and sufficient protection length at the downstream side of the structure to overcome an excessive exit gradient. The Geo-studios oft ware, was used to analyze 1200 different cases. For each case the length of protection and volume of structure required to satisfy the safety factors mentioned previously were estimated. An ANN model was developed and verified using these cases input-output sets as its data base. A MatLAB code was written to perform a genetic algorithm optimization modeling coupled with this ANN model using a formulated optimization model. A sensitivity analysis was done for selecting the cross-over probability, the mutation probability and level ,the number of population, the position of the crossover and the weights distribution for all the terms of the objective function. Results indicate that the most factor that affects the optimum solution is the number of population required. The minimum value that gives stable global optimum solution of this parameters is (30000) while other variables have little effect on the optimum solution.Keywords: inclined cutoff, optimization, genetic algorithm, artificial neural networks, geo-studio, uplift pressure, exit gradient, factor of safety
Procedia PDF Downloads 32624443 Nature of Body Image Distortion in Eating Disorders
Authors: Katri K. Cornelissen, Lise Gulli Brokjob, Kristofor McCarty, Jiri Gumancik, Martin J. Tovee, Piers L. Cornelissen
Abstract:
Recent research has shown that body size estimation of healthy women is driven by independent attitudinal and perceptual components. The attitudinal component represents psychological concerns about body, coupled to low self-esteem and a tendency towards depressive symptomatology, leading to over-estimation of body size, independent of the Body Mass Index (BMI) someone actually has. The perceptual component is a normal bias known as contraction bias, which, for bodies is dependent on actual BMI. Women with a BMI less than the population norm tend to overestimate their size, while women with a BMI greater than the population norm tend to underestimate their size. Women whose BMI is close to the population mean are most accurate. This is indexed by a regression of estimated BMI on actual BMI with a slope less than one. It is well established that body dissatisfaction, i.e. an attitudinal distortion, leads to body size overestimation in eating disordered individuals. However, debate persists as to whether women with eating disorders may also suffer a perceptual body distortion. Therefore, the current study was set to ask whether women with eating disorders exhibit the normal contraction bias when they estimate their own body size. If they do not, this would suggest differences in the way that women with eating disorders process the perceptual aspects of body shape and size in comparison to healthy controls. 100 healthy controls and 33 women with a history of eating disorders were recruited. Critically, it was ensured that both groups of participants represented comparable and adequate ranges of actual BMI (e.g. ~18 to ~40). Of those with eating disorders, 19 had a history of anorexia nervosa, 6 bulimia nervosa, and 8 OSFED. 87.5% of the women with a history of eating disorders self-reported that they were either recovered or recovering, and 89.7% of them self-reported that they had had one or more instances of relapse. The mean time lapsed since first diagnosis was 5 years and on average participants had experienced two relapses. Participants were asked to fill number of psychometric measures (EDE-Q, BSQ, RSE, BDI) to establish the attitudinal component of their body image as well as their tendency to internalize socio-cultural body ideals. Additionally, participants completed a method of adjustment psychophysical task, using photorealistic avatars calibrated for BMI, in order to provide an estimate of their own body size and shape. The data from the healthy controls replicate previous findings, revealing independent contributions to body size estimation from both attitudinal and perceptual (i.e. contraction bias) body image components, as described above. For the eating disorder group, once the adequacy of their actual BMI ranges was established, a regression of estimated BMI on actual BMI had a slope greater than 1, significantly different to that from controls. This suggests that (some) eating disordered individuals process the perceptual aspects of body image differently from healthy controls. It therefore is necessary to develop interventions which are specific to the perceptual processing of body shape and size for the management of (some) individuals with eating disorders.Keywords: body image distortion, perception, recovery, relapse, BMI, eating disorders
Procedia PDF Downloads 7024442 Geographic Information Systems and Remotely Sensed Data for the Hydrological Modelling of Mazowe Dam
Authors: Ellen Nhedzi Gozo
Abstract:
Unavailability of adequate hydro-meteorological data has always limited the analysis and understanding of hydrological behaviour of several dam catchments including Mazowe Dam in Zimbabwe. The problem of insufficient data for Mazowe Dam catchment analysis was solved by extracting catchment characteristics and aerial hydro-meteorological data from ASTER, LANDSAT, Shuttle Radar Topographic Mission SRTM remote sensing (RS) images using ILWIS, ArcGIS and ERDAS Imagine geographic information systems (GIS) software. Available observed hydrological as well as meteorological data complemented the use of the remotely sensed information. Ground truth land cover was mapped using a Garmin Etrex global positioning system (GPS) system. This information was then used to validate land cover classification detail that was obtained from remote sensing images. A bathymetry survey was conducted using a SONAR system connected to GPS. Hydrological modelling using the HBV model was then performed to simulate the hydrological process of the catchment in an effort to verify the reliability of the derived parameters. The model output shows a high Nash-Sutcliffe Coefficient that is close to 1 indicating that the parameters derived from remote sensing and GIS can be applied with confidence in the analysis of Mazowe Dam catchment.Keywords: geographic information systems, hydrological modelling, remote sensing, water resources management
Procedia PDF Downloads 34224441 Hybridization of Manually Extracted and Convolutional Features for Classification of Chest X-Ray of COVID-19
Authors: M. Bilal Ishfaq, Adnan N. Qureshi
Abstract:
COVID-19 is the most infectious disease these days, it was first reported in Wuhan, the capital city of Hubei in China then it spread rapidly throughout the whole world. Later on 11 March 2020, the World Health Organisation (WHO) declared it a pandemic. Since COVID-19 is highly contagious, it has affected approximately 219M people worldwide and caused 4.55M deaths. It has brought the importance of accurate diagnosis of respiratory diseases such as pneumonia and COVID-19 to the forefront. In this paper, we propose a hybrid approach for the automated detection of COVID-19 using medical imaging. We have presented the hybridization of manually extracted and convolutional features. Our approach combines Haralick texture features and convolutional features extracted from chest X-rays and CT scans. We also employ a minimum redundancy maximum relevance (MRMR) feature selection algorithm to reduce computational complexity and enhance classification performance. The proposed model is evaluated on four publicly available datasets, including Chest X-ray Pneumonia, COVID-19 Pneumonia, COVID-19 CTMaster, and VinBig data. The results demonstrate high accuracy and effectiveness, with 0.9925 on the Chest X-ray pneumonia dataset, 0.9895 on the COVID-19, Pneumonia and Normal Chest X-ray dataset, 0.9806 on the Covid CTMaster dataset, and 0.9398 on the VinBig dataset. We further evaluate the effectiveness of the proposed model using ROC curves, where the AUC for the best-performing model reaches 0.96. Our proposed model provides a promising tool for the early detection and accurate diagnosis of COVID-19, which can assist healthcare professionals in making informed treatment decisions and improving patient outcomes. The results of the proposed model are quite plausible and the system can be deployed in a clinical or research setting to assist in the diagnosis of COVID-19.Keywords: COVID-19, feature engineering, artificial neural networks, radiology images
Procedia PDF Downloads 8024440 A Bayesian Model with Improved Prior in Extreme Value Problems
Authors: Eva L. Sanjuán, Jacinto Martín, M. Isabel Parra, Mario M. Pizarro
Abstract:
In Extreme Value Theory, inference estimation for the parameters of the distribution is made employing a small part of the observation values. When block maxima values are taken, many data are discarded. We developed a new Bayesian inference model to seize all the information provided by the data, introducing informative priors and using the relations between baseline and limit parameters. Firstly, we studied the accuracy of the new model for three baseline distributions that lead to a Gumbel extreme distribution: Exponential, Normal and Gumbel. Secondly, we considered mixtures of Normal variables, to simulate practical situations when data do not adjust to pure distributions, because of perturbations (noise).Keywords: bayesian inference, extreme value theory, Gumbel distribution, highly informative prior
Procedia PDF Downloads 20124439 The Future of Fuel Cell Electric Vehicles: Overcoming Barriers to Widespread
Authors: Silvio Carlos Anibal de Almeida
Abstract:
Hydrogen stands out as a promising alternative to fossil fuels due to its significantly high energy density. The promise of a sustainable transportation system with fuel cell electric vehicles (FCEVs) depends on overcoming economic and infrastructure barriers. The high cost of hydrogen and the scarcity of refueling stations require innovative solutions for the widespread adoption of FCEVs. Although developments in fuel cell technology have reduced costs in recent years, FCEVs are still considerably more expensive than internal combustion vehicles. This study analyzes the prospects for cost reduction of FCEVs, hydrogen, and the investments needed to expand the hydrogen distribution network. Projections indicate that the cost of FCEVs will align with that of gasoline cars by 2050, driven by technological maturation and mass production. Reducing the production costs of green hydrogen by reducing renewable energy costs, developing more efficient electrolyzers, and leveraging economies of scale could bring the price down to less than $5/kgH₂ by 2030. Government investment and public-private partnerships are essential to build a robust infrastructure for production, transportation, storage, and refueling stations. The goal of a sustainable transportation future powered by FCEVs can only be achieved by converging these factors.Keywords: alternative vehicles, fuel cell, fuel cell electric vehicles, green hydrogen
Procedia PDF Downloads 024438 Quantitative, Preservative Methodology for Review of Interview Transcripts Using Natural Language Processing
Authors: Rowan P. Martnishn
Abstract:
During the execution of a National Endowment of the Arts grant, approximately 55 interviews were collected from professionals across various fields. These interviews were used to create deliverables – historical connections for creations that began as art and evolved entirely into computing technology. With dozens of hours’ worth of transcripts to be analyzed by qualitative coders, a quantitative methodology was created to sift through the documents. The initial step was to both clean and format all the data. First, a basic spelling and grammar check was applied, as well as a Python script for normalized formatting which used an open-source grammatical formatter to make the data as coherent as possible. 10 documents were randomly selected to manually review, where words often incorrectly translated during the transcription were recorded and replaced throughout all other documents. Then, to remove all banter and side comments, the transcripts were spliced into paragraphs (separated by change in speaker) and all paragraphs with less than 300 characters were removed. Secondly, a keyword extractor, a form of natural language processing where significant words in a document are selected, was run on each paragraph for all interviews. Every proper noun was put into a data structure corresponding to that respective interview. From there, a Bidirectional and Auto-Regressive Transformer (B.A.R.T.) summary model was then applied to each paragraph that included any of the proper nouns selected from the interview. At this stage the information to review had been sent from about 60 hours’ worth of data to 20. The data was further processed through light, manual observation – any summaries which proved to fit the criteria of the proposed deliverable were selected, as well their locations within the document. This narrowed that data down to about 5 hours’ worth of processing. The qualitative researchers were then able to find 8 more connections in addition to our previous 4, exceeding our minimum quota of 3 to satisfy the grant. Major findings of the study and subsequent curation of this methodology raised a conceptual finding crucial to working with qualitative data of this magnitude. In the use of artificial intelligence there is a general trade off in a model between breadth of knowledge and specificity. If the model has too much knowledge, the user risks leaving out important data (too general). If the tool is too specific, it has not seen enough data to be useful. Thus, this methodology proposes a solution to this tradeoff. The data is never altered outside of grammatical and spelling checks. Instead, the important information is marked, creating an indicator of where the significant data is without compromising the purity of it. Secondly, the data is chunked into smaller paragraphs, giving specificity, and then cross-referenced with the keywords (allowing generalization over the whole document). This way, no data is harmed, and qualitative experts can go over the raw data instead of using highly manipulated results. Given the success in deliverable creation as well as the circumvention of this tradeoff, this methodology should stand as a model for synthesizing qualitative data while maintaining its original form.Keywords: B.A.R.T.model, keyword extractor, natural language processing, qualitative coding
Procedia PDF Downloads 3424437 Culture and Commodification: A Study of William Gibson's the Bridge Trilogy
Authors: Aruna Bhat
Abstract:
Culture can be placed within the social structure that embodies both the creation of social groups, and the manner in which they interact with each other. As many critics have pointed out, culture in the Postmodern context has often been considered a commodity, and indeed it shares many attributes with commercial products. Popular culture follows many patterns of behavior derived from Economics, from the simple principle of supply and demand, to the creation of marketable demographics which fit certain criterion. This trend is exemplary visible in contemporary fiction, especially in contemporary science fiction; Cyberpunk fiction in particular which is an off shoot of pure science fiction. William Gibson is one such author who in his works portrays such a scenario, and in his The Bridge Trilogy he adds another level of interpretation to this state of affairs, by describing a world that is centered on industrialization of a new kind – that focuses around data in the cyberspace. In this new world, data has become the most important commodity, and man has become nothing but a nodal point in a vast ocean of raw data resulting into commodification of each thing including Culture. This paper will attempt to study the presence of above mentioned elements in William Gibson’s The Bridge Trilogy. The theories applied will be Postmodernism and Cultural studies.Keywords: culture, commodity, cyberpunk, data, postmodern
Procedia PDF Downloads 51124436 SiC Merged PiN and Schottky (MPS) Power Diodes Electrothermal Modeling in SPICE
Abstract:
This paper sets out a behavioral macro-model of a Merged PiN and Schottky (MPS) diode based on silicon carbide (SiC). This model holds good for both static and dynamic electrothermal simulations for industrial applications. Its parameters have been worked out from datasheets curves by drawing on the optimization method: Simulated Annealing (SA) for the SiC MPS diodes made available in the industry. The model also adopts the Analog Behavioral Model (ABM) of PSPICE in which it has been implemented. The thermal behavior of the devices was also taken into consideration by making use of Foster’ canonical network as figured out from electro-thermal measurement provided by the manufacturer of the device.Keywords: SiC MPS diode, electro-thermal, SPICE model, behavioral macro-model
Procedia PDF Downloads 40924435 Impact of Safety and Quality Considerations of Housing Clients on the Construction Firms’ Intention to Adopt Quality Function Deployment: A Case of Construction Sector
Authors: Saif Ul Haq
Abstract:
The current study intends to examine the safety and quality considerations of clients of housing projects and their impact on the adoption of Quality Function Deployment (QFD) by the construction firm. Mixed method research technique has been used to collect and analyze the data wherein a survey was conducted to collect the data from 220 clients of housing projects in Saudi Arabia. Then, the telephonic and Skype interviews were conducted to collect data of 15 professionals working in the top ten real estate companies of Saudi Arabia. Data were analyzed by using partial least square (PLS) and thematic analysis techniques. Findings reveal that today’s customer prioritizes the safety and quality requirements of their houses and as a result, construction firms adopt QFD to address the needs of customers. The findings are of great importance for the clients of housing projects as well as for the construction firms as they could apply QFD in housing projects to address the safety and quality concerns of their clients.Keywords: construction industry, quality considerations, quality function deployment, safety considerations
Procedia PDF Downloads 12724434 Calculating Quantity of Steel Bar Placed in Mesh Form in a Circular Slab or Dome
Authors: Karam Chand Gupta
Abstract:
When steel reinforcement is placed in mesh form in circular concrete slab at base or domes at top in case of over head service reservoir or any other structure, it is difficult to estimate/measure the total quantity of steel that would be needed or placed. For the purpose of calculating the total length of the steel bars, at present, the practice is – the length of each bar is measured and then added up. This is tiresome and time consuming process. I have derived a mathematics formula with the help of which we can calculate in one line the quantity of total steel that will be needed. This will not only make it easy and time saving but also avoids any error in making entries and calculations.Keywords: dome, mesh, slab, steel
Procedia PDF Downloads 68824433 Higher-Level Return to Female Karate Competition Following Multiple Patella Dislocations
Authors: A. Maso, C. Bellissimo, G. Facchinetti, N. Milani, D. Panzin, D. Pogliana, L. Garlaschelli, L. Rivaroli, S. Rivaroli, M. Zurek, J. Konin
Abstract:
15 year-old female karate athlete experienced two unilateral patella dislocations: one contact and one non-contact. This challenged her from competing as planned at the regional and national competitions as a result of her inability to perform at a high level. Despite these injuries and other complicated factors, she was able to modify her training timeline and successfully perform, winning third at the National Cup. Initial pain numeric rating scale 8/10 during karate training isometric figures, taking the stairs, long walking, a positive rasp test, palpation pain on the lateral patella joint 9/10, pain performing open kinetic chain 0°-45° and close kinetic chain 30°-90°, tensor fascia lata, vastus lateralis, psoas muscles retraction/stiffness. Foot hyper pronation, internally rotated femur, and knee flexion 15° were the postural findings. Exercise prescription for three days/week for three weeks to include exercise-based rehabilitation and soft tissue mobilization with massage and foam rolling. After three weeks, the pain was improved during activity daily living 5/10, and soft tissue stiffness decreased. An additional four weeks of exercise-based rehabilitation was continued. At this time, axial x-rays and TA-GT TAC were taken, and an orthopaedic medical check was recommended to continue conservative treatment. At week seven, she performed 2/4 karate position technique without pain and 2/4 with pain. An isokinetic test was performed at week 12, demonstrating a 10% strength deficit and 6% resistance deficit both to the left hamstrings. Moreover, an 8% strength and resistance surplus to the left quadriceps was found. No pain was present during activity, daily living and sports activity, allowing a return to play training to begin. A plan for the return to play framework collaborated with her trainer, her father, a physiotherapist, a sports scientist, an osteopath, and a nutritionist. Within 4 and 5 months, both non-athlete and athlete movement quality analysis tests were performed. The plan agreed to establish a return to play goal of 7 months and the highest level return to competition goal of 9 months from the start of rehabilitation. This included three days/week of training and repeated testing of movement quality before return to competition with detectable improvements from 77% to 93%. Beginning goals of the rehabilitation plan included the importance of a team approach. The patient’s father and trainer were important to collaborate with to assure a safe and timely return to competition. The possibility of achieving the goals was strongly related to orthopaedic decision-making and progress during the first weeks of rehabilitation. Without complications or setbacks, the patient can successfully return to her highest level of competition. The patient returned to participation after five months of rehabilitation and training, and then she returned to competition at the national level in nine months. The successful return was the result of a team approach and a compliant patient with clear goals.Keywords: karate, knee, performance, rehabilitation
Procedia PDF Downloads 10924432 Customers’ Acceptability of Islamic Banking: Employees’ Perspective in Peshawar
Authors: Tahira Imtiaz, Karim Ullah
Abstract:
This paper aims to incorporate the banks employees’ perspective on acceptability of Islamic banking by the customers of Peshawar. A qualitative approach is adopted for which six in-depth interviews with employees of Islamic banks are conducted. The employees were asked to share their experience regarding customers’ acceptance attitude towards acceptability of Islamic banking. Collected data was analyzed through thematic analysis technique and its synthesis with the current literature. Through data analysis a theoretical framework is developed, which highlights the factors which drive customers towards Islamic banking, as witnessed by the employees. The practical implication of analyzed data evident that a new model could be developed on the basis of four determinants of human preference namely: inner satisfaction, time, faith and market forces.Keywords: customers’ attraction, employees’ perspective, Islamic banking, Riba
Procedia PDF Downloads 33624431 Customized Design of Amorphous Solids by Generative Deep Learning
Authors: Yinghui Shang, Ziqing Zhou, Rong Han, Hang Wang, Xiaodi Liu, Yong Yang
Abstract:
The design of advanced amorphous solids, such as metallic glasses, with targeted properties through artificial intelligence signifies a paradigmatic shift in physical metallurgy and materials technology. Here, we developed a machine-learning architecture that facilitates the generation of metallic glasses with targeted multifunctional properties. Our architecture integrates the state-of-the-art unsupervised generative adversarial network model with supervised models, allowing the incorporation of general prior knowledge derived from thousands of data points across a vast range of alloy compositions, into the creation of data points for a specific type of composition, which overcame the common issue of data scarcity typically encountered in the design of a given type of metallic glasses. Using our generative model, we have successfully designed copper-based metallic glasses, which display exceptionally high hardness or a remarkably low modulus. Notably, our architecture can not only explore uncharted regions in the targeted compositional space but also permits self-improvement after experimentally validated data points are added to the initial dataset for subsequent cycles of data generation, hence paving the way for the customized design of amorphous solids without human intervention.Keywords: metallic glass, artificial intelligence, mechanical property, automated generation
Procedia PDF Downloads 6624430 Making the Right Call for Falls: Evaluating the Efficacy of a Multi-Faceted Trust Wide Approach to Improving Patient Safety Post Falls
Authors: Jawaad Saleem, Hannah Wright, Peter Sommerville, Adrian Hopper
Abstract:
Introduction: Inpatient falls are the most commonly reported patient safety incidents, and carry a significant burden on resources, morbidity, and mortality. Ensuring adequate post falls management of patients by staff is therefore paramount to maintaining patient safety especially in out of hours and resource stretched settings. Aims: This quality improvement project aims to improve the current practice of falls management at Guys St Thomas Hospital, London as compared to our 2016 Quality Improvement Project findings. Furthermore, it looks to increase current junior doctors confidence in managing falls and their use of new guidance protocols. Methods: Multifaceted Interventions implemented included: the development of new trust wide guidelines detailing management pathways for patients post falls, available for intranet access. Furthermore, the production of 2000 lanyard cards distributed amongst junior doctors and staff which summarised these guidelines. Additionally, a ‘safety signal’ email was sent from the Trust chief medical officer to all staff raising awareness of falls and the guidelines. Formal falls teaching was also implemented for new doctors at induction. Using an established incident database, 189 consecutive falls in 2017were retrospectively analysed electronically to assess and compared to the variables measured in 2016 post interventions. A separate serious incident database was used to analyse 50 falls from May 2015 to March 2018 to ascertain the statistical significance of the impact of our interventions on serious incidents. A similar questionnaire for the 2017 cohort of foundation year one (FY1) doctors was performed and compared to 2016 results. Results: Questionnaire data demonstrated improved awareness and utility of guidelines and increased confidence as well as an increase in training. 97% of FY1 trainees felt that the interventions had increased their awareness of the impact of falls on patients in the trust. Data from the incident database demonstrated the time to review patients post fall had decreased from an average of 130 to 86 minutes. Improvement was also demonstrated in the reduced time to order and schedule X-ray and CT imaging, 3 and 5 hours respectively. Data from the serious incident database show that ‘the time from fall until harm was detected’ was statistically significantly lower (P = 0.044) post intervention. We also showed the incidence of significant delays in detecting harm ( > 10 hours) reduced post intervention. Conclusions: Our interventions have helped to significantly reduce the average time to assess, order and schedule appropriate imaging post falls. Delays of over ten hours to detect serious injuries after falls were commonplace; since the intervention, their frequency has markedly reduced. We suggest this will lead to identifying patient harm sooner, reduced clinical incidents relating to falls and thus improve overall patient safety. Our interventions have also helped increase clinical staff confidence, management, and awareness of falls in the trust. Next steps include expanding teaching sessions, improving multidisciplinary team involvement to aid this improvement.Keywords: patient safety, quality improvement, serious incidents, falls, clinical care
Procedia PDF Downloads 12624429 R Data Science for Technology Management
Authors: Sunghae Jun
Abstract:
Technology management (TM) is important issue in a company improving the competitiveness. Among many activities of TM, technology analysis (TA) is important factor, because most decisions for management of technology are decided by the results of TA. TA is to analyze the developed results of target technology using statistics or Delphi. TA based on Delphi is depended on the experts’ domain knowledge, in comparison, TA by statistics and machine learning algorithms use objective data such as patent or paper instead of the experts’ knowledge. Many quantitative TA methods based on statistics and machine learning have been studied, and these have been used for technology forecasting, technological innovation, and management of technology. They applied diverse computing tools and many analytical methods case by case. It is not easy to select the suitable software and statistical method for given TA work. So, in this paper, we propose a methodology for quantitative TA using statistical computing software called R and data science to construct a general framework of TA. From the result of case study, we also show how our methodology is applied to real field. This research contributes to R&D planning and technology valuation in TM areas.Keywords: technology management, R system, R data science, statistics, machine learning
Procedia PDF Downloads 46024428 Mixture statistical modeling for predecting mortality human immunodeficiency virus (HIV) and tuberculosis(TB) infection patients
Authors: Mohd Asrul Affendi Bi Abdullah, Nyi Nyi Naing
Abstract:
The purpose of this study was to identify comparable manner between negative binomial death rate (NBDR) and zero inflated negative binomial death rate (ZINBDR) with died patients with (HIV + T B+) and (HIV + T B−). HIV and TB is a serious world wide problem in the developing country. Data were analyzed with applying NBDR and ZINBDR to make comparison which a favorable model is better to used. The ZINBDR model is able to account for the disproportionately large number of zero within the data and is shown to be a consistently better fit than the NBDR model. Hence, as a results ZINBDR model is a superior fit to the data than the NBDR model and provides additional information regarding the died mechanisms HIV+TB. The ZINBDR model is shown to be a use tool for analysis death rate according age categorical.Keywords: zero inflated negative binomial death rate, HIV and TB, AIC and BIC, death rate
Procedia PDF Downloads 43724427 A Study on the Improvement of Mobile Device Call Buzz Noise Caused by Audio Frequency Ground Bounce
Authors: Jangje Park, So Young Kim
Abstract:
The market demand for audio quality in mobile devices continues to increase, and audible buzz noise generated in time division communication is a chronic problem that goes against the market demand. In the case of time division type communication, the RF Power Amplifier (RF PA) is driven at the audio frequency cycle, and it makes various influences on the audio signal. In this paper, we measured the ground bounce noise generated by the peak current flowing through the ground network in the RF PA with the audio frequency; it was confirmed that the noise is the cause of the audible buzz noise during a call. In addition, a grounding method of the microphone device that can improve the buzzing noise was proposed. Considering that the level of the audio signal generated by the microphone device is -38dBV based on 94dB Sound Pressure Level (SPL), even ground bounce noise of several hundred uV will fall within the range of audible noise if it is induced by the audio amplifier. Through the grounding method of the microphone device proposed in this paper, it was confirmed that the audible buzz noise power density at the RF PA driving frequency was improved by more than 5dB under the conditions of the Printed Circuit Board (PCB) used in the experiment. A fundamental improvement method was presented regarding the buzzing noise during a mobile phone call.Keywords: audio frequency, buzz noise, ground bounce, microphone grounding
Procedia PDF Downloads 13924426 Efficient Reuse of Exome Sequencing Data for Copy Number Variation Callings
Authors: Chen Wang, Jared Evans, Yan Asmann
Abstract:
With the quick evolvement of next-generation sequencing techniques, whole-exome or exome-panel data have become a cost-effective way for detection of small exonic mutations, but there has been a growing desire to accurately detect copy number variations (CNVs) as well. In order to address this research and clinical needs, we developed a sequencing coverage pattern-based method not only for copy number detections, data integrity checks, CNV calling, and visualization reports. The developed methodologies include complete automation to increase usability, genome content-coverage bias correction, CNV segmentation, data quality reports, and publication quality images. Automatic identification and removal of poor quality outlier samples were made automatically. Multiple experimental batches were routinely detected and further reduced for a clean subset of samples before analysis. Algorithm improvements were also made to improve somatic CNV detection as well as germline CNV detection in trio family. Additionally, a set of utilities was included to facilitate users for producing CNV plots in focused genes of interest. We demonstrate the somatic CNV enhancements by accurately detecting CNVs in whole exome-wide data from the cancer genome atlas cancer samples and a lymphoma case study with paired tumor and normal samples. We also showed our efficient reuses of existing exome sequencing data, for improved germline CNV calling in a family of the trio from the phase-III study of 1000 Genome to detect CNVs with various modes of inheritance. The performance of the developed method is evaluated by comparing CNV calling results with results from other orthogonal copy number platforms. Through our case studies, reuses of exome sequencing data for calling CNVs have several noticeable functionalities, including a better quality control for exome sequencing data, improved joint analysis with single nucleotide variant calls, and novel genomic discovery of under-utilized existing whole exome and custom exome panel data.Keywords: bioinformatics, computational genetics, copy number variations, data reuse, exome sequencing, next generation sequencing
Procedia PDF Downloads 26024425 Colloidal Gas Aphron Generated by a Cationic Surfactant as an Alternative Technique to Recovery Natural Colorants from Fermented Broth
Authors: V. C. Santos-Ebinuma, J. F. B. Pereira, M. F. S. Teixeira, A. Pessoa Jr., P. Jauregi
Abstract:
There is worldwide interest in process development for colorants production from natural sources. Microorganisms provide an alternative source of natural colorants which can be produced by cultivation technology and extracted from fermented broth. The aim of the present work was to study the recovery of red colorants from fermented broth of Penicillium purpurogenum DPUA 1275 using the technique of Colloidal Gas Aphrons (CGA); CGA are surfactant-stabilized microbubbles generated by intense stirring of a surfactant solution. CGA were generated by the cationic, hexadecyl trimethyl ammonium bromide (CTAB) surfactant. Firstly, experiments were carried out at different surfactant/fermented broth volumetric ratios (VCGA/VFB, VRATIO) varying between 3 and 18 at pH 6.9. Secondly, the experiments were carried out at VRATIO of 6 and 12 in different pH, namely, 6.9, 8.0, 9.0 and 10.0. The first results of recovery showed that an increase in the VRATIO from 3 to 6 and 12 promoted an increase as recovery as partition coefficient. However, at VRATIO of 18 the lowest partition coefficient was obtained. The best results were achieved at VRATIO of 6 and 12, namely recovery, Re, around 60% and partition coefficient, K, of 2.5 and 3.0 to 6 and 12 VRATIO, respectively. The second set of experiments showed that the pH 9.0 promoted the best results at VRATIO of 12 as follow: Re=70%, K=5.39, proteins and sugar selectivity (SePROT, 3.75 and SeSUGAR, 7.20, respectively). These results indicate that with CTAB the recovery is mainly driven by electrostatic interactions. In conclusion, the results above show that CGA employing a cationic surfactant is a promissory technique and it can be used as the first step of purification to recovery red colorants from fermented broth.Keywords: liquid-liquid extraction, colloidal gas aphrons, recovery, natural colorants
Procedia PDF Downloads 360