Search results for: project progress prediction
7751 The Effects of Teacher Efficacy, Instructional Leadership and Professional Learning Communities on Student Achievement in Literacy and Numeracy: A Look at Primary Schools within Sibu Division
Authors: Jarrod Sio Jyh Lih
Abstract:
This paper discusses the factors contributing to student achievement in literacy and numeracy in primary schools within Sibu division. The study involved 694 level 1 primary schoolteachers. Using descriptive statistics, the study observed high levels of practice for teacher efficacy, instructional leadership and professional learning communities (PLCs). The differences between gender, teaching experience and academic qualification were analyzed using the t-test and one-way analysis of variance (ANOVA). The study reported significant differences in respondent perceptions based on teaching experience vis-à-vis teacher efficacy. Here, the post hoc Tukey test revealed that efficaciousness grows with experience. A correlation test observed positive and significant correlations between all independent variables. Binary logistic regression was applied to predict the independent variables’ influence on student achievement. The findings revealed that a dimension of instructional leadership – ‘monitoring student progress’ - emerged as the best predictor of student achievement for literacy and numeracy. The result indicated the students were more than 4 times more likely to achieve the national key performance index for both literacy and numeracy when student progress was monitored. In conclusion, ‘monitoring student progress’ had a positive influence on students’ achievement for literacy and numeracy, hence making it a possible course of action for school heads. However, more comprehensive studies are needed to ascertain its consistency within the context of Malaysia.Keywords: efficacy, instructional, literacy, numeracy
Procedia PDF Downloads 2617750 Analysis of Brain Signals Using Neural Networks Optimized by Co-Evolution Algorithms
Authors: Zahra Abdolkarimi, Naser Zourikalatehsamad,
Abstract:
Up to 40 years ago, after recognition of epilepsy, it was generally believed that these attacks occurred randomly and suddenly. However, thanks to the advance of mathematics and engineering, such attacks can be predicted within a few minutes or hours. In this way, various algorithms for long-term prediction of the time and frequency of the first attack are presented. In this paper, by considering the nonlinear nature of brain signals and dynamic recorded brain signals, ANFIS model is presented to predict the brain signals, since according to physiologic structure of the onset of attacks, more complex neural structures can better model the signal during attacks. Contribution of this work is the co-evolution algorithm for optimization of ANFIS network parameters. Our objective is to predict brain signals based on time series obtained from brain signals of the people suffering from epilepsy using ANFIS. Results reveal that compared to other methods, this method has less sensitivity to uncertainties such as presence of noise and interruption in recorded signals of the brain as well as more accuracy. Long-term prediction capacity of the model illustrates the usage of planted systems for warning medication and preventing brain signals.Keywords: co-evolution algorithms, brain signals, time series, neural networks, ANFIS model, physiologic structure, time prediction, epilepsy suffering, illustrates model
Procedia PDF Downloads 2827749 Rainfall-Runoff Forecasting Utilizing Genetic Programming Technique
Authors: Ahmed Najah Ahmed Al-Mahfoodh, Ali Najah Ahmed Al-Mahfoodh, Ahmed Al-Shafie
Abstract:
In this study, genetic programming (GP) technique has been investigated in prediction of set of rainfall-runoff data. To assess the effect of input parameters on the model, the sensitivity analysis was adopted. To evaluate the performance of the proposed model, three statistical indexes were used, namely; Correlation Coefficient (CC), Mean Square Error (MSE) and Correlation of Efficiency (CE). The principle aim of this study is to develop a computationally efficient and robust approach for predict of rainfall-runoff which could reduce the cost and labour for measuring these parameters. This research concentrates on the Johor River in Johor State, Malaysia.Keywords: genetic programming, prediction, rainfall-runoff, Malaysia
Procedia PDF Downloads 4817748 A Study for Area-level Mosquito Abundance Prediction by Using Supervised Machine Learning Point-level Predictor
Authors: Theoktisti Makridou, Konstantinos Tsaprailis, George Arvanitakis, Charalampos Kontoes
Abstract:
In the literature, the data-driven approaches for mosquito abundance prediction relaying on supervised machine learning models that get trained with historical in-situ measurements. The counterpart of this approach is once the model gets trained on pointlevel (specific x,y coordinates) measurements, the predictions of the model refer again to point-level. These point-level predictions reduce the applicability of those solutions once a lot of early warning and mitigation actions applications need predictions for an area level, such as a municipality, village, etc... In this study, we apply a data-driven predictive model, which relies on public-open satellite Earth Observation and geospatial data and gets trained with historical point-level in-Situ measurements of mosquito abundance. Then we propose a methodology to extract information from a point-level predictive model to a broader area-level prediction. Our methodology relies on the randomly spatial sampling of the area of interest (similar to the Poisson hardcore process), obtaining the EO and geomorphological information for each sample, doing the point-wise prediction for each sample, and aggregating the predictions to represent the average mosquito abundance of the area. We quantify the performance of the transformation from the pointlevel to the area-level predictions, and we analyze it in order to understand which parameters have a positive or negative impact on it. The goal of this study is to propose a methodology that predicts the mosquito abundance of a given area by relying on point-level prediction and to provide qualitative insights regarding the expected performance of the area-level prediction. We applied our methodology to historical data (of Culex pipiens) of two areas of interest (Veneto region of Italy and Central Macedonia of Greece). In both cases, the results were consistent. The mean mosquito abundance of a given area can be estimated with similar accuracy to the point-level predictor, sometimes even better. The density of the samples that we use to represent one area has a positive effect on the performance in contrast to the actual number of sampling points which is not informative at all regarding the performance without the size of the area. Additionally, we saw that the distance between the sampling points and the real in-situ measurements that were used for training did not strongly affect the performance.Keywords: mosquito abundance, supervised machine learning, culex pipiens, spatial sampling, west nile virus, earth observation data
Procedia PDF Downloads 1477747 A Novel Algorithm for Parsing IFC Models
Authors: Raninder Kaur Dhillon, Mayur Jethwa, Hardeep Singh Rai
Abstract:
Information technology has made a pivotal progress across disparate disciplines, one of which is AEC (Architecture, Engineering and Construction) industry. CAD is a form of computer-aided building modulation that architects, engineers and contractors use to create and view two- and three-dimensional models. The AEC industry also uses building information modeling (BIM), a newer computerized modeling system that can create four-dimensional models; this software can greatly increase productivity in the AEC industry. BIM models generate open source IFC (Industry Foundation Classes) files which aim for interoperability for exchanging information throughout the project lifecycle among various disciplines. The methods developed in previous studies require either an IFC schema or MVD and software applications, such as an IFC model server or a Building Information Modeling (BIM) authoring tool, to extract a partial or complete IFC instance model. This paper proposes an efficient algorithm for extracting a partial and total model from an Industry Foundation Classes (IFC) instance model without an IFC schema or a complete IFC model view definition (MVD). Procedia PDF Downloads 3007746 The Development of Electronic Health Record Adoption in Indonesian Hospitals: 2008-2015
Authors: Adistya Maulidya, Mujuna Abbas, Nur Assyifa, Putri Dewi Gutiyani
Abstract:
Countries are moving forward to develop databases from electronic health records for monitoring and research. Since the issuance of Information and Electonic Transaction Constitution No. 11 of 2008 as well as Minister Regulation No. 269 of 2008, there has been a gradual progress of Indonesian hospitals adopting Electonic Health Record (EHR) in its systems. This paper is the result of a literature study about the progress that has been made in Indonesia to develop national health information infrastructure through EHR within the hospitals. The purpose of this study was to describe trends in adoption of EHR systems among hospitals in Indonesia from 2008 to 2015 as well as to assess the preparedness of Indonesian national health information infrastructure facing ASEAN Economic Community.Keywords: adoption, Indonesian hospitals, electronic health record, ASEAN economic community
Procedia PDF Downloads 2967745 The Combination of the Mel Frequency Cepstral Coefficients, Perceptual Linear Prediction, Jitter and Shimmer Coefficients for the Improvement of Automatic Recognition System for Dysarthric Speech
Authors: Brahim Fares Zaidi
Abstract:
Our work aims to improve our Automatic Recognition System for Dysarthria Speech based on the Hidden Models of Markov and the Hidden Markov Model Toolkit to help people who are sick. With pronunciation problems, we applied two techniques of speech parameterization based on Mel Frequency Cepstral Coefficients and Perceptual Linear Prediction and concatenated them with JITTER and SHIMMER coefficients in order to increase the recognition rate of a dysarthria speech. For our tests, we used the NEMOURS database that represents speakers with dysarthria and normal speakers.Keywords: ARSDS, HTK, HMM, MFCC, PLP
Procedia PDF Downloads 1087744 Genodata: The Human Genome Variation Using BigData
Authors: Surabhi Maiti, Prajakta Tamhankar, Prachi Uttam Mehta
Abstract:
Since the accomplishment of the Human Genome Project, there has been an unparalled escalation in the sequencing of genomic data. This project has been the first major vault in the field of medical research, especially in genomics. This project won accolades by using a concept called Bigdata which was earlier, extensively used to gain value for business. Bigdata makes use of data sets which are generally in the form of files of size terabytes, petabytes, or exabytes and these data sets were traditionally used and managed using excel sheets and RDBMS. The voluminous data made the process tedious and time consuming and hence a stronger framework called Hadoop was introduced in the field of genetic sciences to make data processing faster and efficient. This paper focuses on using SPARK which is gaining momentum with the advancement of BigData technologies. Cloud Storage is an effective medium for storage of large data sets which is generated from the genetic research and the resultant sets produced from SPARK analysis.Keywords: human genome project, Bigdata, genomic data, SPARK, cloud storage, Hadoop
Procedia PDF Downloads 2597743 Predicting the Diagnosis of Alzheimer’s Disease: Development and Validation of Machine Learning Models
Authors: Jay L. Fu
Abstract:
Patients with Alzheimer's disease progressively lose their memory and thinking skills and, eventually, the ability to carry out simple daily tasks. The disease is irreversible, but early detection and treatment can slow down the disease progression. In this research, publicly available MRI data and demographic data from 373 MRI imaging sessions were utilized to build models to predict dementia. Various machine learning models, including logistic regression, k-nearest neighbor, support vector machine, random forest, and neural network, were developed. Data were divided into training and testing sets, where training sets were used to build the predictive model, and testing sets were used to assess the accuracy of prediction. Key risk factors were identified, and various models were compared to come forward with the best prediction model. Among these models, the random forest model appeared to be the best model with an accuracy of 90.34%. MMSE, nWBV, and gender were the three most important contributing factors to the detection of Alzheimer’s. Among all the models used, the percent in which at least 4 of the 5 models shared the same diagnosis for a testing input was 90.42%. These machine learning models allow early detection of Alzheimer’s with good accuracy, which ultimately leads to early treatment of these patients.Keywords: Alzheimer's disease, clinical diagnosis, magnetic resonance imaging, machine learning prediction
Procedia PDF Downloads 1437742 Shedding Light on the Black Box: Explaining Deep Neural Network Prediction of Clinical Outcome
Authors: Yijun Shao, Yan Cheng, Rashmee U. Shah, Charlene R. Weir, Bruce E. Bray, Qing Zeng-Treitler
Abstract:
Deep neural network (DNN) models are being explored in the clinical domain, following the recent success in other domains such as image recognition. For clinical adoption, outcome prediction models require explanation, but due to the multiple non-linear inner transformations, DNN models are viewed by many as a black box. In this study, we developed a deep neural network model for predicting 1-year mortality of patients who underwent major cardio vascular procedures (MCVPs), using temporal image representation of past medical history as input. The dataset was obtained from the electronic medical data warehouse administered by Veteran Affairs Information and Computing Infrastructure (VINCI). We identified 21,355 veterans who had their first MCVP in 2014. Features for prediction included demographics, diagnoses, procedures, medication orders, hospitalizations, and frailty measures extracted from clinical notes. Temporal variables were created based on the patient history data in the 2-year window prior to the index MCVP. A temporal image was created based on these variables for each individual patient. To generate the explanation for the DNN model, we defined a new concept called impact score, based on the presence/value of clinical conditions’ impact on the predicted outcome. Like (log) odds ratio reported by the logistic regression (LR) model, impact scores are continuous variables intended to shed light on the black box model. For comparison, a logistic regression model was fitted on the same dataset. In our cohort, about 6.8% of patients died within one year. The prediction of the DNN model achieved an area under the curve (AUC) of 78.5% while the LR model achieved an AUC of 74.6%. A strong but not perfect correlation was found between the aggregated impact scores and the log odds ratios (Spearman’s rho = 0.74), which helped validate our explanation.Keywords: deep neural network, temporal data, prediction, frailty, logistic regression model
Procedia PDF Downloads 1537741 Prediction of Rotating Machines with Rolling Element Bearings and Its Components Deterioration
Authors: Marimuthu Gurusamy
Abstract:
In vibration analysis (with accelerometers) of rotating machines with rolling element bearing, the customers are interested to know the failure of the machine well in advance to plan the spare inventory and maintenance. But in real world most of the machines fails before the prediction of vibration analyst or Expert analysis software. Presently the prediction of failure is based on ISO 10816 vibration limits only. But this is not enough to monitor the failure of machines well in advance. Because more than 50% of the machines will fail even the vibration readings are within acceptable zone as per ISO 10816.Hence it requires further detail analysis and different techniques to predict the failure well in advance. In vibration Analysis, the velocity spectrum is used to analyse the root cause of the mechanical problems like unbalance, misalignment and looseness etc. The envelope spectrum are used to analyse the bearing frequency components, hence the failure in inner race, outer race and rolling elements are identified. But so far there is no correlation made between these two concepts. The author used both velocity spectrum and Envelope spectrum to analyse the machine behaviour and bearing condition to correlated the changes in dynamic load (by unbalance, misalignment and looseness etc.) and effect of impact on the bearing. Hence we could able to predict the expected life of the machine and bearings in the rotating equipment (with rolling element bearings). Also we used process parameters like temperature, flow and pressure to correlate with flow induced vibration and load variations, when abnormal vibration occurs due to changes in process parameters. Hence by correlation of velocity spectrum, envelope spectrum and process data with 20 years of experience in vibration analysis, the author could able to predict the rotating Equipment and its component’s deterioration and expected duration for maintenance.Keywords: vibration analysis, velocity spectrum, envelope spectrum, prediction of deterioration
Procedia PDF Downloads 4517740 3D-Vehicle Associated Research Fields for Smart City via Semantic Search Approach
Authors: Haluk Eren, Mucahit Karaduman
Abstract:
This paper presents 15-year trends for scientific studies in a scientific database considering 3D and vehicle words. Two words are selected to find their associated publications in IEEE scholar database. Both of keywords are entered individually for the years 2002, 2012, and 2016 on the database to identify the preferred subjects of researchers in same years. We have classified closer research fields after searching and listing. Three years (2002, 2012, and 2016) have been investigated to figure out progress in specified time intervals. The first one is assumed as the initial progress in between 2002-2012, and the second one is in 2012-2016 that is fast development duration. We have found very interesting and beneficial results to understand the scholars’ research field preferences for a decade. This information will be highly desirable in smart city-based research purposes consisting of 3D and vehicle-related issues.Keywords: Vehicle, three-dimensional, smart city, scholarly search, semantic
Procedia PDF Downloads 3287739 Profitability Assessment of Granite Aggregate Production and the Development of a Profit Assessment Model
Authors: Melodi Mbuyi Mata, Blessing Olamide Taiwo, Afolabi Ayodele David
Abstract:
The purpose of this research is to create empirical models for assessing the profitability of granite aggregate production in Akure, Ondo state aggregate quarries. In addition, an artificial neural network (ANN) model and multivariate predicting models for granite profitability were developed in the study. A formal survey questionnaire was used to collect data for the study. The data extracted from the case study mine for this study includes granite marketing operations, royalty, production costs, and mine production information. The following methods were used to achieve the goal of this study: descriptive statistics, MATLAB 2017, and SPSS16.0 software in analyzing and modeling the data collected from granite traders in the study areas. The ANN and Multi Variant Regression models' prediction accuracy was compared using a coefficient of determination (R²), Root mean square error (RMSE), and mean square error (MSE). Due to the high prediction error, the model evaluation indices revealed that the ANN model was suitable for predicting generated profit in a typical quarry. More quarries in Nigeria's southwest region and other geopolitical zones should be considered to improve ANN prediction accuracy.Keywords: national development, granite, profitability assessment, ANN models
Procedia PDF Downloads 1017738 A Study on Exploring and Prioritizing Critical Risks in Construction Project Assessment
Authors: A. Swetha
Abstract:
This study aims to prioritize and explore critical risks in construction project assessment, employing the Weighted Average Index method and Principal Component Analysis (PCA). Through extensive literature review and expert interviews, project assessment risk factors were identified across Budget and Cost Management Risk, Schedule and Time Management Risk, Scope and Planning Risk, Safety and Regulatory Compliance Risk, Resource Management Risk, Communication and Stakeholder Management Risk, and Environmental and Sustainability Risk domains. A questionnaire was distributed to stakeholders involved in construction activities in Hyderabad, India, with 180 completed responses analyzed using the Weighted Average Index method to prioritize risk factors. Subsequently, PCA was used to understand relationships between these factors and uncover underlying patterns. Results highlighted dependencies on critical resources, inadequate risk assessment, cash flow constraints, and safety concerns as top priorities, while factors like currency exchange rate fluctuations and delayed information dissemination ranked lower but remained significant. These insights offer valuable guidance for stakeholders to mitigate risks effectively and enhance project outcomes. By adopting systematic risk assessment and management approaches, construction projects in Hyderabad and beyond can navigate challenges more efficiently, ensuring long-term viability and resilience.Keywords: construction project assessment risk factor, risk prioritization, weighted average index, principal component analysis, project risk factors
Procedia PDF Downloads 407737 Challenges and Opportunities for Implementing Integrated Project Delivery Method in Public Sector Construction
Authors: Ahsan Ahmed, Ming Lu, Syed Zaidi, Farhan Khan
Abstract:
The Integrated Project Delivery (IPD) method has been proposed as the solution to tackle complexity and fragmentation in the real world while addressing the construction industry’s growing needs for productivity and sustainability. Although the private sector has taken the initiative in implementing IPD and taken advantage of new technology such as building information modeling (BIM) in delivering projects, IPD remains less known and rarely used in public sector construction. The focus of this paper is set on the use of IPD in projects in public sector, which is potentially complemented by the use of analytical functionalities for workface planning and construction oriented design enabled by recent research advances in BIM. Experiences and lessons learned from implementing IPD in the private sector and in BIM-based construction automation research would play a vital role in reducing barriers and eliminating issues in connection with project delivery in the public sector. The paper elaborates issues challenges, contractual relationships and the interactions throughout the planning, design and construction phases in the context of implementing IPD on construction projects in the public sector. A slab construction case is used as a ‘sandbox’ model to elaborate (1) the ideal way of communication, integration, and collaboration among all the parties involved in project delivery in planning and (2) the execution of projects by using IDP principles and optimization, simulation analyses.Keywords: integrated project delivery, IPD, building information modeling, BIM
Procedia PDF Downloads 2027736 Application of Causal Inference and Discovery in Curriculum Evaluation and Continuous Improvement
Authors: Lunliang Zhong, Bin Duan
Abstract:
The undergraduate graduation project is a vital part of the higher education curriculum, crucial for engineering accreditation. Current evaluations often summarize data without identifying underlying issues. This study applies the Peter-Clark algorithm to analyze causal relationships within the graduation project data of an Electronics and Information Engineering program, creating a causal model. Structural equation modeling confirmed the model's validity. The analysis reveals key teaching stages affecting project success, uncovering problems in the process. Introducing causal discovery and inference into project evaluation helps identify issues and propose targeted improvement measures. The effectiveness of these measures is validated by comparing the learning outcomes of two student cohorts, stratified by confounding factors, leading to improved teaching quality.Keywords: causal discovery, causal inference, continuous improvement, Peter-Clark algorithm, structural equation modeling
Procedia PDF Downloads 187735 A Strategy of Green Sukuk to Promote Sustainable Development Goals (SDGs) in Indonesia
Authors: Amrial, Yuri Oktaviani, Ziyan Muhammad Farhan
Abstract:
On the phase of shifting paradigm into sustainability, Indonesia is involved in Sustainable Development Goals (SDGs) project. That act is revealed by creating Medium and Long Term Roadmap for Sustainable Finance in Indonesia which collaborated design by Indonesia Financial Service Board (OJK) and Ministry of Environment and Forestry. One of alternative for that infrastructure financing is sharia-based financing, Green Sukuk (Sukuk specified on sustainable infrastructure project). Green Sukuk for infrastructure financing in Indonesia can be issued by the government in the form of Sukuk Project Financing. Moreover, banks in Indonesia can also participate for the issuance of Green Sukuk. So that the banks can create a financing for people who are concerned about environmental issues. By using qualitative methods and literature review, this paper aims to discuss potential, strategy and planning of Green Sukuk for financing sustainable infrastructure in the purpose of SDGs. This paper will benefit for government to give scientific discussion on the strategy of Green Sukuk in promoting sustainable goals infrastructure project in Indonesia.Keywords: green sukuk, infrastructure, SDGs, sustainable
Procedia PDF Downloads 3617734 Connecting the Dots: Bridging Academia and National Community Partnerships When Delivering Healthy Relationships Programming
Authors: Nicole Vlasman, Karamjeet Dhillon
Abstract:
Over the past four years, the Healthy Relationships Program has been delivered in community organizations and schools across Canada. More than 240 groups have been facilitated in collaboration with 33 organizations. As a result, 2157 youth have been engaged in the programming. The purpose and scope of the Healthy Relationships Program are to offer sustainable, evidence-based skills through small group implementation to prevent violence and promote positive, healthy relationships in youth. The program development has included extensive networking at regional and national levels. The Healthy Relationships Program is currently being implemented, adapted, and researched within the Resilience and Inclusion through Strengthening and Enhancing Relationships (RISE-R) project. Alongside the project’s research objectives, the RISE-R team has worked to virtually share the ongoing findings of the project through a slow ontology approach. Slow ontology is a practice integrated into project systems and structures whereby slowing the pace and volume of outputs offers creative opportunities. Creative production reveals different layers of success and complements the project, the building blocks for sustainability. As a result of integrating a slow ontology approach, the RISE-R team has developed a Geographic Information System (GIS) that documents local landscapes through a Story Map feature, and more specifically, video installations. Video installations capture the cartography of space and place within the context of singular diverse community spaces (case studies). By documenting spaces via human connections, the project captures narratives, which further enhance the voices and faces of the community within the larger project scope. This GIS project aims to create a visual and interactive flow of information that complements the project's mixed-method research approach. Conclusively, creative project development in the form of a geographic information system can provide learning and engagement opportunities at many levels (i.e., within community organizations and educational spaces or with the general public). In each of these disconnected spaces, fragmented stories are connected through a visual display of project outputs. A slow ontology practice within the context of the RISE-R project documents activities on the fringes and within internal structures; primarily through documenting project successes as further contributions to the Centre for School Mental Health framework (philosophy, recruitment techniques, allocation of resources and time, and a shared commitment to evidence-based products).Keywords: community programming, geographic information system, project development, project management, qualitative, slow ontology
Procedia PDF Downloads 1557733 The Reason of Principles of Construction Engineering and Management Being Necessary for Contracting Firms and Their Projects Managers
Authors: Mamoon Mousa Atout
Abstract:
The industries of construction are in continuous growth not only in Middle East rejoin but almost all over the world. For the last fifteen years, big expansion and increase of different types of projects has been observed. Many infrastructural projects have been developed, high rise buildings, big shopping malls, power sub-stations, roads, bridges, schools, universities and developing many of new cities with full and complete facilities. The growth and enlargement of the mentioned developed projects has been accomplished through many international and local contracting organizations. Senior management of these organizations depend on their qualified and experienced team whom are aware of the implications of project management, construction management, engineering management and resource management during tendering till final completion of the project. This research aims to find out why reasons of principles of construction engineering and management are necessary for contracting firms and their managers. Principles of construction management help contracting organizations to accomplish and deliver projects without delay. This can be maintained by establishing guidelines’ details for updating the adopted system of construction management that they have through qualified and experienced project managers. The research focuses on benefits of other essential skills of projects planning, monitoring and control. Defining roles and responsibilities of contractor project managers during tendering and execution is a part of the investigated factors that will be analyzed. Other skills like optimizing and utilizing the obtainable project resources to deliver the project within time, cost and quality will be also investigated to find out how these factors are affecting the performance of contracting firms, projects managers and projects. The conclusion of the research will help senior management team and the contractors project managers about the benefits of implications and benefits construction management system and its effect upon the performance and knowledge of contract values that they have, and the optimal profit margin of the firm it.Keywords: construction management, contracting firms, project managers, planning processes, roles and responsibilities
Procedia PDF Downloads 2997732 Prediction of Coronary Heart Disease Using Fuzzy Logic
Authors: Elda Maraj, Shkelqim Kuka
Abstract:
Coronary heart disease causes many deaths in the world. Unfortunately, this problem will continue to increase in the future. In this paper, a fuzzy logic model to predict coronary heart disease is presented. This model has been developed with seven input variables and one output variable that was implemented for 30 patients in Albania. Here fuzzy logic toolbox of MATLAB is used. Fuzzy model inputs are considered as cholesterol, blood pressure, physical activity, age, BMI, smoking, and diabetes, whereas the output is the disease classification. The fuzzy sets and membership functions are chosen in an appropriate manner. Centroid method is used for defuzzification. The database is taken from University Hospital Center "Mother Teresa" in Tirana, Albania.Keywords: coronary heart disease, fuzzy logic toolbox, membership function, prediction model
Procedia PDF Downloads 1617731 Case Study Approach Using Scenario Analysis to Analyze Unabsorbed Head Office Overheads
Authors: K. C. Iyer, T. Gupta, Y. M. Bindal
Abstract:
Head office overhead (HOOH) is an indirect cost and is recovered through individual project billings by the contractor. Delay in a project impacts the absorption of HOOH cost allocated to that particular project and thus diminishes the expected profit of the contractor. This unabsorbed HOOH cost is later claimed by contractors as damages. The subjective nature of the available formulae to compute unabsorbed HOOH is the difficulty that contractors and owners face and thus dispute it. The paper attempts to bring together the rationale of various HOOH formulae by gathering contractor’s HOOH cost data on all of its project, using case study approach and comparing variations in values of HOOH using scenario analysis. The case study approach uses project data collected from four construction projects of a contractor in India to calculate unabsorbed HOOH costs from various available formulae. Scenario analysis provides further variations in HOOH values after considering two independent situations mainly scope changes and new projects during the delay period. Interestingly, one of the findings in this study reveals that, in spite of HOOH getting absorbed by additional works available during the period of delay, a few formulae depict an increase in the value of unabsorbed HOOH, neglecting any absorption by the increase in scope. This indicates that these formulae are inappropriate for use in case of a change to the scope of work. Results of this study can help both parties in deciding on an appropriate formula more objectively, considering the events on a project causing the delay and contractor's position in respect of obtaining new projects.Keywords: absorbed and unabsorbed overheads, head office overheads, scenario analysis, scope variation
Procedia PDF Downloads 1657730 Prediction of Scour Profile Caused by Submerged Three-Dimensional Wall Jets
Authors: Abdullah Al Faruque, Ram Balachandar
Abstract:
Series of laboratory tests were carried out to study the extent of scour caused by a three-dimensional wall jets exiting from a square cross-section nozzle and into a non-cohesive sand beds. Previous observations have indicated that the effect of the tailwater depth was significant for densimetric Froude number greater than ten. However, the present results indicate that the cut off value could be lower depending on the value of grain size-to-nozzle width ratio. Numbers of equations are drawn out for a better scaling of numerous scour parameters. Also suggested the empirical prediction of scour to predict the scour centre line profile and plan view of scour profile at any particular time.Keywords: densimetric froude number, jets, nozzle, sand, scour, tailwater, time
Procedia PDF Downloads 4357729 Analysis of Delays during Initial Phase of Construction Projects and Mitigation Measures
Authors: Sunaitan Al Mutairi
Abstract:
A perfect start is a key factor for project completion on time. The study examined the effects of delayed mobilization of resources during the initial phases of the project. This paper mainly highlights the identification and categorization of all delays during the initial construction phase and their root cause analysis with corrective/control measures for the Kuwait Oil Company oil and gas projects. A relatively good percentage of the delays identified during the project execution (Contract award to end of defects liability period) attributed to mobilization/preliminary activity delays. Data analysis demonstrated significant increase in average project delay during the last five years compared to the previous period. Contractors had delays/issues during the initial phase, which resulted in slippages and progressively increased, resulting in time and cost overrun. Delays/issues not mitigated on time during the initial phase had very high impact on project completion. Data analysis of the delays for the past five years was carried out using trend chart, scatter plot, process map, box plot, relative importance index and Pareto chart. Construction of any project inside the Gathering Centers involves complex management skills related to work force, materials, plant, machineries, new technologies etc. Delay affects completion of projects and compromises quality, schedule and budget of project deliverables. Works executed as per plan during the initial phase and start-up duration of the project construction activities resulted in minor slippages/delays in project completion. In addition, there was a good working environment between client and contractor resulting in better project execution and management. Mainly, the contractor was on the front foot in the execution of projects, which had minimum/no delays during the initial and construction period. Hence, having a perfect start during the initial construction phase shall have a positive influence on the project success. Our research paper studies each type of delay with some real example supported by statistic results and suggests mitigation measures. Detailed analysis carried out with all stakeholders based on impact and occurrence of delays to have a practical and effective outcome to mitigate the delays. The key to improvement is to have proper control measures and periodic evaluation/audit to ensure implementation of the mitigation measures. The focus of this research is to reduce the delays encountered during the initial construction phase of the project life cycle.Keywords: construction activities delays, delay analysis for construction projects, mobilization delays, oil & gas projects delays
Procedia PDF Downloads 3187728 Ergonomics Management and Sustainability: An Exploratory Study Applied to Automaker Industry in South of Brazil
Authors: Giles Balbinotti, Lucas Balbinotti, Paula Hembecker
Abstract:
The management of the productive process project activities, for the conception of future work and for the financial health of the companies, is an important condition in an organizational model that corroborates the management of the human aspects and their variabilities existing in the work. It is important to seek, at all levels of the organization, understanding and consequent cultural change, and so that factors associated with human aspects are considered and prioritized in the projects. In this scenario, the central question of research for this study is placed from the context of the work, in which the managers and project coordinators are inserted, as follows: How is the top management convinced, in the design stages, to take The ‘Ergonomics’ as strategy for the performance and sustainability of the business? In this perspective, this research has as general objective to analyze how the application of the management of the human aspects in a real project of productive process in the automotive industry, including the activity of the manager and coordinator of the project beyond the strategies of convincing to act in the ergonomics of design. For this, the socio-technical and ergonomic approach is adopted, given its anthropocentric premise in the sense of acting on the social system simultaneously to the technical system, besides the support of the Modapts system that measures the non-value-added times and the correlation with the Critical positions. The methodological approach adopted in this study is based on a review of the literature and the analysis of the activity of the project coordinators of an industry, including the management of human aspects in the context of work variability and the strategies applied in project activities. It was observed in the study that the loss of performance of the serial production lines reaches the important number of the order of 30%, which can make the operation with not value-added, and this loss has as one of the causes, the ergonomic problems present in the professional activity.Keywords: human aspects in production process project, ergonomics in design, sociotechnical project management, sociotechnical, ergonomic principles, sustainability
Procedia PDF Downloads 2517727 Parallel Computing: Offloading Matrix Multiplication to GPU
Authors: Bharath R., Tharun Sai N., Bhuvan G.
Abstract:
This project focuses on developing a Parallel Computing method aimed at optimizing matrix multiplication through GPU acceleration. Addressing algorithmic challenges, GPU programming intricacies, and integration issues, the project aims to enhance efficiency and scalability. The methodology involves algorithm design, GPU programming, and optimization techniques. Future plans include advanced optimizations, extended functionality, and integration with high-level frameworks. User engagement is emphasized through user-friendly interfaces, open- source collaboration, and continuous refinement based on feedback. The project's impact extends to significantly improving matrix multiplication performance in scientific computing and machine learning applications.Keywords: matrix multiplication, parallel processing, cuda, performance boost, neural networks
Procedia PDF Downloads 587726 The Application of Data Mining Technology in Building Energy Consumption Data Analysis
Authors: Liang Zhao, Jili Zhang, Chongquan Zhong
Abstract:
Energy consumption data, in particular those involving public buildings, are impacted by many factors: the building structure, climate/environmental parameters, construction, system operating condition, and user behavior patterns. Traditional methods for data analysis are insufficient. This paper delves into the data mining technology to determine its application in the analysis of building energy consumption data including energy consumption prediction, fault diagnosis, and optimal operation. Recent literature are reviewed and summarized, the problems faced by data mining technology in the area of energy consumption data analysis are enumerated, and research points for future studies are given.Keywords: data mining, data analysis, prediction, optimization, building operational performance
Procedia PDF Downloads 8527725 An Implementation of Fuzzy Logic Technique for Prediction of the Power Transformer Faults
Authors: Omar M. Elmabrouk., Roaa Y. Taha., Najat M. Ebrahim, Sabbreen A. Mohammed
Abstract:
Power transformers are the most crucial part of power electrical system, distribution and transmission grid. This part is maintained using predictive or condition-based maintenance approach. The diagnosis of power transformer condition is performed based on Dissolved Gas Analysis (DGA). There are five main methods utilized for analyzing these gases. These methods are International Electrotechnical Commission (IEC) gas ratio, Key Gas, Roger gas ratio, Doernenburg, and Duval Triangle. Moreover, due to the importance of the transformers, there is a need for an accurate technique to diagnose and hence predict the transformer condition. The main objective of this technique is to avoid the transformer faults and hence to maintain the power electrical system, distribution and transmission grid. In this paper, the DGA was utilized based on the data collected from the transformer records available in the General Electricity Company of Libya (GECOL) which is located in Benghazi-Libya. The Fuzzy Logic (FL) technique was implemented as a diagnostic approach based on IEC gas ratio method. The FL technique gave better results and approved to be used as an accurate prediction technique for power transformer faults. Also, this technique is approved to be a quite interesting for the readers and the concern researchers in the area of FL mathematics and power transformer.Keywords: dissolved gas-in-oil analysis, fuzzy logic, power transformer, prediction
Procedia PDF Downloads 1437724 Prediction of Marine Ecosystem Changes Based on the Integrated Analysis of Multivariate Data Sets
Authors: Prozorkevitch D., Mishurov A., Sokolov K., Karsakov L., Pestrikova L.
Abstract:
The current body of knowledge about the marine environment and the dynamics of marine ecosystems includes a huge amount of heterogeneous data collected over decades. It generally includes a wide range of hydrological, biological and fishery data. Marine researchers collect these data and analyze how and why the ecosystem changes from past to present. Based on these historical records and linkages between the processes it is possible to predict future changes. Multivariate analysis of trends and their interconnection in the marine ecosystem may be used as an instrument for predicting further ecosystem evolution. A wide range of information about the components of the marine ecosystem for more than 50 years needs to be used to investigate how these arrays can help to predict the future.Keywords: barents sea ecosystem, abiotic, biotic, data sets, trends, prediction
Procedia PDF Downloads 1167723 Red-Tide Detection and Prediction Using MODIS Data in the Arabian Gulf of Qatar
Authors: Yasir E. Mohieldeen
Abstract:
Qatar is one of the most water scarce countries in the World. In 2014, the average per capita rainfall was less than 29 m3/y/ca, while the global average is 6,000 m3/y/ca. However, the per capita water consumption in Qatar is among the highest in the World: more than 500 liters per person per day, whereas the global average is 160 liters per person per day. Since the early 2000s, Qatar has been relying heavily on desalinated water from the Arabian Gulf as the main source of fresh water. In 2009, about 99.9% of the total potable water produced was desalinated. Reliance on desalinated water makes Qatar very vulnerable to water related natural disasters, such as the red-tide phenomenon. Qatar’s strategic water reserve lasts for only 7 days. In case of red-tide outbreak, the country would not be able to desalinate water for days, let alone the months that this disaster would bring about (as it clogs the desalination equipment). The 2008-09 red-tide outbreak, for instance, lasted for more than eight months and forced the closure of desalination plants in the region for weeks. This study aims at identifying favorite conditions for red-tide outbreaks, using satellite data along with in-situ measurements. This identification would allow the prediction of these outbreaks and their hotspots. Prediction and monitoring of outbreaks are crucial to water security in the country, as different measures could be put in place in advance to prevent an outbreak and mitigate its impact if it happened. Red-tide outbreaks are detected using different algorithms for chlorophyll concentration in the Gulf waters. Vegetation indices, such as Normalized Difference Vegetation Index (NDVI) and Enhanced Vegetation Index (EVI) were used along with Surface Algae Bloom Index (SABI) to detect known outbreaks. MODIS (or Moderate Resolution Imaging Spectroradiometer) bands are used to calculate these indices. A red-tide outbreaks atlas in the Arabian Gulf is being produced. Prediction of red-tide outbreaks ahead of their occurrences would give critical information on possible water-shortage in the country. Detecting known outbreaks in the past few decades and related parameters (e.g. water salinity, water surface temperature, nutrition, sandstorms, … etc) enables the identification of favorite conditions of red-tide outbreak that are key to the prediction of these outbreaks.Keywords: Arabian Gulf, MODIS, red-tide detection, strategic water reserve, water desalination
Procedia PDF Downloads 1077722 Police Violence, Activism, and the Changing Rural United States: A Digital History and Mapping Narrative
Authors: Joel Zapata
Abstract:
Chicana/o Activism in the Southern Plains Through Time and Space, a digital history project available at PlainsMovement.com, helps reveal an understudied portion of the Chicana/o Civil Rights Movement: the way it unfolded on the Southern Plains. The project centers around an approachable interactive map and timeline along with a curated collection of materials. Therefore, the project provides a digital museum experience that has not emerged within the region’s museums. That is, this digital history project takes scholarly research to the wider public, making it is also a publicly facing history project. In this way, the project adds to both scholarly and socially significant conversations, showing that the region was home to a burgeoning wing of the Chicana/o Movement and that instances of police brutality largely spurred this wing of the social justice movement. Moreover, the curated collection of materials demonstrates that police brutality united the plains’ Mexican population across political ideology, a largely overlooked aspect within the study of Mexican American civil rights movements. Such a finding can be of use today since contemporary Latina/o social justice organizations generally ignore policing issues even amid a rise in national awareness regarding police abuse. In making history accessible to Mexican origin and Latina/o communities, these same communities may in-turn use the knowledge gained from historical research towards the betterment of their social positions—the foundational goal of Chicana/o history and the related field of Chicana/o Studies. Ultimately, this digital history project is intended to draw visitors to further explore the Chicana/o Civil Rights Movement within and beyond the plains.Keywords: Chicana/o Movement, digital history, police brutality, newspapers, protests, student activism
Procedia PDF Downloads 122