Search results for: time workflow network
18768 Comparisons of Individual and Group Replacement Policies for a Series Connection System with Two Machines
Authors: Wen Liang Chang, Mei Wei Wang, Ruey Huei Yeh
Abstract:
This paper studies the comparisons of individual and group replacement policies for a series connection system with two machines. Suppose that manufacturer’s production system is a series connection system which is combined by two machines. For two machines, when machines fail within the operating time, minimal repair is performed for machines by the manufacturer. The manufacturer plans to a preventive replacement for machines at a pre-specified time to maintain system normal operation. Under these maintenance policies, the maintenance cost rate models of individual and group replacement for a series connection system with two machines is derived and further, optimal preventive replacement time is obtained such that the expected total maintenance cost rate is minimized. Finally, some numerical examples are given to illustrate the influences of individual and group replacement policies to the maintenance cost rate.Keywords: individual replacement, group replacement, replacement time, two machines, series connection system
Procedia PDF Downloads 48818767 Study on Sharp V-Notch Problem under Dynamic Loading Condition Using Symplectic Analytical Singular Element
Authors: Xiaofei Hu, Zhiyu Cai, Weian Yao
Abstract:
V-notch problem under dynamic loading condition is considered in this paper. In the time domain, the precise time domain expanding algorithm is employed, in which a self-adaptive technique is carried out to improve computing accuracy. By expanding variables in each time interval, the recursive finite element formulas are derived. In the space domain, a Symplectic Analytical Singular Element (SASE) for V-notch problem is constructed addressing the stress singularity of the notch tip. Combining with the conventional finite elements, the proposed SASE can be used to solve the dynamic stress intensity factors (DSIFs) in a simple way. Numerical results show that the proposed SASE for V-notch problem subjected to dynamic loading condition is effective and efficient.Keywords: V-notch, dynamic stress intensity factor, finite element method, precise time domain expanding algorithm
Procedia PDF Downloads 17218766 Grey Wolf Optimization Technique for Predictive Analysis of Products in E-Commerce: An Adaptive Approach
Authors: Shital Suresh Borse, Vijayalaxmi Kadroli
Abstract:
E-commerce industries nowadays implement the latest AI, ML Techniques to improve their own performance and prediction accuracy. This helps to gain a huge profit from the online market. Ant Colony Optimization, Genetic algorithm, Particle Swarm Optimization, Neural Network & GWO help many e-commerce industries for up-gradation of their predictive performance. These algorithms are providing optimum results in various applications, such as stock price prediction, prediction of drug-target interaction & user ratings of similar products in e-commerce sites, etc. In this study, customer reviews will play an important role in prediction analysis. People showing much interest in buying a lot of services& products suggested by other customers. This ultimately increases net profit. In this work, a convolution neural network (CNN) is proposed which further is useful to optimize the prediction accuracy of an e-commerce website. This method shows that CNN is used to optimize hyperparameters of GWO algorithm using an appropriate coding scheme. Accurate model results are verified by comparing them to PSO results whose hyperparameters have been optimized by CNN in Amazon's customer review dataset. Here, experimental outcome proves that this proposed system using the GWO algorithm achieves superior execution in terms of accuracy, precision, recovery, etc. in prediction analysis compared to the existing systems.Keywords: prediction analysis, e-commerce, machine learning, grey wolf optimization, particle swarm optimization, CNN
Procedia PDF Downloads 11318765 In vitro Effects of Viscum album on the Functionality of Rabbit Spermatozoa
Authors: Marek Halenár, Eva Tvrdá, Simona Baldovská, Ľubomír Ondruška, Peter Massányi, Adriana Kolesárová
Abstract:
This study aimed to assess the in vitro effects of different concentrations of the Viscum album extract on the motility, viability, and reactive oxygen species (ROS) production by rabbit spermatozoa during different time periods (0, 2, and 8h). Spermatozoa motility was assessed by using the CASA (Computer aided sperm analysis) system. Cell viability was evaluated by using the metabolic activity MTT assay, and the luminol-based luminometry was applied to quantify the ROS formation. The CASA analysis revealed that low Viscum concentrations were able to prevent a rapid decline of spermatozoa motility, especially in the case of concentrations ranging between 1 and 5 µg/mL (P<0.05 with respect to time 8h). At the same time, concentrations ranging between 1 and 100 µg/mL of the extract led to a significant preservation of the cell viability (P<0.05 in case of 5, 50 and 100 µg/mL; P<0.01 with respect to 1 and 10 µg/mL, time 8h). 1 and 5 µg/mL of the extract exhibited antioxidant characteristics, translated into a significant reduction of the ROS production, particularly notable at time 8h (P<0.01). The results indicate that the Viscum extract is capable of delaying the damage inflicted to the spermatozoon by the in vitro environment.Keywords: CASA, mistletoe, mitochondrial activity, motility, reactive oxygen species, rabbits, spermatozoa, Viscum album
Procedia PDF Downloads 39418764 A Method To Assess Collaboration Using Perception of Risk from the Architectural Engineering Construction Industry
Authors: Sujesh F. Sujan, Steve W. Jones, Arto Kiviniemi
Abstract:
The use of Building Information Modelling (BIM) in the Architectural-Engineering-Construction (AEC) industry is a form of systemic innovation. Unlike incremental innovation, (such as the technological development of CAD from hand based drawings to 2D electronically printed drawings) any form of systemic innovation in Project-Based Inter-Organisational Networks requires complete collaboration and results in numerous benefits if adopted and utilised properly. Proper use of BIM involves people collaborating with the use of interoperable BIM compliant tools. The AEC industry globally has been known for its adversarial and fragmented nature where firms take advantage of one another to increase their own profitability. Due to the industry’s nature, getting people to collaborate by unifying their goals is critical to successful BIM adoption. However, this form of innovation is often being forced artificially in the old ways of working which do not suit collaboration. This may be one of the reasons for its low global use even though the technology was developed more than 20 years ago. Therefore, there is a need to develop a metric/method to support and allow industry players to gain confidence in their investment into BIM software and workflow methods. This paper departs from defining systemic risk as a risk that affects all the project participants at a given stage of a project and defines categories of systemic risks. The need to generalise is to allow method applicability to any industry where the category will be the same, but the example of the risk will depend on the industry the study is done in. The method proposed seeks to use individual perception of an example of systemic risk as a key parameter. The significance of this study lies in relating the variance of individual perception of systemic risk to how much the team is collaborating. The method bases its notions on the claim that a more unified range of individual perceptions would mean a higher probability that the team is collaborating better. Since contracts and procurement devise how a project team operates, the method could also break the methodological barrier of highly subjective findings that case studies inflict, which has limited the possibility of generalising between global industries. Since human nature applies in all industries, the authors’ intuition is that perception can be a valuable parameter to study collaboration which is essential especially in projects that utilise systemic innovation such as BIM.Keywords: building information modelling, perception of risk, systemic innovation, team collaboration
Procedia PDF Downloads 18518763 Optimisation of Energy Harvesting for a Composite Aircraft Wing Structure Bonded with Discrete Macro Fibre Composite Sensors
Authors: Ali H. Daraji, Ye Jianqiao
Abstract:
The micro electrical devices of the wireless sensor network are continuously developed and become very small and compact with low electric power requirements using limited period life conventional batteries. The low power requirement for these devices, cost of conventional batteries and its replacement have encouraged researcher to find alternative power supply represented by energy harvesting system to provide an electric power supply with infinite period life. In the last few years, the investigation of energy harvesting for structure health monitoring has increased to powering wireless sensor network by converting waste mechanical vibration into electricity using piezoelectric sensors. Optimisation of energy harvesting is an important research topic to ensure a flowing of efficient electric power from structural vibration. The harvesting power is mainly based on the properties of piezoelectric material, dimensions of piezoelectric sensor, its position on a structure and value of an external electric load connected between sensor electrodes. Larger surface area of sensor is not granted larger power harvesting when the sensor area is covered positive and negative mechanical strain at the same time. Thus lead to reduction or cancellation of piezoelectric output power. Optimisation of energy harvesting is achieved by locating these sensors precisely and efficiently on the structure. Limited published work has investigated the energy harvesting for aircraft wing. However, most of the published studies have simplified the aircraft wing structure by a cantilever flat plate or beam. In these studies, the optimisation of energy harvesting was investigated by determination optimal value of an external electric load connected between sensor electrode terminals or by an external electric circuit or by randomly splitting piezoelectric sensor to two segments. However, the aircraft wing structures are complex than beam or flat plate and mostly constructed from flat and curved skins stiffened by stringers and ribs with more complex mechanical strain induced on the wing surfaces. This aircraft wing structure bonded with discrete macro fibre composite sensors was modelled using multiphysics finite element to optimise the energy harvesting by determination of the optimal number of sensors, location and the output resistance load. The optimal number and location of macro fibre sensors were determined based on the maximization of the open and close loop sensor output voltage using frequency response analysis. It was found different optimal distribution, locations and number of sensors bounded on the top and the bottom surfaces of the aircraft wing.Keywords: energy harvesting, optimisation, sensor, wing
Procedia PDF Downloads 30218762 Classification of Foliar Nitrogen in Common Bean (Phaseolus Vulgaris L.) Using Deep Learning Models and Images
Authors: Marcos Silva Tavares, Jamile Raquel Regazzo, Edson José de Souza Sardinha, Murilo Mesquita Baesso
Abstract:
Common beans are a widely cultivated and consumed legume globally, serving as a staple food for humans, especially in developing countries, due to their nutritional characteristics. Nitrogen (N) is the most limiting nutrient for productivity, and foliar analysis is crucial to ensure balanced nitrogen fertilization. Excessive N applications can cause, either isolated or cumulatively, soil and water contamination, plant toxicity, and increase their susceptibility to diseases and pests. However, the quantification of N using conventional methods is time-consuming and costly, demanding new technologies to optimize the adequate supply of N to plants. Thus, it becomes necessary to establish constant monitoring of the foliar content of this macronutrient in plants, mainly at the V4 stage, aiming at precision management of nitrogen fertilization. In this work, the objective was to evaluate the performance of a deep learning model, Resnet-50, in the classification of foliar nitrogen in common beans using RGB images. The BRS Estilo cultivar was sown in a greenhouse in a completely randomized design with four nitrogen doses (T1 = 0 kg N ha-1, T2 = 25 kg N ha-1, T3 = 75 kg N ha-1, and T4 = 100 kg N ha-1) and 12 replications. Pots with 5L capacity were used with a substrate composed of 43% soil (Neossolo Quartzarênico), 28.5% crushed sugarcane bagasse, and 28.5% cured bovine manure. The water supply of the plants was done with 5mm of water per day. The application of urea (45% N) and the acquisition of images occurred 14 and 32 days after sowing, respectively. A code developed in Matlab© R2022b was used to cut the original images into smaller blocks, originating an image bank composed of 4 folders representing the four classes and labeled as T1, T2, T3, and T4, each containing 500 images of 224x224 pixels obtained from plants cultivated under different N doses. The Matlab© R2022b software was used for the implementation and performance analysis of the model. The evaluation of the efficiency was done by a set of metrics, including accuracy (AC), F1-score (F1), specificity (SP), area under the curve (AUC), and precision (P). The ResNet-50 showed high performance in the classification of foliar N levels in common beans, with AC values of 85.6%. The F1 for classes T1, T2, T3, and T4 was 76, 72, 74, and 77%, respectively. This study revealed that the use of RGB images combined with deep learning can be a promising alternative to slow laboratory analyses, capable of optimizing the estimation of foliar N. This can allow rapid intervention by the producer to achieve higher productivity and less fertilizer waste. Future approaches are encouraged to develop mobile devices capable of handling images using deep learning for the classification of the nutritional status of plants in situ.Keywords: convolutional neural network, residual network 50, nutritional status, artificial intelligence
Procedia PDF Downloads 1918761 Radiofrequency Ablation: A Technique in the Management of Low Anal Fistula
Authors: R. Suresh, C. B. Singh, A. K. Sarda
Abstract:
Background: Over the decades, several surgical techniques have been developed to treat anal fistulas with variable success rates and complications. Large amount of work has been done in radiofrequency excision of the fistula for several years but no work has been done for ablating the tract. Therefore one can consider for obliteration ofanal fistula by Radiofrequency ablation (RFA). Material and Methods: A randomized controlled clinical trial was conducted at Lok Nayak Hospital, where a total of 40 patients were enrolled in the study and they were randomly assigned to Group I (fistulectomy)(n=20) and Group II (RFA) (n=20). Aim of the study was to compare the efficacy of RFA of fistula versus fistulectomy in the treatment of a low anal fistula and to evaluate RFA as an effective alternative to fistulectomy with respect to time taken for wound healing as primary outcome and post-operative pain, time taken to return to work as secondary outcomes. Patients with simple low anal fistulas, single internal and external opening, not more than two secondary tracts were included. Patients with high complex fistula, fistulas communicating with cavity, fistula due to condition like tuberculosis, Crohn's, malignancy were excluded from the study. Results: Both groups were comparable with respect to age, sex ratio, type of fistula. Themean healing time was significantly shorter in group II (41.02 days) than in group I(62.68 days).The mean operative time was significantly shorter in groupII (21.40 min) than in group I(28.50 min). The mean time taken to return to work was significantly shorter in group II(8.30 days)than in group I(12.01 days).There was no significant difference in the post operative hospital stay, mean postoperative pain score, wound infection and recurrence between the two groups. Conclusion: The patients who underwent RFA of fistula had shorter wound healing time, operative time and time taken to return to work when compared to those who underwent fistulectomy and therefore RFA shows outcome comparable to fistulectomy in the treatment of low anal fistula.Keywords: fistulectomy, low anal fistula, radio frequency ablation, wound healing
Procedia PDF Downloads 34418760 Upgrades for Hydric Supply in Water System Distribution: Use of the Bayesian Network and Technical Expedients
Authors: Elena Carcano, James Ball
Abstract:
This work details the strategies adopted by the Italian Water Utilities during the distribution of water in emergency conditions which glide from earthquakes and droughts to floods and fires. Several water bureaus located over the national territory have been interviewed, and the collected information has been used in a database of potential interventions to be taken. The work discusses the actions adopted by water utilities. These are generally prioritized in order to minimize the social, temporal, and economic burden that the damaged and nearby areas need to support. Actions are defined relying on the Bayesian Network Approach, which constitutes the hard core of any decision support system. The Bayesian Networks give answers to interventions to real and most likely risky cases. The added value of this research consists in supplying the National Bureau, namely Protezione Civile, in charge of managing havoc and catastrophic situations with a univocal plot outline so as to be able to handle actions uniformly at the expense of different local laws or contradictory customs which squander any recovery conditions, proper technical service, and economic aids. The paper is organized as follows: in section 1, the introduction is stated; section 2 provides a brief discussion of BNNs (Bayesian Networks), section 3 introduces the adopted methodology; and in the last sections, results are presented, and conclusions are drawn.Keywords: hierarchical process, strategic plan, water emergency conditions, water supply
Procedia PDF Downloads 16018759 Experimental Study on the Preparation of Pelletizing of the Panzhihua's Fine Ilmenite Concentrate
Authors: Han Kexi, Lv Xuewei, Song Bing
Abstract:
This paper focuses on the preparation of pelletizing with the Panzhihua ilmenite concentrate to satisfy the requirement of smelting titania slag. The effects of the moisture content, mixing time of raw materials, pressure of pellet, roller rotating speed of roller, drying temperature and time on the pelletizing yield and compressive strength were investigated. The experimental results show that the moister content was controlled at 2.0%~2.5%, mixing time at 20 min, the pressure of the ball forming machine at 13~15 mpa, the pelletizing yield can reach up 85%. When the roller rotating speed is 6~8 r/min while the drying temperature and time respectively is 350 ℃ and 40~60 min, the compressive strength of pelletizing more than 1500 N. The preparation of pelletizing can meet the requirement of smelting titania slag.Keywords: Panzhihua fine ilmenite concentrate, pelletizing, pelletizing yield, compressive strength, drying
Procedia PDF Downloads 21618758 Improved Technology Portfolio Management via Sustainability Analysis
Authors: Ali Al-Shehri, Abdulaziz Al-Qasim, Abdulkarim Sofi, Ali Yousef
Abstract:
The oil and gas industry has played a major role in improving the prosperity of mankind and driving the world economy. According to the International Energy Agency (IEA) and Integrated Environmental Assessment (EIA) estimates, the world will continue to rely heavily on hydrocarbons for decades to come. This growing energy demand mandates taking sustainability measures to prolong the availability of reliable and affordable energy sources, and ensure lowering its environmental impact. Unlike any other industry, the oil and gas upstream operations are energy-intensive and scattered over large zonal areas. These challenging conditions require unique sustainability solutions. In recent years there has been a concerted effort by the oil and gas industry to develop and deploy innovative technologies to: maximize efficiency, reduce carbon footprint, reduce CO2 emissions, and optimize resources and material consumption. In the past, the main driver for research and development (R&D) in the exploration and production sector was primarily driven by maximizing profit through higher hydrocarbon recovery and new discoveries. Environmental-friendly and sustainable technologies are increasingly being deployed to balance sustainability and profitability. Analyzing technology and its sustainability impact is increasingly being used in corporate decision-making for improved portfolio management and allocating valuable resources toward technology R&D.This paper articulates and discusses a novel workflow to identify strategic sustainable technologies for improved portfolio management by addressing existing and future upstream challenges. It uses a systematic approach that relies on sustainability key performance indicators (KPI’s) including energy efficiency quotient, carbon footprint, and CO2 emissions. The paper provides examples of various technologies including CCS, reducing water cuts, automation, using renewables, energy efficiency, etc. The use of 4IR technologies such as Artificial Intelligence, Machine Learning, and Data Analytics are also discussed. Overlapping technologies, areas of collaboration and synergistic relationships are identified. The unique sustainability analyses provide improved decision-making on technology portfolio management.Keywords: sustainability, oil& gas, technology portfolio, key performance indicator
Procedia PDF Downloads 18318757 Pedestrian Areas, Development Stimulus in Urban Old Fabrics; Analyzing Stroget, Pedestrian Street in Copenhagen
Authors: Kiomars Habibi, Mostafa Behzadfar, Airin Jaberi
Abstract:
Designing appropriate places for the comfort of pedestrians is one of the most important aspects of modern urbanization and renovation and rehabilitation stimulus of urban old fabrics. So, that special cities designed for pedestrians with a complete network of streets without cars, can be considered as one of the best habitations in the world. The number of these cities with a network of streets and squares in which beauty, enjoyment and comfort are mostly concerned for the pedestrians designed regions is increasing around the world, such as Stockholm, Copenhagen, Munich, Frankfurt, Venice, Rome, etc. In this paper, we are going to explain the influential factors regarding the efficiency of these cities by identifying one of the most important pedestrian ways of the world; Strøget is a car free zone in Copenhagen, Denmark. This popular tourist attraction in the center of town is the longest pedestrian shopping area in Europe. Analyses indicate that world-wide experience concerning the renovation and rehabilitation of old fabrics has many advantages in exploiting the idea of pedestrian way for regeneration of old fabrics. Transforming the streets to appropriate places for the comfort of pedestrians, expanding the public spaces such as city squares, and decreasing the masses of building alongside the brought comfort and peace is the main reason in the success of Strøget pedestrian street in urban old fabrics of Copenhagen. Hypothesis: The Strøget pedestrian street has been the development stimulus in Copenhagen and the urban old fabrics development as a resultKeywords: development, stimulus, pedestrian street, urban landscape, Stroget
Procedia PDF Downloads 10718756 Investigations of the Crude Oil Distillation Preheat Section in Unit 100 of Abadan Refinery and Its Recommendation
Authors: Mahdi GoharRokhi, Mohammad H. Ruhipour, Mohammad R. ZamaniZadeh, Mohsen Maleki, Yusef Shamsayi, Mahdi FarhaniNejad, Farzad FarrokhZadeh
Abstract:
Possessing massive resources of natural gas and petroleum, Iran has a special place among all other oil producing countries, according to international institutions of energy. In order to use these resources, development and functioning optimization of refineries and industrial units is mandatory. Heat exchanger is one of the most important and strategic equipment which its key role in the process of production is clear to everyone. For instance, if the temperature of a processing fluid is not set as needed by heat exchangers, the specifications of desired product can change profoundly. Crude oil enters a network of heat exchangers in atmospheric distillation section before getting into the distillation tower; in this case, well-functioning of heat exchangers can significantly affect the operation of distillation tower. In this paper, different scenarios for pre-heating of oil are studied using oil and gas simulation software, and the results are discussed. As we reviewed various scenarios, adding a heat exchanger to pre-heating network is proposed as the most efficient factor in improving all governing parameters of the tower i.e. temperature, pressure, and reflux rate. This exchanger is embedded in crude oil’s path. Crude oil enters the exchanger after E-101 and exchanges heat with discharging kerosene pump around from E-136. As depicted in the results, it will efficiently assist the improvement of process operation and side expenses.Keywords: atmospheric distillation unit, heat exchanger, preheat, simulation
Procedia PDF Downloads 66018755 Reducing Hazardous Materials Releases from Railroad Freights through Dynamic Trip Plan Policy
Authors: Omar A. Abuobidalla, Mingyuan Chen, Satyaveer S. Chauhan
Abstract:
Railroad transportation of hazardous materials freights is important to the North America economics that supports the national’s supply chain. This paper introduces various extensions of the dynamic hazardous materials trip plan problems. The problem captures most of the operational features of a real-world railroad transportations systems that dynamically initiates a set of blocks and assigns each shipment to a single block path or multiple block paths. The dynamic hazardous materials trip plan policies have distinguishing features that are integrating the blocking plan, and the block activation decisions. We also present a non-linear mixed integer programming formulation for each variant and present managerial insights based on a hypothetical railroad network. The computation results reveal that the dynamic car scheduling policies are not only able to take advantage of the capacity of the network but also capable of diminishing the population, and environment risks by rerouting the active blocks along the least risky train services without sacrificing the cost advantage of the railroad. The empirical results of this research illustrate that the issue of integrating the blocking plan, and the train makeup of the hazardous materials freights must receive closer attentions.Keywords: dynamic car scheduling, planning and scheduling hazardous materials freights, airborne hazardous materials, gaussian plume model, integrated blocking and routing plans, box model
Procedia PDF Downloads 20518754 An Evaluation of the Artificial Neural Network and Adaptive Neuro Fuzzy Inference System Predictive Models for the Remediation of Crude Oil-Contaminated Soil Using Vermicompost
Authors: Precious Ehiomogue, Ifechukwude Israel Ahuchaogu, Isiguzo Edwin Ahaneku
Abstract:
Vermicompost is the product of the decomposition process using various species of worms, to create a mixture of decomposing vegetable or food waste, bedding materials, and vemicast. This process is called vermicomposting, while the rearing of worms for this purpose is called vermiculture. Several works have verified the adsorption of toxic metals using vermicompost but the application is still scarce for the retention of organic compounds. This research brings to knowledge the effectiveness of earthworm waste (vermicompost) for the remediation of crude oil contaminated soils. The remediation methods adopted in this study were two soil washing methods namely, batch and column process which represent laboratory and in-situ remediation. Characterization of the vermicompost and crude oil contaminated soil were performed before and after the soil washing using Fourier transform infrared (FTIR), scanning electron microscopy (SEM), X-ray fluorescence (XRF), X-ray diffraction (XRD) and Atomic adsorption spectrometry (AAS). The optimization of washing parameters, using response surface methodology (RSM) based on Box-Behnken Design was performed on the response from the laboratory experimental results. This study also investigated the application of machine learning models [Artificial neural network (ANN), Adaptive neuro fuzzy inference system (ANFIS). ANN and ANFIS were evaluated using the coefficient of determination (R²) and mean square error (MSE)]. Removal efficiency obtained from the Box-Behnken design experiment ranged from 29% to 98.9% for batch process remediation. Optimization of the experimental factors carried out using numerical optimization techniques by applying desirability function method of the response surface methodology (RSM) produce the highest removal efficiency of 98.9% at absorbent dosage of 34.53 grams, adsorbate concentration of 69.11 (g/ml), contact time of 25.96 (min), and pH value of 7.71, respectively. Removal efficiency obtained from the multilevel general factorial design experiment ranged from 56% to 92% for column process remediation. The coefficient of determination (R²) for ANN was (0.9974) and (0.9852) for batch and column process, respectively, showing the agreement between experimental and predicted results. For batch and column precess, respectively, the coefficient of determination (R²) for RSM was (0.9712) and (0.9614), which also demonstrates agreement between experimental and projected findings. For the batch and column processes, the ANFIS coefficient of determination was (0.7115) and (0.9978), respectively. It can be concluded that machine learning models can predict the removal of crude oil from polluted soil using vermicompost. Therefore, it is recommended to use machines learning models to predict the removal of crude oil from contaminated soil using vermicompost.Keywords: ANFIS, ANN, crude-oil, contaminated soil, remediation and vermicompost
Procedia PDF Downloads 11118753 The Effect of the Parameters of the Grinding on the Characteristics of the Deposit Phosphate Ore of Kef Es Sennoun, Djebel Onk-Tebessa, Algeria
Authors: N. Benabdeslam, N. Bouzidi, F. Atmani, R. Boucif, A. Sakhri
Abstract:
The objective of this study was to provide answers for a better understanding of the mechanisms involved during grinding. To obtain a phosphate powder, we carry out sieving - grinding circuits for each parameter influencing the process. The analysis of the average particle size of the different tests carried out served in the first place as a basis for the determination of the granulometric curve area, the characteristics and the granular coefficients, then the exploitation of the different results for the calculation of the energies consumed for the fragmentation of different ore types, the energy coefficients as well as the ability to grind. Indeed, a time of 5 to 10 minutes can be chosen as the optimal grinding time in a disc mill for a % in weight of the highest pass. However, grinding time can influence the granular characteristics of ore.Keywords: characteristic granular, grinding, mineralogical composition, phosphate ore, parameters
Procedia PDF Downloads 20218752 Artificial Neural Network Modeling and Genetic Algorithm Based Optimization of Hydraulic Design Related to Seepage under Concrete Gravity Dams on Permeable Soils
Authors: Muqdad Al-Juboori, Bithin Datta
Abstract:
Hydraulic structures such as gravity dams are classified as essential structures, and have the vital role in providing strong and safe water resource management. Three major aspects must be considered to achieve an effective design of such a structure: 1) The building cost, 2) safety, and 3) accurate analysis of seepage characteristics. Due to the complexity and non-linearity relationships of the seepage process, many approximation theories have been developed; however, the application of these theories results in noticeable errors. The analytical solution, which includes the difficult conformal mapping procedure, could be applied for a simple and symmetrical problem only. Therefore, the objectives of this paper are to: 1) develop a surrogate model based on numerical simulated data using SEEPW software to approximately simulate seepage process related to a hydraulic structure, 2) develop and solve a linked simulation-optimization model based on the developed surrogate model to describe the seepage occurring under a concrete gravity dam, in order to obtain optimum and safe design at minimum cost. The result shows that the linked simulation-optimization model provides an efficient and optimum design of concrete gravity dams.Keywords: artificial neural network, concrete gravity dam, genetic algorithm, seepage analysis
Procedia PDF Downloads 22418751 Synergism in the Inquiry Lab: An Analysis of Time Targets and Achievement
Authors: John M. Basey, Clinton D. Francis, Maxwell B. Joseph
Abstract:
After gathering data from experimental procedures, inquiry-oriented-science labs often allow students the freedom to stay and complete the write up in class or leave lab early and complete the write up later. Teachers must decide whether to allow students this freedom to self-regulate this time. Student interviews have indicated four time-target strategies that may influence how students utilize this time: grade-target-A, grade-target-C, time-limited, and proficiency. The hypothesis tested was that variability in class composition relative to the four grade-target strategies has an impact on when students leave class, which in turn may influence their overall learning as exemplified by grades. Students were divided into the four indicated groups with a survey. Class composition and the GTA teaching the class had significant impacts on how long students stayed in class with class composition having the greatest impact. A factor analysis identified two factors. Factor 1 included classes with percentages of grade-target students opposite time-limited/proficiency students and explained 43% of the variance. Factor 2 included classes with percentages of grade-target-A/proficiency students opposite grade-target-C students and explained 33% of the variance. Students who stayed longer received significantly higher grades (P = 0.008) with no significant relationships between grade and Factor 1 or Factor 2 (P > 0.05). The time students stayed in class was significantly positively related to Factor 1 (P = 0.006) and significantly negatively related to Factor 2 (P = 0.008). These results support the hypothesis and indicate that teachers may want to know the composition of student-target strategies before deciding on how to have students allocate study time at the end of inquiry-oriented labs. According to these results, ideal classes for self-regulation have a high proportion of proficiency and time-limited students and a low proportion of grade-target students, or a high proportion of grade-target-A and proficiency students and a low proportion of grade-target-C students. Non-ideal classes for self-regulation were comprised of the inverse proportions.Keywords: grades, inquiry lab design, synergism in student motivation, class composition
Procedia PDF Downloads 12918750 Survival Analysis after a First Ischaemic Stroke Event: A Case-Control Study in the Adult Population of England.
Authors: Padma Chutoo, Elena Kulinskaya, Ilyas Bakbergenuly, Nicholas Steel, Dmitri Pchejetski
Abstract:
Stroke is associated with a significant risk of morbidity and mortality. There is scarcity of research on the long-term survival after first-ever ischaemic stroke (IS) events in England with regards to effects of different medical therapies and comorbidities. The objective of this study was to model the all-cause mortality after an IS diagnosis in the adult population of England. Using a retrospective case-control design, we extracted the electronic medical records of patients born prior to or in year 1960 in England with a first-ever ischaemic stroke diagnosis from January 1986 to January 2017 within the Health and Improvement Network (THIN) database. Participants with a history of ischaemic stroke were matched to 3 controls by sex and age at diagnosis and general practice. The primary outcome was the all-cause mortality. The hazards of the all-cause mortality were estimated using a Weibull-Cox survival model which included both scale and shape effects and a shared random effect of general practice. The model included sex, birth cohort, socio-economic status, comorbidities and medical therapies. 20,250 patients with a history of IS (cases) and 55,519 controls were followed up to 30 years. From 2008 to 2015, the one-year all-cause mortality for the IS patients declined with an absolute change of -0.5%. Preventive treatments to cases increased considerably over time. These included prescriptions of statins and antihypertensives. However, prescriptions for antiplatelet drugs decreased in the routine general practice since 2010. The survival model revealed a survival benefit of antiplatelet treatment to stroke survivors with hazard ratio (HR) of 0.92 (0.90 – 0.94). IS diagnosis had significant interactions with gender and age at entry and hypertension diagnosis. IS diagnosis was associated with high risk of all-cause mortality with HR= 3.39 (3.05-3.72) for cases compared to controls. Hypertension was associated with poor survival with HR = 4.79 (4.49 - 5.09) for hypertensive cases relative to non-hypertensive controls, though the detrimental effect of hypertension has not reached significance for hypertensive controls, HR = 1.19(0.82-1.56). This study of English primary care data showed that between 2008 and 2015, the rates of prescriptions of stroke preventive treatments increased, and a short-term all-cause mortality after IS stroke declined. However, stroke resulted in poor long-term survival. Hypertension, a modifiable risk factor, was found to be associated with poor survival outcomes in IS patients. Antiplatelet drugs were found to be protective to survival. Better efforts are required to reduce the burden of stroke through health service development and primary prevention.Keywords: general practice, hazard ratio, health improvement network (THIN), ischaemic stroke, multiple imputation, Weibull-Cox model.
Procedia PDF Downloads 18618749 Self-Carried Theranostic Nanoparticles for in vitro and in vivo Cancer Therapy with Real-Time Monitoring of Drug Release
Authors: Jinfeng Zhang, Chun-Sing Lee
Abstract:
The use of different nanocarriers for delivering hydrophobic pharmaceutical agents to tumor sites has garnered major attention. Despite the merits of these nanocarriers, further studies are needed for improving their drug loading capacities (typically less than 10%) and reducing their potential systemic toxicity. So development of alternative self-carried nanodrug delivery strategies without using any inert carriers is highly desirable. In this study, we developed a self-carried theranostic curcumin (Cur) nanodrug for highly effective cancer therapy in vitro and in vivo with real-time monitoring of drug release. With a biocompatible C18PMH-PEG functionalization, the Cur nanoparticles (NPs) showed excellent dispersibility and outstanding stability in physiological environment, with drug loading capacity higher than 78 wt.%. Both confocal microscopy and flow cytometry confirmed the cellular fluorescent “OFF-ON” activation and real-time monitoring of Cur molecule release, showing its potential for cancer diagnosis. In vitro and in vivo experiments clearly show that therapeutic efficacy of the PEGylated Cur NPs is much better than that of free Cur. This self-carried theranostic strategy with real-time monitoring of drug release may open a new way for simultaneous cancer therapy and diagnosis.Keywords: drug delivery, in vitro and in vivo cancer therapy, real-time monitoring, self-carried
Procedia PDF Downloads 39918748 Discovering the Effects of Meteorological Variables on the Air Quality of Bogota, Colombia, by Data Mining Techniques
Authors: Fabiana Franceschi, Martha Cobo, Manuel Figueredo
Abstract:
Bogotá, the capital of Colombia, is its largest city and one of the most polluted in Latin America due to the fast economic growth over the last ten years. Bogotá has been affected by high pollution events which led to the high concentration of PM10 and NO2, exceeding the local 24-hour legal limits (100 and 150 g/m3 each). The most important pollutants in the city are PM10 and PM2.5 (which are associated with respiratory and cardiovascular problems) and it is known that their concentrations in the atmosphere depend on the local meteorological factors. Therefore, it is necessary to establish a relationship between the meteorological variables and the concentrations of the atmospheric pollutants such as PM10, PM2.5, CO, SO2, NO2 and O3. This study aims to determine the interrelations between meteorological variables and air pollutants in Bogotá, using data mining techniques. Data from 13 monitoring stations were collected from the Bogotá Air Quality Monitoring Network within the period 2010-2015. The Principal Component Analysis (PCA) algorithm was applied to obtain primary relations between all the parameters, and afterwards, the K-means clustering technique was implemented to corroborate those relations found previously and to find patterns in the data. PCA was also used on a per shift basis (morning, afternoon, night and early morning) to validate possible variation of the previous trends and a per year basis to verify that the identified trends have remained throughout the study time. Results demonstrated that wind speed, wind direction, temperature, and NO2 are the most influencing factors on PM10 concentrations. Furthermore, it was confirmed that high humidity episodes increased PM2,5 levels. It was also found that there are direct proportional relationships between O3 levels and wind speed and radiation, while there is an inverse relationship between O3 levels and humidity. Concentrations of SO2 increases with the presence of PM10 and decreases with the wind speed and wind direction. They proved as well that there is a decreasing trend of pollutant concentrations over the last five years. Also, in rainy periods (March-June and September-December) some trends regarding precipitations were stronger. Results obtained with K-means demonstrated that it was possible to find patterns on the data, and they also showed similar conditions and data distribution among Carvajal, Tunal and Puente Aranda stations, and also between Parque Simon Bolivar and las Ferias. It was verified that the aforementioned trends prevailed during the study period by applying the same technique per year. It was concluded that PCA algorithm is useful to establish preliminary relationships among variables, and K-means clustering to find patterns in the data and understanding its distribution. The discovery of patterns in the data allows using these clusters as an input to an Artificial Neural Network prediction model.Keywords: air pollution, air quality modelling, data mining, particulate matter
Procedia PDF Downloads 25818747 Detection of Arterial Stiffness in Diabetes Using Photoplethysmograph
Authors: Neelamshobha Nirala, R. Periyasamy, Awanish Kumar
Abstract:
Diabetes is a metabolic disorder and with the increase of global prevalence of diabetes, cardiovascular diseases and mortality related to diabetes has also increased. Diabetes causes the increase of arterial stiffness by elusive hormonal and metabolic abnormalities. We used photoplethysmograph (PPG), a simple non-invasive method to study the change in arterial stiffness due to diabetes. Toe PPG signals were taken from 29 diabetic subjects with mean age of (65±8.4) years and 21 non-diabetic subjects of mean age of (49±14) years. Mean duration of diabetes is 12±8 years for diabetic group. Rise-time (RT) and area under rise time (AUR) were calculated from the PPG signal of each subject and Welch’s t-test is used to find the significant difference between two groups. We obtained a significant difference of (p-value) 0.0005 and 0.03 for RT and AUR respectively between diabetic and non-diabetic subjects. Average value of RT and AUR is 0.298±0.003 msec and 14.4±4.2 arbitrary units respectively for diabetic subject compared to 0.277±0.0005 msec and 13.66±2.3 a.u respectively for non-diabetic subjects. In conclusion, this study support that arterial stiffness is increased in diabetes and can be detected early using PPG.Keywords: area under rise-time, AUR, arterial stiffness, diabetes, photoplethysmograph, PPG, rise-time (RT)
Procedia PDF Downloads 25918746 Consultation Time and Its Impact on Length of Stay in the Emergency Department
Authors: Esam Roshdy, Saleh AlRashdi, Turki Alharbi, Rawan Eskandarani, Zurina Cabilo
Abstract:
Introduction/ background: Consultation in the Emergency Department constitute a major part of the work flow every day. Any delay in the consultation process have a major impact on the length of stay and patient disposition and thus affect the total waiting time of patients in the ED. King Fahad medical City in Riyadh City, Saudi Arabia is considered a major Tertiary hospital where there is high flow of patients of different categories visiting the ED. The importance of decreasing consultation time and decision for final disposition of patients was recognized and interpreted in this project to find ways to improve the patient flow in the department and thus the total patient disposition and outcome. Aim / Objectives: 1. To monitor the time of consultation for patients in the Emergency department and its impact on the length of stay of patients in the ED. 2. To detect and assess the problems that lead to long consultation times in the ED, and reach a targeted time of 2 hours for final disposition of patients, according to recognized international and our institutional consultation policy, to reach the final goal of decreasing total length of stay and thus improve the patient flow in the ED. Methods: Data was collected retrospectively for a 92 charts of consultations done in the ED over 2 month’s period. The data was analyzed to get the median of Total Consultation Time. A survey was conducted among all ED staff to determine the level of knowledge about the total consultation time and the compliance to the institutional policy target of 2 hours. A second Data sample of 168 chart was collected after awareness campaign and education of all ED staff about the importance of reaching the target consultation time and compliance to the institutional policy. Results: We have found that there is room for improvement in our overall consultation time. This was found to be more frequent with certain specialties. Our surveys have showed that many ED staff are not familiar or not compliant with our consultation policy which was not clear for everyone. Post-intervention data have showed that awareness of the importance to decrease the total consultation time and compliance alone to the targeted goal have had a huge impact on overall improvement and decreasing the time of final decision and disposition of the patient and the overall patient length of stay in the ED. Conclusion: Working on improving Consultation time in the Emergency Department is a major factor in improving overall length of stay and patient flow. This improvement helps in the overall patient disposition and satisfaction. Plan: As a continuation of our project we are planning to focus on the conflict of admission cases where more than one specialty is involved in the care of patients. We are planning to collect data on the time it takes to resolve and reach final disposition of those patients, and its impact on the length of stay and our department flow and the overall patient outcome and satisfaction.Keywords: consultation time, impact, length of stay, in the ED
Procedia PDF Downloads 28918745 Implementation in Python of a Method to Transform One-Dimensional Signals in Graphs
Authors: Luis Andrey Fajardo Fajardo
Abstract:
We are immersed in complex systems. The human brain, the galaxies, the snowflakes are examples of complex systems. An area of interest in Complex systems is the chaos theory. This revolutionary field of science presents different ways of study than determinism and reductionism. Here is where in junction with the Nonlinear DSP, chaos theory offer valuable techniques that establish a link between time series and complex theory in terms of complex networks, so that, the study of signals can be explored from the graph theory. Recently, some people had purposed a method to transform time series in graphs, but no one had developed a suitable implementation in Python with signals extracted from Chaotic Systems or Complex systems. That’s why the implementation in Python of an existing method to transform one dimensional chaotic signals from time domain to graph domain and some measures that may reveal information not extracted in the time domain is proposed.Keywords: Python, complex systems, graph theory, dynamical systems
Procedia PDF Downloads 50918744 Physics-Informed Machine Learning for Displacement Estimation in Solid Mechanics Problem
Authors: Feng Yang
Abstract:
Machine learning (ML), especially deep learning (DL), has been extensively applied to many applications in recently years and gained great success in solving different problems, including scientific problems. However, conventional ML/DL methodologies are purely data-driven which have the limitations, such as need of ample amount of labelled training data, lack of consistency to physical principles, and lack of generalizability to new problems/domains. Recently, there is a growing consensus that ML models need to further take advantage of prior knowledge to deal with these limitations. Physics-informed machine learning, aiming at integration of physics/domain knowledge into ML, has been recognized as an emerging area of research, especially in the recent 2 to 3 years. In this work, physics-informed ML, specifically physics-informed neural network (NN), is employed and implemented to estimate the displacements at x, y, z directions in a solid mechanics problem that is controlled by equilibrium equations with boundary conditions. By incorporating the physics (i.e. the equilibrium equations) into the learning process of NN, it is showed that the NN can be trained very efficiently with a small set of labelled training data. Experiments with different settings of the NN model and the amount of labelled training data were conducted, and the results show that very high accuracy can be achieved in fulfilling the equilibrium equations as well as in predicting the displacements, e.g. in setting the overall displacement of 0.1, a root mean square error (RMSE) of 2.09 × 10−4 was achieved.Keywords: deep learning, neural network, physics-informed machine learning, solid mechanics
Procedia PDF Downloads 15018743 Numerical Analysis of a Reaction Diffusion System of Lambda-Omega Type
Authors: Hassan J. Al Salman, Ahmed A. Al Ghafli
Abstract:
In this study, we consider a nonlinear in time finite element approximation of a reaction diffusion system of lambda-omega type. We use a fixed-point theorem to prove existence of the approximations at each time level. Then, we derive some essential stability estimates and discuss the uniqueness of the approximations. In addition, we employ Nochetto mathematical framework to prove an optimal error bound in time for d= 1, 2 and 3 space dimensions. Finally, we present some numerical experiments to verify the obtained theoretical results.Keywords: reaction diffusion system, finite element approximation, stability estimates, error bound
Procedia PDF Downloads 43018742 Development of Prediction Models of Day-Ahead Hourly Building Electricity Consumption and Peak Power Demand Using the Machine Learning Method
Authors: Dalin Si, Azizan Aziz, Bertrand Lasternas
Abstract:
To encourage building owners to purchase electricity at the wholesale market and reduce building peak demand, this study aims to develop models that predict day-ahead hourly electricity consumption and demand using artificial neural network (ANN) and support vector machine (SVM). All prediction models are built in Python, with tool Scikit-learn and Pybrain. The input data for both consumption and demand prediction are time stamp, outdoor dry bulb temperature, relative humidity, air handling unit (AHU), supply air temperature and solar radiation. Solar radiation, which is unavailable a day-ahead, is predicted at first, and then this estimation is used as an input to predict consumption and demand. Models to predict consumption and demand are trained in both SVM and ANN, and depend on cooling or heating, weekdays or weekends. The results show that ANN is the better option for both consumption and demand prediction. It can achieve 15.50% to 20.03% coefficient of variance of root mean square error (CVRMSE) for consumption prediction and 22.89% to 32.42% CVRMSE for demand prediction, respectively. To conclude, the presented models have potential to help building owners to purchase electricity at the wholesale market, but they are not robust when used in demand response control.Keywords: building energy prediction, data mining, demand response, electricity market
Procedia PDF Downloads 31618741 A Network Economic Analysis of Friendship, Cultural Activity, and Homophily
Authors: Siming Xie
Abstract:
In social networks, the term homophily refers to the tendency of agents with similar characteristics to link with one another and is so robustly observed across many contexts and dimensions. The starting point of my research is the observation that the “type” of agents is not a single exogenous variable. Agents, despite their differences in race, religion, and other hard to alter characteristics, may share interests and engage in activities that cut across those predetermined lines. This research aims to capture the interactions of homophily effects in a model where agents have two-dimension characteristics (i.e., race and personal hobbies such as basketball, which one either likes or dislikes) and with biases in meeting opportunities and in favor of same-type friendships. A novel feature of my model is providing a matching process with biased meeting probability on different dimensions, which could help to understand the structuring process in multidimensional networks without missing layer interdependencies. The main contribution of this study is providing a welfare based matching process for agents with multi-dimensional characteristics. In particular, this research shows that the biases in meeting opportunities on one dimension would lead to the emergence of homophily on the other dimension. The objective of this research is to determine the pattern of homophily in network formations, which will shed light on our understanding of segregation and its remedies. By constructing a two-dimension matching process, this study explores a method to describe agents’ homophilous behavior in a social network with multidimension and construct a game in which the minorities and majorities play different strategies in a society. It also shows that the optimal strategy is determined by the relative group size, where society would suffer more from social segregation if the two racial groups have a similar size. The research also has political implications—cultivating the same characteristics among agents helps diminishing social segregation, but only if the minority group is small enough. This research includes both theoretical models and empirical analysis. Providing the friendship formation model, the author first uses MATLAB to perform iteration calculations, then derives corresponding mathematical proof on previous results, and last shows that the model is consistent with empirical evidence from high school friendships. The anonymous data comes from The National Longitudinal Study of Adolescent Health (Add Health).Keywords: homophily, multidimension, social networks, friendships
Procedia PDF Downloads 17018740 User Experience Measurement of User Interfaces
Authors: Mohammad Hashemi, John Herbert
Abstract:
Quantifying and measuring Quality of Experience (QoE) are important and difficult concerns in Human Computer Interaction (HCI). Quality of Service (QoS) and the actual User Interface (UI) of the application are both important contributors to the QoE of a user. This paper describes a framework that measures accurately the way a user uses the UI in order to model users' behaviours and profiles. It monitors the use of the mouse and use of UI elements with accurate time measurement. It does this in real-time and does so unobtrusively and efficiently allowing the user to work as normal with the application. This real-time accurate measurement of the user's interaction provides valuable data and insight into the use of the UI, and is also the basis for analysis of the user's QoE.Keywords: user modelling, user interface experience, quality of experience, user experience, human and computer interaction
Procedia PDF Downloads 50418739 From Ride-Hailing App to Diversified and Sustainable Platform Business Model
Authors: Ridwan Dewayanto Rusli
Abstract:
We show how prisoner's dilemma-type competition problems can be mitigated through rapid platform diversification and ecosystem expansion. We analyze a ride-hailing company in Southeast Asia, Gojek, whose network grew to more than 170 million users comprising consumers, partner drivers, merchants, and complementors within a few years and has already achieved higher contribution margins than ride-hailing peers Uber and Lyft. Its ecosystem integrates ride-hailing, food delivery and logistics, merchant solutions, e-commerce, marketplace and advertising, payments, and fintech offerings. The company continues growing its network of complementors and App developers, expanding content and gaining critical mass in consumer data analytics and advertising. We compare the company's growth and diversification trajectory with those of its main international rivals and peers. The company's rapid growth and future potential are analyzed using Cusumano's (2012) Staying Power and Six Principles, Hax and Wilde's (2003) and Hax's (2010) The Delta Model as well as Santos' (2016) home-market advantages frameworks. The recently announced multi-billion-dollar merger with one of Southeast Asia's largest e-commerce majors lends additional support to the above arguments.Keywords: ride-hailing, prisoner's dilemma, platform and ecosystem strategy, digital applications, diversification, home market advantages, e-commerce
Procedia PDF Downloads 94