Search results for: time series models
23295 A Survey on the Blockchain Smart Contract System: Security Strengths and Weaknesses
Authors: Malaw Ndiaye, Karim Konate
Abstract:
Smart contracts are computer protocols that facilitate, verify, and execute the negotiation or execution of a contract, or that render a contractual term unnecessary. Blockchain and smart contracts can be used to facilitate almost any financial transaction. Thanks to these smart contracts, the settlement of dividends and coupons could be automated. Smart contracts have become lucrative and profitable targets for attackers because they can hold a great amount of money. Smart contracts, although widely used in blockchain technology, are far from perfect due to security concerns. Since there are recent studies on smart contract security, none of them systematically study the strengths and weaknesses of smart contract security. Some have focused on an analysis of program-related vulnerabilities by providing a taxonomy of vulnerabilities. Other studies are responsible for listing the series of attacks linked to smart contracts. Although a series of attacks are listed, there is a lack of discussions and proposals on improving security. This survey takes stock of smart contract security from a more comprehensive perspective by correlating the level of vulnerability and systematic review of security levels in smart contracts.Keywords: blockchain, Bitcoin, smart contract, criminal smart contract, security
Procedia PDF Downloads 16823294 Filtering Momentum Life Cycles, Price Acceleration Signals and Trend Reversals for Stocks, Credit Derivatives and Bonds
Authors: Periklis Brakatsoulas
Abstract:
Recent empirical research shows a growing interest in investment decision-making under market anomalies that contradict the rational paradigm. Momentum is undoubtedly one of the most robust anomalies in the empirical asset pricing research and remains surprisingly lucrative ever since first documented. Although predominantly phenomena identified across equities, momentum premia are now evident across various asset classes. Yet few many attempts are made so far to provide traders a diversified portfolio of strategies across different assets and markets. Moreover, literature focuses on patterns from past returns rather than mechanisms to signal future price directions prior to momentum runs. The aim of this paper is to develop a diversified portfolio approach to price distortion signals using daily position data on stocks, credit derivatives, and bonds. An algorithm allocates assets periodically, and new investment tactics take over upon price momentum signals and across different ranking groups. We focus on momentum life cycles, trend reversals, and price acceleration signals. The main effort here concentrates on the density, time span and maturity of momentum phenomena to identify consistent patterns over time and measure the predictive power of buy-sell signals generated by these anomalies. To tackle this, we propose a two-stage modelling process. First, we generate forecasts on core macroeconomic drivers. Secondly, satellite models generate market risk forecasts using the core driver projections generated at the first stage as input. Moreover, using a combination of the ARFIMA and FIGARCH models, we examine the dependence of consecutive observations across time and portfolio assets since long memory behavior in volatilities of one market appears to trigger persistent volatility patterns across other markets. We believe that this is the first work that employs evidence of volatility transmissions among derivatives, equities, and bonds to identify momentum life cycle patterns.Keywords: forecasting, long memory, momentum, returns
Procedia PDF Downloads 10223293 Predicting Provider Service Time in Outpatient Clinics Using Artificial Intelligence-Based Models
Authors: Haya Salah, Srinivas Sharan
Abstract:
Healthcare facilities use appointment systems to schedule their appointments and to manage access to their medical services. With the growing demand for outpatient care, it is now imperative to manage physician's time effectively. However, high variation in consultation duration affects the clinical scheduler's ability to estimate the appointment duration and allocate provider time appropriately. Underestimating consultation times can lead to physician's burnout, misdiagnosis, and patient dissatisfaction. On the other hand, appointment durations that are longer than required lead to doctor idle time and fewer patient visits. Therefore, a good estimation of consultation duration has the potential to improve timely access to care, resource utilization, quality of care, and patient satisfaction. Although the literature on factors influencing consultation length abound, little work has done to predict it using based data-driven approaches. Therefore, this study aims to predict consultation duration using supervised machine learning algorithms (ML), which predicts an outcome variable (e.g., consultation) based on potential features that influence the outcome. In particular, ML algorithms learn from a historical dataset without explicitly being programmed and uncover the relationship between the features and outcome variable. A subset of the data used in this study has been obtained from the electronic medical records (EMR) of four different outpatient clinics located in central Pennsylvania, USA. Also, publicly available information on doctor's characteristics such as gender and experience has been extracted from online sources. This research develops three popular ML algorithms (deep learning, random forest, gradient boosting machine) to predict the treatment time required for a patient and conducts a comparative analysis of these algorithms with respect to predictive performance. The findings of this study indicate that ML algorithms have the potential to predict the provider service time with superior accuracy. While the current approach of experience-based appointment duration estimation adopted by the clinic resulted in a mean absolute percentage error of 25.8%, the Deep learning algorithm developed in this study yielded the best performance with a MAPE of 12.24%, followed by gradient boosting machine (13.26%) and random forests (14.71%). Besides, this research also identified the critical variables affecting consultation duration to be patient type (new vs. established), doctor's experience, zip code, appointment day, and doctor's specialty. Moreover, several practical insights are obtained based on the comparative analysis of the ML algorithms. The machine learning approach presented in this study can serve as a decision support tool and could be integrated into the appointment system for effectively managing patient scheduling.Keywords: clinical decision support system, machine learning algorithms, patient scheduling, prediction models, provider service time
Procedia PDF Downloads 12123292 Functionality Based Composition of Web Services to Attain Maximum Quality of Service
Authors: M. Mohemmed Sha Mohamed Kunju, Abdalla A. Al-Ameen Abdurahman, T. Manesh Thankappan, A. Mohamed Mustaq Ahmed Hameed
Abstract:
Web service composition is an effective approach to complete the web based tasks with desired quality. A single web service with limited functionality is inadequate to execute a specific task with series of action. So, it is very much required to combine multiple web services with different functionalities to reach the target. Also, it will become more and more challenging, when these services are from different providers with identical functionalities and varying QoS, so while composing the web services, the overall QoS is considered to be the major factor. Also, it is not true that the expected QoS is always attained when the task is completed. A single web service in the composed chain may affect the overall performance of the task. So care should be taken in different aspects such as functionality of the service, while composition. Dynamic and automatic service composition is one of the main option available. But to achieve the actual functionality of the task, quality of the individual web services are also important. Normally the QoS of the individual service can be evaluated by using the non-functional parameters such as response time, throughput, reliability, availability, etc. At the same time, the QoS is not needed to be at the same level for all the composed services. So this paper proposes a framework that allows composing the services in terms of QoS by setting the appropriate weight to the non-functional parameters of each individual web service involved in the task. Experimental results show that the importance given to the non-functional parameter while composition will definitely improve the performance of the web services.Keywords: composition, non-functional parameters, quality of service, web service
Procedia PDF Downloads 33323291 Preparation of Sorbent Materials for the Removal of Hardness and Organic Pollutants from Water and Wastewater
Authors: Thanaa Abdel Moghny, Mohamed Keshawy, Mahmoud Fathy, Abdul-Raheim M. Abdul-Raheim, Khalid I. Kabel, Ahmed F. El-Kafrawy, Mahmoud Ahmed Mousa, Ahmed E. Awadallah
Abstract:
Ecological pollution is of great concern for human health and the environment. Numerous organic and inorganic pollutants usually discharged into the water caused carcinogenic or toxic effect for human and different life form. In this respect, this work aims to treat water contaminated by organic and inorganic waste using sorbent based on polystyrene. Therefore, two different series of adsorbent material were prepared; the first one included the preparation of polymeric sorbent from the reaction of styrene acrylate ester and alkyl acrylate. The second series involved syntheses of composite ion exchange resins of waste polystyrene and amorphous carbon thin film (WPS/ACTF) by solvent evaporation using micro emulsion polymerization. The produced ACTF/WPS nanocomposite was sulfonated to produce cation exchange resins ACTF/WPSS nanocomposite. The sorbents of the first series were characterized using FTIR, 1H NMR, and gel permeation chromatography. The thermal properties of the cross-linked sorbents were investigated using thermogravimetric analysis, and the morphology was characterized by scanning electron microscope (SEM). The removal of organic pollutant was determined through absorption tests in a various organic solvent. The chemical and crystalline structure of nanocomposite of second series has been proven by studies of FTIR spectrum, X-rays, thermal analysis, SEM and TEM analysis to study morphology of resins and ACTF that assembled with polystyrene chain. It is found that the composite resins ACTF/WPSS are thermally stable and show higher chemical stability than ion exchange WPSS resins. The composite resin was evaluated for calcium hardness removal. The result is evident that the ACTF/WPSS composite has more prominent inorganic pollutant removal than WPSS resin. So, we recommend the using of nanocomposite resin as new potential applications for water treatment process.Keywords: nanocomposite, sorbent materials, waste water, waste polystyrene
Procedia PDF Downloads 42923290 The Changing Face of Pedagogy and Curriculum Development Sub-Components of Teacher Education in Nigeria: A Comparative Evaluation of the University of Lagos, Lagos State University, and Sokoto State University Models
Authors: Saheed A. Rufai
Abstract:
Courses in Pedagogy and Curriculum Development expectedly occupy a core place in the professional education components of teacher education at Lagos, Lagos State, and Sokoto State Universities. This is in keeping with the National Teacher Education Policy statement that stipulates that for student teachers to learn effectively teacher education institutions must be equipped to prepare them adequately. However, there is a growing concern over the unfaithfulness of some of the dominant Nigerian models of teacher education, to this policy statement on teacher educators’ knowledge and skills. The purpose of this paper is to comparatively evaluate both the curricular provisions and the manpower for the pedagogy and curriculum development sub-components of the Lagos, Lagos State, and Sokoto State models of teacher preparation. The paper employs a combination of quantitative and qualitative methods. Preliminary analysis revealed a new trend in teacher educators’ pedagogical knowledge and understanding, with regard to the two intertwined sub-components. The significance of such a study lies in its potential to determine the degree of conformity of each of the three models to the stipulated standards. The paper’s contribution to scholarship lies in its correlation of deficiencies in teacher educators’ professional knowledge and skills and articulation of the implications of such deficiencies for the professional knowledge and skills of the prospective teachers, with a view to providing a framework for reforms.Keywords: curriculum development, pedagogy, teacher education, dominant Nigerian teacher preparation models
Procedia PDF Downloads 44323289 Overall Determinants of Foreign Direct Investment Inflows in Kenya
Authors: George Ogono Muok, N. Obange, S. A. Odhiambo
Abstract:
Empirical literature on the determinants of foreign direct investments (FDI) flows is extensive but controversial over some determinants of FDI in-flows in developing countries. The objective of this study therefore was to investigate the overall determinants of FDI inflows in Kenya. Dynamic macroeconomic theory and correlational study design provided theoretical framework for specification of a time series model. The study used data observed from 1970 to 2015 in World Development Indicators (WDI) data bank. The results show that annual growth rate of GDP, inflation rates and external debt as a proportion of GDP are significant determinants of FDI inflows in Kenya and are therefore important macroeconomic parameters for policy formulation for promotion of FDI inflows in Kenya.Keywords: determinants of foreign, direct, investment inflows in, Kenya, Africa
Procedia PDF Downloads 28523288 Statistical Analysis of Natural Images after Applying ICA and ISA
Authors: Peyman Sheikholharam Mashhadi
Abstract:
Difficulties in analyzing real world images in classical image processing and machine vision framework have motivated researchers towards considering the biology-based vision. It is a common belief that mammalian visual cortex has been adapted to the statistics of the real world images through the evolution process. There are two well-known successful models of mammalian visual cortical cells: Independent Component Analysis (ICA) and Independent Subspace Analysis (ISA). In this paper, we statistically analyze the dependencies which remain in the components after applying these models to the natural images. Also, we investigate the response of feature detectors to gratings with various parameters in order to find optimal parameters of the feature detectors. Finally, the selectiveness of feature detectors to phase, in both models is considered.Keywords: statistics, independent component analysis, independent subspace analysis, phase, natural images
Procedia PDF Downloads 33923287 Time Management in the Public Sector in Nigeria
Authors: Sunny Ewankhiwimen Aigbomian
Abstract:
Time, is a scarce resource and in everything we do, time is required to accomplish any given task. The need for this presentation is predicated on the way majority of Nigerian especially in the public sector operators see “Time Management”. Time as resources cannot be regained if lost or managed badly. As a significant aspect of human life it should be handled with diligence and utmost seriousness if the public sector is to function as a coordinated entity. In our homes, private life and offices, we schedule different things to ensure that some things do not go the unexpected. When it comes to service delivery on the part of government, it ought to be more serious because government is all about effect and efficient service delivery and “Time” is a significant variable necessary to successful accomplishment. The need for Nigerian government to re-examine time management in her public sector with a view of repositioning the sector to be able to compete well with other public sectors in the world. The peculiarity of Time management in Public Sector in Nigerian context as examined and some useful recommendations of immerse assistance proffered.Keywords: Nigeria, public sector, time management, task
Procedia PDF Downloads 9923286 Modeling and Shape Prediction for Elastic Kinematic Chains
Authors: Jiun Jeon, Byung-Ju Yi
Abstract:
This paper investigates modeling and shape prediction of elastic kinematic chains such as colonoscopy. 2D and 3D models of elastic kinematic chains are suggested and their behaviors are demonstrated through simulation. To corroborate the effectiveness of those models, experimental work is performed using a magnetic sensor system.Keywords: elastic kinematic chain, shape prediction, colonoscopy, modeling
Procedia PDF Downloads 60523285 Batch Kinetic, Isotherm and Thermodynamic Studies of Copper (II) Removal from Wastewater Using HDL as Adsorbent
Authors: Nadjet Taoualit, Zoubida Chemat, Djamel-Eddine Hadj-Boussaad
Abstract:
This study aims the removal of copper Cu (II) contained in wastewater by adsorption on a perfect synthesized mud. It is the materials Hydroxides Double Lamellar, HDL, prepared and synthesized by co-precipitation method at constant pH, which requires a simple titration assembly, with an inexpensive and available material in the laboratory, and also allows us better control of the composition of the reaction medium, and gives well crystallized products. A characterization of the adsorbent proved essential. Thus a range of physic-chemical analysis was performed including: FTIR spectroscopy, X-ray diffraction… The adsorption of copper ions was investigated in dispersed medium (batch). A systematic study of various parameters (amount of support, contact time, initial copper concentration, temperature, pH…) was performed. Adsorption kinetic data were tested using pseudo-first order, pseudo-second order, Bangham's equation and intra-particle diffusion models. The equilibrium data were analyzed using Langmuir, Freundlich, Tempkin and other isotherm models at different doses of HDL. The thermodynamics parameters were evaluated at different temperatures. The results have established good potentiality for the HDL to be used as a sorbent for the removal of Copper from wastewater.Keywords: adsoption, copper, HDL, isotherm
Procedia PDF Downloads 27523284 Synthesis, Characterization, and Evaluation of New Series of Oil Sorbers Based on Maleate Esters
Authors: Nora A. Hamad, Ayman M. Atta, Adel A. H. Abdel-Rahman
Abstract:
Two malice anhydride esters were prepared using long chain aliphatic alcohols (C8H17OH and C12H25OH, 1:1 mole ratio). Three series of crosslinked homo and copolymers of maleate esters with octadecyl acrylate and acrylic acid were prepared respectively through suspension copolymerization. The monomers were mixed with 0.02 Wt% of BP initiator, PVA 1% (170 ml for each 100g of monomers) and different weight ratios of DVB crosslinked (1% and 4%) in cyclohexane. The prepared crosslinked homo and copolymers were characterized by SEM, TGA and FTIR spectroscopic analyses. The prepared polymers were coated onto poly (ethylene terephethalate) nonwoven fiber (NWPET). The effect of copolymerization feed composition, crosslinker wt% and reaction media or solvent on swelling properties of crosslinked polymers were studied through the oil absorption tests in toluene and 10% of diluted crude oil with toluene.Keywords: acrylic acid, crosslinked copolymers, maleate ester, poly(ethylene terephethalate) nonwoven fiber (NWPET), oil absorbency, octadecyl acrylat
Procedia PDF Downloads 39123283 Study of Strontium Sorption onto Indian Bentonite
Authors: Pankaj Pathak, Susmita Sharma
Abstract:
Incessant industrial growth fulfill the energy demand of present day society, at the same time it produces huge amount of waste which could be hazardous or non-hazardous in nature. These wastes are coming out from different sources viz, nuclear power, thermal power, coal mines which contain different types of contaminants and one of the emergent contaminant is strontium, used in the present study. The isotope of strontium (Sr90) is radioactive in nature with half-life of 28.8 years and permissible limit of strontium in drinking water is 1.5 ppm. Above the permissible limit causes several types of diseases in human being. Therefore, safe disposal of strontium into ground becomes a biggest challenge for the researchers. In this context, bentonite is being used as an efficient material to retain strontium onto ground due to its specific physical, chemical and mineralogical properties which exhibits higher cation exchange capacity and specific surface area. These properties influence the interaction between strontium and bentonite, which is quantified by employing a parameter known as distribution coefficient. Batch test was conducted, and sorption isotherms were modelled at different interaction time. The pseudo first-order and pseudo second order kinetic models have been used to fit experimental data, which helps to determine the sorption rate and mechanism.Keywords: bentonite, interaction time, sorption, strontium
Procedia PDF Downloads 30523282 The Models of Character Development Bali Police to Improve Quality of Moral Members in Bali Police Headquarters
Authors: Agus Masrukhin
Abstract:
This research aims to find and analyze the model of character building in the Police Headquarters in Bali with a case study of Muslim members in improving the quality of the morality of its members. The formation of patterns of thinking, behavior, mentality, and police officers noble character, later can be used as a solution to reduce the hedonistic nature of the challenges in the era of globalization. The benefit of this study is expected to be a positive recommendation to find a constructive character building models of police officers in the Republic of Indonesia, especially Bali Police. For the long term, the discovery of the character building models can be developed for the entire police force in Indonesia. The type of research that would apply in this study researchers mix the qualitative research methods based on the narrative between the subject and the concrete experience of field research and quantitative research methods with 92 respondents from the police regional police Bali. This research used a descriptive analysis and SWOT analysis then it is presented in the FGD (focus group discussion). The results of this research indicate that the variable modeling the leadership of the police and variable police offices culture have significant influence on the implementation of spiritual development.Keywords: positive constructive, hedonistic, character models, morality
Procedia PDF Downloads 36523281 Innovation Trends in Latin America Countries
Authors: José Carlos Rodríguez, Mario Gómez
Abstract:
This paper analyses innovation trends in Latin America countries by means of the number of patent applications filed by residents and non-residents during the period 1965 to 2012. Making use of patent data released by the World Intellectual Property Organization (WIPO), we search for the presence of multiple structural changes in patent application series in Argentina, Brazil Chile, and Mexico. These changes may suggest that firms’ innovative activity has been modified as a result of implementing a particular science, technology and innovation (STI) policy. Accordingly, the new regulations implemented in these countries during 1980s and 1990s have influenced their intellectual property regimes. The question conducting this research is thus how STI policies in these countries have affected their innovation activity? The results achieved in this research confirm the existence of multiple structural changes in the series of patent applications resulting from STI policies implemented in these countries.Keywords: econometric methods, innovation activity, Latin America countries, patents, science, technology and innovation policy
Procedia PDF Downloads 28323280 Comparative Mesh Sensitivity Study of Different Reynolds Averaged Navier Stokes Turbulence Models in OpenFOAM
Authors: Zhuoneng Li, Zeeshan A. Rana, Karl W. Jenkins
Abstract:
In industry, to validate a case, often a multitude of simulation are required and in order to demonstrate confidence in the process where users tend to use a coarser mesh. Therefore, it is imperative to establish the coarsest mesh that could be used while keeping reasonable simulation accuracy. To date, the two most reliable, affordable and broadly used advanced simulations are the hybrid RANS (Reynolds Averaged Navier Stokes)/LES (Large Eddy Simulation) and wall modelled LES. The potentials in these two simulations will still be developed in the next decades mainly because the unaffordable computational cost of a DNS (Direct Numerical Simulation). In the wall modelled LES, the turbulence model is applied as a sub-grid scale model in the most inner layer near the wall. The RANS turbulence models cover the entire boundary layer region in a hybrid RANS/LES (Detached Eddy Simulation) and its variants, therefore, the RANS still has a very important role in the state of art simulations. This research focuses on the turbulence model mesh sensitivity analysis where various turbulence models such as the S-A (Spalart-Allmaras), SSG (Speziale-Sarkar-Gatski), K-Omega transitional SST (Shear Stress Transport), K-kl-Omega, γ-Reθ transitional model, v2f are evaluated within the OpenFOAM. The simulations are conducted on a fully developed turbulent flow over a flat plate where the skin friction coefficient as well as velocity profiles are obtained to compare against experimental values and DNS results. A concrete conclusion is made to clarify the mesh sensitivity for different turbulence models.Keywords: mesh sensitivity, turbulence models, OpenFOAM, RANS
Procedia PDF Downloads 26123279 Coordination of Traffic Signals on Arterial Streets in Duhok City
Authors: Dilshad Ali Mohammed, Ziyad Nayef Shamsulddin Aldoski, Millet Salim Mohammed
Abstract:
The increase in levels of traffic congestion along urban signalized arterials needs efficient traffic management. The application of traffic signal coordination can improve the traffic operation and safety for a series of signalized intersection along the arterials. The objective of this study is to evaluate the benefits achievable through actuated traffic signal coordination and make a comparison in control delay against the same signalized intersection in case of being isolated. To accomplish this purpose, a series of eight signalized intersections located on two major arterials in Duhok City was chosen for conducting the study. Traffic data (traffic volumes, link and approach speeds, and passenger car equivalent) were collected at peak hours. Various methods had been used for collecting data such as video recording technique, moving vehicle method and manual methods. Geometric and signalization data were also collected for the purpose of the study. The coupling index had been calculated to check the coordination attainability, and then time space diagrams were constructed representing one-way coordination for the intersections on Barzani and Zakho Streets, and others represented two-way coordination for the intersections on Zakho Street with accepted progression bandwidth efficiency. The results of this study show great progression bandwidth of 54 seconds for east direction coordination and 17 seconds for west direction coordination on Barzani Street under suggested controlled speed of 60 kph agreeable with the present data. For Zakho Street, the progression bandwidth is 19 seconds for east direction coordination and 18 seconds for west direction coordination under suggested controlled speed of 40 kph. The results show that traffic signal coordination had led to high reduction in intersection control delays on both arterials.Keywords: bandwidth, congestion, coordination, traffic, signals, streets
Procedia PDF Downloads 30723278 Modeling of System Availability and Bayesian Analysis of Bivariate Distribution
Authors: Muhammad Farooq, Ahtasham Gul
Abstract:
To meet the desired standard, it is important to monitor and analyze different engineering processes to get desired output. The bivariate distributions got a lot of attention in recent years to describe the randomness of natural as well as artificial mechanisms. In this article, a bivariate model is constructed using two independent models developed by the nesting approach to study the effect of each component on reliability for better understanding. Further, the Bayes analysis of system availability is studied by considering prior parametric variations in the failure time and repair time distributions. Basic statistical characteristics of marginal distribution, like mean median and quantile function, are discussed. We use inverse Gamma prior to study its frequentist properties by conducting Monte Carlo Markov Chain (MCMC) sampling scheme.Keywords: reliability, system availability Weibull, inverse Lomax, Monte Carlo Markov Chain, Bayesian
Procedia PDF Downloads 7123277 Bayesian Value at Risk Forecast Using Realized Conditional Autoregressive Expectiel Mdodel with an Application of Cryptocurrency
Authors: Niya Chen, Jennifer Chan
Abstract:
In the financial market, risk management helps to minimize potential loss and maximize profit. There are two ways to assess risks; the first way is to calculate the risk directly based on the volatility. The most common risk measurements are Value at Risk (VaR), sharp ratio, and beta. Alternatively, we could look at the quantile of the return to assess the risk. Popular return models such as GARCH and stochastic volatility (SV) focus on modeling the mean of the return distribution via capturing the volatility dynamics; however, the quantile/expectile method will give us an idea of the distribution with the extreme return value. It will allow us to forecast VaR using return which is direct information. The advantage of using these non-parametric methods is that it is not bounded by the distribution assumptions from the parametric method. But the difference between them is that expectile uses a second-order loss function while quantile regression uses a first-order loss function. We consider several quantile functions, different volatility measures, and estimates from some volatility models. To estimate the expectile of the model, we use Realized Conditional Autoregressive Expectile (CARE) model with the bayesian method to achieve this. We would like to see if our proposed models outperform existing models in cryptocurrency, and we will test it by using Bitcoin mainly as well as Ethereum.Keywords: expectile, CARE Model, CARR Model, quantile, cryptocurrency, Value at Risk
Procedia PDF Downloads 10923276 Electroforming of 3D Digital Light Processing Printed Sculptures Used as a Low Cost Option for Microcasting
Authors: Cecile Meier, Drago Diaz Aleman, Itahisa Perez Conesa, Jose Luis Saorin Perez, Jorge De La Torre Cantero
Abstract:
In this work, two ways of creating small-sized metal sculptures are proposed: the first by means of microcasting and the second by electroforming from models printed in 3D using an FDM (Fused Deposition Modeling) printer or using a DLP (Digital Light Processing) printer. It is viable to replace the wax in the processes of the artistic foundry with 3D printed objects. In this technique, the digital models are manufactured with resin using a low-cost 3D FDM printer in polylactic acid (PLA). This material is used, because its properties make it a viable substitute to wax, within the processes of artistic casting with the technique of lost wax through Ceramic Shell casting. This technique consists of covering a sculpture of wax or in this case PLA with several layers of thermoresistant material. This material is heated to melt the PLA, obtaining an empty mold that is later filled with the molten metal. It is verified that the PLA models reduce the cost and time compared with the hand modeling of the wax. In addition, one can manufacture parts with 3D printing that are not possible to create with manual techniques. However, the sculptures created with this technique have a size limit. The problem is that when printed pieces with PLA are very small, they lose detail, and the laminar texture hides the shape of the piece. DLP type printer allows obtaining more detailed and smaller pieces than the FDM. Such small models are quite difficult and complex to melt using the lost wax technique of Ceramic Shell casting. But, as an alternative, there are microcasting and electroforming, which are specialized in creating small metal pieces such as jewelry ones. The microcasting is a variant of the lost wax that consists of introducing the model in a cylinder in which the refractory material is also poured. The molds are heated in an oven to melt the model and cook them. Finally, the metal is poured into the still hot cylinders that rotate in a machine at high speed to properly distribute all the metal. Because microcasting requires expensive material and machinery to melt a piece of metal, electroforming is an alternative for this process. The electroforming uses models in different materials; for this study, micro-sculptures printed in 3D are used. These are subjected to an electroforming bath that covers the pieces with a very thin layer of metal. This work will investigate the recommended size to use 3D printers, both with PLA and resin and first tests are being done to validate use the electroforming process of microsculptures, which are printed in resin using a DLP printer.Keywords: sculptures, DLP 3D printer, microcasting, electroforming, fused deposition modeling
Procedia PDF Downloads 13523275 Online Monitoring and Control of Continuous Mechanosynthesis by UV-Vis Spectrophotometry
Authors: Darren A. Whitaker, Dan Palmer, Jens Wesholowski, James Flaherty, John Mack, Ahmad B. Albadarin, Gavin Walker
Abstract:
Traditional mechanosynthesis has been performed by either ball milling or manual grinding. However, neither of these techniques allow the easy application of process control. The temperature may change unpredictably due to friction in the process. Hence the amount of energy transferred to the reactants is intrinsically non-uniform. Recently, it has been shown that the use of Twin-Screw extrusion (TSE) can overcome these limitations. Additionally, TSE enables a platform for continuous synthesis or manufacturing as it is an open-ended process, with feedstocks at one end and product at the other. Several materials including metal-organic frameworks (MOFs), co-crystals and small organic molecules have been produced mechanochemically using TSE. The described advantages of TSE are offset by drawbacks such as increased process complexity (a large number of process parameters) and variation in feedstock flow impacting on product quality. To handle the above-mentioned drawbacks, this study utilizes UV-Vis spectrophotometry (InSpectroX, ColVisTec) as an online tool to gain real-time information about the quality of the product. Additionally, this is combined with real-time process information in an Advanced Process Control system (PharmaMV, Perceptive Engineering) allowing full supervision and control of the TSE process. Further, by characterizing the dynamic behavior of the TSE, a model predictive controller (MPC) can be employed to ensure the process remains under control when perturbed by external disturbances. Two reactions were studied; a Knoevenagel condensation reaction of barbituric acid and vanillin and, the direct amidation of hydroquinone by ammonium acetate to form N-Acetyl-para-aminophenol (APAP) commonly known as paracetamol. Both reactions could be carried out continuously using TSE, nuclear magnetic resonance (NMR) spectroscopy was used to confirm the percentage conversion of starting materials to product. This information was used to construct partial least squares (PLS) calibration models within the PharmaMV development system, which relates the percent conversion to product to the acquired UV-Vis spectrum. Once this was complete, the model was deployed within the PharmaMV Real-Time System to carry out automated optimization experiments to maximize the percentage conversion based on a set of process parameters in a design of experiments (DoE) style methodology. With the optimum set of process parameters established, a series of PRBS process response tests (i.e. Pseudo-Random Binary Sequences) around the optimum were conducted. The resultant dataset was used to build a statistical model and associated MPC. The controller maximizes product quality whilst ensuring the process remains at the optimum even as disturbances such as raw material variability are introduced into the system. To summarize, a combination of online spectral monitoring and advanced process control was used to develop a robust system for optimization and control of two TSE based mechanosynthetic processes.Keywords: continuous synthesis, pharmaceutical, spectroscopy, advanced process control
Procedia PDF Downloads 17723274 A Semi-supervised Classification Approach for Trend Following Investment Strategy
Authors: Rodrigo Arnaldo Scarpel
Abstract:
Trend following is a widely accepted investment strategy that adopts a rule-based trading mechanism that rather than striving to predict market direction or on information gathering to decide when to buy and when to sell a stock. Thus, in trend following one must respond to market’s movements that has recently happen and what is currently happening, rather than on what will happen. Optimally, in trend following strategy, is to catch a bull market at its early stage, ride the trend, and liquidate the position at the first evidence of the subsequent bear market. For applying the trend following strategy one needs to find the trend and identify trade signals. In order to avoid false signals, i.e., identify fluctuations of short, mid and long terms and to separate noise from real changes in the trend, most academic works rely on moving averages and other technical analysis indicators, such as the moving average convergence divergence (MACD) and the relative strength index (RSI) to uncover intelligible stock trading rules following trend following strategy philosophy. Recently, some works has applied machine learning techniques for trade rules discovery. In those works, the process of rule construction is based on evolutionary learning which aims to adapt the rules to the current environment and searches for the global optimum rules in the search space. In this work, instead of focusing on the usage of machine learning techniques for creating trading rules, a time series trend classification employing a semi-supervised approach was used to early identify both the beginning and the end of upward and downward trends. Such classification model can be employed to identify trade signals and the decision-making procedure is that if an up-trend (down-trend) is identified, a buy (sell) signal is generated. Semi-supervised learning is used for model training when only part of the data is labeled and Semi-supervised classification aims to train a classifier from both the labeled and unlabeled data, such that it is better than the supervised classifier trained only on the labeled data. For illustrating the proposed approach, it was employed daily trade information, including the open, high, low and closing values and volume from January 1, 2000 to December 31, 2022, of the São Paulo Exchange Composite index (IBOVESPA). Through this time period it was visually identified consistent changes in price, upwards or downwards, for assigning labels and leaving the rest of the days (when there is not a consistent change in price) unlabeled. For training the classification model, a pseudo-label semi-supervised learning strategy was used employing different technical analysis indicators. In this learning strategy, the core is to use unlabeled data to generate a pseudo-label for supervised training. For evaluating the achieved results, it was considered the annualized return and excess return, the Sortino and the Sharpe indicators. Through the evaluated time period, the obtained results were very consistent and can be considered promising for generating the intended trading signals.Keywords: evolutionary learning, semi-supervised classification, time series data, trading signals generation
Procedia PDF Downloads 8923273 Institutional Capacity and Corruption: Evidence from Brazil
Authors: Dalson Figueiredo, Enivaldo Rocha, Ranulfo Paranhos, José Alexandre
Abstract:
This paper analyzes the effects of institutional capacity on corruption. Methodologically, the research design combines both descriptive and multivariate statistics to examine two original datasets based on secondary data. In particular, we employ a principal component model to estimate an indicator of institutional capacity for both state audit institutions and subnational judiciary courts. Then, we estimate the effect of institutional capacity on two dependent variables: (1) incidence of administrative irregularities and (2) time elapsed to judge corruption cases. The preliminary results using ordinary least squares, negative binomial and Tobit models suggest the same conclusions: higher the institutional audit capacity, higher is the probability of detecting a corruption case. On the other hand, higher the institutional capacity of state judiciary, the lower is the time to judge corruption cases.Keywords: institutional capacity, corruption, state level institutions, evidence from Brazil
Procedia PDF Downloads 37223272 The Network Relative Model Accuracy (NeRMA) Score: A Method to Quantify the Accuracy of Prediction Models in a Concurrent External Validation
Authors: Carl van Walraven, Meltem Tuna
Abstract:
Background: Network meta-analysis (NMA) quantifies the relative efficacy of 3 or more interventions from studies containing a subgroup of interventions. This study applied the analytical approach of NMA to quantify the relative accuracy of prediction models with distinct inclusion criteria that are evaluated on a common population (‘concurrent external validation’). Methods: We simulated binary events in 5000 patients using a known risk function. We biased the risk function and modified its precision by pre-specified amounts to create 15 prediction models with varying accuracy and distinct patient applicability. Prediction model accuracy was measured using the Scaled Brier Score (SBS). Overall prediction model accuracy was measured using fixed-effects methods that accounted for model applicability patterns. Prediction model accuracy was summarized as the Network Relative Model Accuracy (NeRMA) Score which ranges from -∞ through 0 (accuracy of random guessing) to 1 (accuracy of most accurate model in concurrent external validation). Results: The unbiased prediction model had the highest SBS. The NeRMA score correctly ranked all simulated prediction models by the extent of bias from the known risk function. A SAS macro and R-function was created to implement the NeRMA Score. Conclusions: The NeRMA Score makes it possible to quantify the accuracy of binomial prediction models having distinct inclusion criteria in a concurrent external validation.Keywords: prediction model accuracy, scaled brier score, fixed effects methods, concurrent external validation
Procedia PDF Downloads 23623271 A Non-Invasive Blood Glucose Monitoring System Using near-Infrared Spectroscopy with Remote Data Logging
Authors: Bodhayan Nandi, Shubhajit Roy Chowdhury
Abstract:
This paper presents the development of a portable blood glucose monitoring device based on Near-Infrared Spectroscopy. The system supports Internet connectivity through WiFi and uploads the time series data of glucose concentration of patients to a server. In addition, the server is given sufficient intelligence to predict the future pathophysiological state of a patient given the current and past pathophysiological data. This will enable to prognosticate the approaching critical condition of the patient much before the critical condition actually occurs.The server hosts web applications to allow authorized users to monitor the data remotely.Keywords: non invasive, blood glucose concentration, microcontroller, IoT, application server, database server
Procedia PDF Downloads 22023270 Developing a Web GIS Tool for the Evaluation of Soil Erosion of a Watershed
Authors: Y. Fekir, K. Mederbal, M. A. Hamadouche, D. Anteur
Abstract:
The soil erosion by water has become one of the biggest problems of the environment in the world, threatening the majority of countries. There are several models to evaluate erosion. These models are still a simplified representation of reality. They permit the analysis of complex systems, measurements are complementary to allow an extrapolation in time and space and may combine different factors. The empirical model of soil loss proposed by Wischmeier and Smith (Universal Soil Loss Equation), is widely used in many countries. He considers that erosion is a multiplicative function of five factors: rainfall erosivity (the R factor) the soil erodibility factor (K), topography (LS), the erosion control practices (P) and vegetation cover and agricultural practices (C). In this work, we tried to develop a tool based on Web GIS functionality to evaluate soil losses caused by erosion taking into account five factors. This tool allows the user to integrate all the data needed for the evaluation (DEM, Land use, rainfall ...) in the form of digital layers to calculate the five factors taken into account in the USLE equation (R, K, C, P, LS). Accordingly, and after treatment of the integrated data set, a map of the soil losses will be achieved as a result. We tested the proposed tool on a watershed basin located in the weste of Algeria where a dataset was collected and prepared.Keywords: USLE, erosion, web gis, Algeria
Procedia PDF Downloads 33023269 Impact of Enhanced Business Models on Technology Companies in the Pandemic: A Case Study about the Revolutionary Change in Management Styles
Authors: Murat Colak, Berkay Cakir Saridogan
Abstract:
Since the dawn of modern corporations, almost every single employee has been working in the same loop, which contains three basic steps: going to work, providing the needs for the work, and getting back home. Only a small amount of people were able to break that standard and live outside the box. As the 2019 pandemic hit the Earth and most companies shut down their physical offices, that loop had to change for everyone. This means that the old management styles had to be significantly re-arranged to the "work from home" type of business methods. The methods include online conferences and meetings, time and task tracking using algorithms, globalization of the work, and, most importantly, remote working. After the global epidemic started, even the tech giants were concerned. Now, it can be seen those technology companies have an incredible step-up in their shares compared to the other companies because they know how to manage such situations even better than every other industry. This study aims to take the old traditional management styles in big companies and compare them with the post-covid methods (2019-2022). As a result of this comparison made using the annual reports and shared statistics, this study aims to explain why the winners of this crisis are the technology companies.Keywords: Covid-19, technology companies, business models, remote work
Procedia PDF Downloads 6423268 Modelling Kinetics of Colour Degradation in American Pokeweed (Phytolacca americana) Extract Concentration
Authors: Seyed-Ahmad Shahidi, Salemeh Kazemzadeh, Mehdi Sharifi Soltani, Azade Ghorbani-HasanSaraei
Abstract:
The kinetics of colour changes of American Pokeweed extract, due to concentration by various heating methods was studied. Three different heating/evaporation processes were employed for production of American Pokeweed extract concentrate. The American Pokeweed extract was concentrated to a final 40 °Brix from an initial °Brix of 4 by microwave heating, rotary vacuum evaporator and evaporating at atmospheric pressure. The final American Pokeweed extract concentration of 40 °Brix was achieved in 188, 216 and 320 min by using microwave, rotary vacuum and atmospheric heating processes, respectively. The colour change during concentration processes was investigated. Total colour differences, Hunter L, a and b parameters were used to estimate the extent of colour loss. All Hunter colour parameters decreased with time. The zero-order, first-order and a combined kinetics model were applied to the changes in colour parameters. All models were found to describe the L, a and b-data adequately. Results indicated that variation in TCD followed both first-order and combined kinetics models. This model implied that the colour formation and pigment destruction occurred during concentration processes of American Pokeweed extract.Keywords: American pokeweed, colour, concentration, kinetics
Procedia PDF Downloads 49823267 Time "And" Dimension(s) - Visualizing the 4th and 4+ Dimensions
Authors: Siddharth Rana
Abstract:
As we know so far, there are 3 dimensions that we are capable of interpreting and perceiving, and there is a 4th dimension, called time, about which we don’t know much yet. We, as humans, live in the 4th dimension, not the 3rd. We travel 3 dimensionally but cannot yet travel 4 dimensionally; perhaps if we could, then visiting the past and the future would be like climbing a mountain or going down a road. So far, we humans are not even capable of imagining any higher dimensions than the three dimensions in which we can travel. We are the beings of the 4th dimension; we are the beings of time; that is why we can travel 3 dimensionally; however, if, say, there were beings of the 5th dimension, then they would easily be able to travel 4 dimensionally, i.e., they could travel in the 4th dimension as well. Beings of the 5th dimension can easily time travel. However, beings of the 4th dimension, like us, cannot time travel because we live in a 4-D world, traveling 3 dimensionally. That means to ever do time travel, we just need to go to a higher dimension and not only perceive it but also be able to travel in it. However, traveling to the past is not very possible, unlike traveling to the future. Even if traveling to the past were possible, it would be very unlikely that an event in the past would be changed. In this paper, some approaches are provided to define time, our movement in time to the future, some aspects of time travel using dimensions, and how we can perceive a higher dimension.Keywords: time, dimensions, String theory, relativity
Procedia PDF Downloads 10723266 Investigating the Factors Affecting Generalization of Deep Learning Models for Plant Disease Detection
Authors: Praveen S. Muthukumarana, Achala C. Aponso
Abstract:
A large percentage of global crop harvest is lost due to crop diseases. Timely identification and treatment of crop diseases is difficult in many developing nations due to insufficient trained professionals in the field of agriculture. Many crop diseases can be accurately diagnosed by visual symptoms. In the past decade, deep learning has been successfully utilized in domains such as healthcare but adoption in agriculture for plant disease detection is rare. The literature shows that models trained with popular datasets such as PlantVillage does not generalize well on real world images. This paper attempts to find out how to make plant disease identification models that generalize well with real world images.Keywords: agriculture, convolutional neural network, deep learning, plant disease classification, plant disease detection, plant disease diagnosis
Procedia PDF Downloads 145