Search results for: block model
17115 Cellular Automata Using Fractional Integral Model
Authors: Yasser F. Hassan
Abstract:
In this paper, a proposed model of cellular automata is studied by means of fractional integral function. A cellular automaton is a decentralized computing model providing an excellent platform for performing complex computation with the help of only local information. The paper discusses how using fractional integral function for representing cellular automata memory or state. The architecture of computing and learning model will be given and the results of calibrating of approach are also given.Keywords: fractional integral, cellular automata, memory, learning
Procedia PDF Downloads 41217114 A Strategic Communication Design Model for Indigenous Knowledge Management
Authors: Dilina Janadith Nawarathne
Abstract:
This article presents the initial development of a communication model (Model_isi) as the means of gathering, preserving and transferring indigenous knowledge in the field of knowledge management. The article first discusses the need for an appropriate complimentary model for indigenous knowledge management which differs from the existing methods and models. Then the paper suggests the newly developed model for indigenous knowledge management which generate as result of blending key aspects of different disciplines, which can be implemented as a complementary approach for the existing scientific method. The paper further presents the effectiveness of the developed method in reflecting upon a pilot demonstration carried out on selected indigenous communities of Sri Lanka.Keywords: indigenous knowledge management, knowledge transferring, tacit knowledge, research model, asian centric philosophy
Procedia PDF Downloads 48017113 Study on the Model Predicting Post-Construction Settlement of Soft Ground
Authors: Pingshan Chen, Zhiliang Dong
Abstract:
In order to estimate the post-construction settlement more objectively, the power-polynomial model is proposed, which can reflect the trend of settlement development based on the observed settlement data. It was demonstrated by an actual case history of an embankment, and during the prediction. Compared with the other three prediction models, the power-polynomial model can estimate the post-construction settlement more accurately with more simple calculation.Keywords: prediction, model, post-construction settlement, soft ground
Procedia PDF Downloads 42517112 Finding Data Envelopment Analysis Targets Using Multi-Objective Programming in DEA-R with Stochastic Data
Authors: R. Shamsi, F. Sharifi
Abstract:
In this paper, we obtain the projection of inefficient units in data envelopment analysis (DEA) in the case of stochastic inputs and outputs using the multi-objective programming (MOP) structure. In some problems, the inputs might be stochastic while the outputs are deterministic, and vice versa. In such cases, we propose a multi-objective DEA-R model because in some cases (e.g., when unnecessary and irrational weights by the BCC model reduce the efficiency score), an efficient decision-making unit (DMU) is introduced as inefficient by the BCC model, whereas the DMU is considered efficient by the DEA-R model. In some other cases, only the ratio of stochastic data may be available (e.g., the ratio of stochastic inputs to stochastic outputs). Thus, we provide a multi-objective DEA model without explicit outputs and prove that the input-oriented MOP DEA-R model in the invariable return to scale case can be replaced by the MOP-DEA model without explicit outputs in the variable return to scale and vice versa. Using the interactive methods for solving the proposed model yields a projection corresponding to the viewpoint of the DM and the analyst, which is nearer to reality and more practical. Finally, an application is provided.Keywords: DEA-R, multi-objective programming, stochastic data, data envelopment analysis
Procedia PDF Downloads 10517111 Potential of Dredged Material for CSEB in Building Structure
Authors: BoSheng Liu
Abstract:
The research goal is to re-image a locally-sourced waste product as abuilding material. The author aims to contribute to the compressed stabilized earth block (CSEB) by investigating the promising role of dredged material as an alternative building ingredient in the production of bricks and tiles. Dredged material comes from the sediment deposited near the shore or downstream, where the water current velocity decreases. This sediment needs to be dredged to provide water transportation; thus, there are mounds of the dredged material stored at bay. It is the interest of this research to reduce the filtered un-organic soil in the production of CSEB and replace it with locally dredged material from the Atchafalaya River in Morgan City, Louisiana. Technology and mechanical innovations have evolved the traditional adobe production method, which mixes the soil and natural fiber into molded bricks, into chemically stabilized CSEB made by compressing the clay mixture and stabilizer in a compression chamber with particular loads. In the case of dredged material CSEB (DM-CSEB), cement plays an essential role as the bending agent contributing to the unit strength while sustaining the filtered un-organic soil. Each DM-CSEB unit is made in a compression chamber with 580 PSI (i.e., 4 MPa) force. The research studied the cement content from 5% to 10% along with the range of dredged material mixtures, which differed from 20% to 80%. The material mixture content affected the DM-CSEB's strength and workability during and after its compression. Results indicated two optimal workabilities of the mixture: 27% fine clay content and 63% dredged material with 10% cement, or 28% fine clay content, and 67% dredged material with 5% cement. The final product of DM-CSEB emitted between 10 to 13 times fewer carbon emissions compared to the conventional fired masonry structure. DM-CSEB satisfied the strength requirement given by the ASTM C62 and ASTM C34 standards for construction material. One of the final evaluations tested and validated the material performance by designing and constructing an architectural, conical tile-vault prototype that was 28" by 40" by 24." The vault utilized a computational form-finding approach to generate the form's geometry, which optimized the correlation between the vault geometry and structural load distribution. A series of scaffolding was deployed to create the framework for the tile-vault construction. The final tile-vault structure was made from 2 layers of DM-CSEB tiles jointed by mortar, and the construction of the structure used over 110 tiles. The tile-vault prototype was capable of carrying over 400 lbs of live loads, which further demonstrated the dredged material feasibility as a construction material. The presented case study of Dredged Material Compressed Stabilized Earth Block (DM-CSEB) provides the first impression of dredged material in the clayey mixture process, structural performance, and construction practice. Overall, the approach of integrating dredged material in building material can be feasible, regionally sourced, cost-effective, and environment-friendly.Keywords: dredged material, compressed stabilized earth block, tile-vault, regionally sourced, environment-friendly
Procedia PDF Downloads 11517110 The Participation of Experts in the Criminal Policy on Drugs: The Proposal of a Cannabis Regulation Model in Spain by the Cannabis Policy Studies Group
Authors: Antonio Martín-Pardo
Abstract:
With regard to the context in which this paper is inserted, it is noteworthy that the current criminal policy model in which we find immersed, denominated by some doctrine sector as the citizen security model, is characterized by a marked tendency towards the discredit of expert knowledge. This type of technic knowledge has been displaced by the common sense and by the daily experience of the people at the time of legislative drafting, as well as by excessive attention to the short-term political effects of the law. Despite this criminal-political adverse scene, we still find valuable efforts in the side of experts to bring some rationality to the legislative development. This is the case of the proposal for a new cannabis regulation model in Spain carried out by the Cannabis Policy Studies Group (hereinafter referred as ‘GEPCA’). The GEPCA is a multidisciplinary group composed by authors with multiple/different orientations, trajectories and interests, but with a common minimum objective: the conviction that the current situation regarding cannabis is unsustainable and, that a rational legislative solution must be given to the growing social pressure for the regulation of their consumption and production. This paper details the main lines through which this technical proposal is developed with the purpose of its dissemination and discussion in the Congress. The basic methodology of the proposal is inductive-expository. In that way, firstly, we will offer a brief, but solid contextualization of the situation of cannabis in Spain. This contextualization will touch on issues such as the national regulatory situation and its relationship with the international context; the criminal, judicial and penitentiary impact of the offer and consumption of cannabis, or the therapeutic use of the substance, among others. In second place, we will get down to the business properly by detailing the minutia of the three main cannabis access channels that are proposed. Namely: the regulated market, the associations of cannabis users and personal self-cultivation. In each of these options, especially in the first two, special attention will be paid to both, the production and processing of the substance and the necessary administrative control of the activity. Finally, in a third block, some notes will be given on a series of subjects that surround the different access options just mentioned above and that give fullness and coherence to the proposal outlined. Among those related issues we find some such as consumption and tenure of the substance; the issue of advertising and promotion of cannabis; consumption in areas of special risk (work or driving v. g.); the tax regime; the need to articulate evaluation instruments for the entire process; etc. The main conclusion drawn from the analysis of the proposal is the unsustainability of the current repressive system, clearly unsuccessful, and the need to develop new access routes to cannabis that guarantee both public health and the rights of people who have freely chosen to consume it.Keywords: cannabis regulation proposal, cannabis policies studies group, criminal policy, expertise participation
Procedia PDF Downloads 11917109 Application of Generalized Autoregressive Score Model to Stock Returns
Authors: Katleho Daniel Makatjane, Diteboho Lawrence Xaba, Ntebogang Dinah Moroke
Abstract:
The current study investigates the behaviour of time-varying parameters that are based on the score function of the predictive model density at time t. The mechanism to update the parameters over time is the scaled score of the likelihood function. The results revealed that there is high persistence of time-varying, as the location parameter is higher and the skewness parameter implied the departure of scale parameter from the normality with the unconditional parameter as 1.5. The results also revealed that there is a perseverance of the leptokurtic behaviour in stock returns which implies the returns are heavily tailed. Prior to model estimation, the White Neural Network test exposed that the stock price can be modelled by a GAS model. Finally, we proposed further researches specifically to model the existence of time-varying parameters with a more detailed model that encounters the heavy tail distribution of the series and computes the risk measure associated with the returns.Keywords: generalized autoregressive score model, South Africa, stock returns, time-varying
Procedia PDF Downloads 50017108 Reduction of Rotor-Bearing-Support Finite Element Model through Substructuring
Authors: Abdur Rosyid, Mohamed El-Madany, Mohanad Alata
Abstract:
Due to simplicity and low cost, rotordynamic system is often modeled by using lumped parameters. Recently, finite elements have been used to model rotordynamic system as it offers higher accuracy. However, it involves high degrees of freedom. In some applications such as control design, this requires higher cost. For this reason, various model reduction methods have been proposed. This work demonstrates the quality of model reduction of rotor-bearing-support system through substructuring. The quality of the model reduction is evaluated by comparing some first natural frequencies, modal damping ratio, critical speeds and response of both the full system and the reduced system. The simulation shows that the substructuring is proven adequate to reduce finite element rotor model in the frequency range of interest as long as the numbers and the locations of master nodes are determined appropriately. However, the reduction is less accurate in an unstable or nearly-unstable system.Keywords: rotordynamic, finite element model, timoshenko beam, 3D solid elements, Guyan reduction method
Procedia PDF Downloads 27217107 A Unified Model for Orotidine Monophosphate Synthesis: Target for Inhibition of Growth of Mycobacterium tuberculosis
Authors: N. Naga Subrahmanyeswara Rao, Parag Arvind Deshpande
Abstract:
Understanding nucleotide synthesis reaction of any organism is beneficial to know the growth of it as in Mycobacterium tuberculosis to design anti TB drug. One of the reactions of de novo pathway which takes place in all organisms was considered. The reaction takes places between phosphoribosyl pyrophosphate and orotate catalyzed by orotate phosphoribosyl transferase and divalent metal ion gives orotdine monophosphate, a nucleotide. All the reaction steps of three experimentally proposed mechanisms for this reaction were considered to develop kinetic rate expression. The model was validated using the data for four organisms. This model could successfully describe the kinetics for the reported data. The developed model can serve as a reliable model to describe the kinetics in new organisms without the need of mechanistic determination. So an organism-independent model was developed.Keywords: mechanism, nucleotide, organism, tuberculosis
Procedia PDF Downloads 33417106 Controlling the Expense of Political Contests Using a Modified N-Players Tullock’s Model
Abstract:
This work introduces a generalization of the classical Tullock’s model of one-stage contests under complete information with multiple unlimited numbers of contestants. In classical Tullock’s model, the contest winner is not necessarily the highest bidder. Instead, the winner is determined according to a draw in which the winning probabilities are the relative contestants’ efforts. The Tullock modeling fits well political contests, in which the winner is not necessarily the highest effort contestant. This work presents a modified model which uses a simple non-discriminating rule, namely, a parameter to influence the total costs planned for an election, for example, the contest designer can control the contestants' efforts. The winner pays a fee, and the losers are reimbursed the same amount. Our proposed model includes a mechanism that controls the efforts exerted and balances competition, creating a tighter, less predictable and more interesting contest. Additionally, the proposed model follows the fairness criterion in the sense that it does not alter the contestants' probabilities of winning compared to the classic Tullock’s model. We provide an analytic solution for the contestant's optimal effort and expected reward.Keywords: contests, Tullock's model, political elections, control expenses
Procedia PDF Downloads 14517105 Fama French Four Factor Model: A Study of Nifty Fifty Companies
Authors: Deeksha Arora
Abstract:
The study aims to explore the applicability of the widely used asset pricing models, namely, Capital Asset Pricing Model (CAPM) and the Fama-French Four Factor Model in the Indian equity market. The study will be based on the companies that form part of the Nifty Fifty Index for a period of five years: 2011 to 2016. The asset pricing model is examined by forming portfolios on the basis of three variables – market capitalization (size effect), book-to-market equity ratio (value effect) and profitability. The study provides a basis to test the presence of the Fama-French Four factor model in Indian stock market. This study may provide a basis for future research in the generalized asset pricing model comprising of multiple risk factors.Keywords: book to market equity, Fama French four factor model, market capitalization, profitability, size effect, value effect
Procedia PDF Downloads 26317104 The Effectiveness of a Hybrid Diffie-Hellman-RSA-Advanced Encryption Standard Model
Authors: Abdellahi Cheikh
Abstract:
With the emergence of quantum computers with very powerful capabilities, the security of the exchange of shared keys between two interlocutors poses a big problem in terms of the rapid development of technologies such as computing power and computing speed. Therefore, the Diffie-Hellmann (DH) algorithm is more vulnerable than ever. No mechanism guarantees the security of the key exchange, so if an intermediary manages to intercept it, it is easy to intercept. In this regard, several studies have been conducted to improve the security of key exchange between two interlocutors, which has led to interesting results. The modification made on our model Diffie-Hellman-RSA-AES (DRA), which encrypts the information exchanged between two users using the three-encryption algorithms DH, RSA and AES, by using stenographic photos to hide the contents of the p, g and ClesAES values that are sent in an unencrypted state at the level of DRA model to calculate each user's public key. This work includes a comparative study between the DRA model and all existing solutions, as well as the modification made to this model, with an emphasis on the aspect of reliability in terms of security. This study presents a simulation to demonstrate the effectiveness of the modification made to the DRA model. The obtained results show that our model has a security advantage over the existing solution, so we made these changes to reinforce the security of the DRA model.Keywords: Diffie-Hellmann, DRA, RSA, advanced encryption standard
Procedia PDF Downloads 9317103 Project Management Agile Model Based on Project Management Body of Knowledge Guideline
Authors: Mehrzad Abdi Khalife, Iraj Mahdavi
Abstract:
This paper presents the agile model for project management process. For project management process, the Project Management Body of Knowledge (PMBOK) guideline has been selected as platform. Combination of computational science and artificial intelligent methodology has been added to the guideline to transfer the standard to agile project management process. The model is the combination of practical standard, computational science and artificial intelligent. In this model, we present communication model and protocols to keep process agile. Here, we illustrate the collaboration man and machine in project management area with artificial intelligent approach.Keywords: artificial intelligent, conceptual model, man-machine collaboration, project management, standard
Procedia PDF Downloads 34117102 Parameter Estimation for the Oral Minimal Model and Parameter Distinctions Between Obese and Non-obese Type 2 Diabetes
Authors: Manoja Rajalakshmi Aravindakshana, Devleena Ghosha, Chittaranjan Mandala, K. V. Venkateshb, Jit Sarkarc, Partha Chakrabartic, Sujay K. Maity
Abstract:
Oral Glucose Tolerance Test (OGTT) is the primary test used to diagnose type 2 diabetes mellitus (T2DM) in a clinical setting. Analysis of OGTT data using the Oral Minimal Model (OMM) along with the rate of appearance of ingested glucose (Ra) is performed to study differences in model parameters for control and T2DM groups. The differentiation of parameters of the model gives insight into the behaviour and physiology of T2DM. The model is also studied to find parameter differences among obese and non-obese T2DM subjects and the sensitive parameters were co-related to the known physiological findings. Sensitivity analysis is performed to understand changes in parameter values with model output and to support the findings, appropriate statistical tests are done. This seems to be the first preliminary application of the OMM with obesity as a distinguishing factor in understanding T2DM from estimated parameters of insulin-glucose model and relating the statistical differences in parameters to diabetes pathophysiology.Keywords: oral minimal model, OGTT, obese and non-obese T2DM, mathematical modeling, parameter estimation
Procedia PDF Downloads 9117101 Model Order Reduction Using Hybrid Genetic Algorithm and Simulated Annealing
Authors: Khaled Salah
Abstract:
Model order reduction has been one of the most challenging topics in the past years. In this paper, a hybrid solution of genetic algorithm (GA) and simulated annealing algorithm (SA) are used to approximate high-order transfer functions (TFs) to lower-order TFs. In this approach, hybrid algorithm is applied to model order reduction putting in consideration improving accuracy and preserving the properties of the original model which are two important issues for improving the performance of simulation and computation and maintaining the behavior of the original complex models being reduced. Compared to conventional mathematical methods that have been used to obtain a reduced order model of high order complex models, our proposed method provides better results in terms of reducing run-time. Thus, the proposed technique could be used in electronic design automation (EDA) tools.Keywords: genetic algorithm, simulated annealing, model reduction, transfer function
Procedia PDF Downloads 14317100 Increasing Sulfur Handling Cost Efficiency Using the Eco Sulfur Paving Block Method at PT Pertamina EP Field Cepu
Authors: Adha Bayu Wijaya, A. Zainal Abidin, Naufal Baihaqi, Joko Suprayitno, Astika Titistiti, Muslim Adi Wijaya, Endah Tri Lestari, Agung Wibowo
Abstract:
Sulfur is a non-metallic chemical element in the form of a yellow crystalline solid with the chemical formula, and is formed from several types of natural and artificial chemical reactions. Commercial applications of sulfur processed products can be found in various aspects of life, for example in the use of processed sulfur as paving blocks. The Gundih Central Processing Plant (CPP) is capable of producing 14 tons/day of sulfur pellets. This amount comes from the high H2S content of the wells with a total concentration of 20,000 ppm and a volume accumulation of 14 MMSCFD acid gas. H2S is converted to sulfur using the thiobacillus microbe in the Biological Sulfur Recovery Unit (BSRU) with a sulfur product purity level greater than 95%. In 2018 sulfur production at Gundih CPP was recorded at 4044 tons which could potentially trigger serious problems from an environmental aspect. The use of sulfur as material for making paving blocks is an alternative solution in addressing the potential impact on the environment, as regulated by Government Regulation No.22 of Year 2021 concerning the Waste Management of Non-Hazardous and Toxic Substances (B3), and the high cost of handling sulfur by third parties. The design mix of ratio sulfur paving blocks is 22% cements, rock ash 67%, and 11% of sulfur pellets. The sulfur used in making the paving mixture is pure sulfur, namely the side product category without any contaminants, thereby eliminating the potential for environmental pollution when implementing sulfur paving. Strength tests of sulfur paving materials have also been confirmed by external laboratories. The standard used in making sulfur paving blocks refers to the SNI 03-0691-1996 standard. With the results of sulfur paving blocks made according to quality B. Currently, sulfur paving blocks are used in building access to wells locations and in public roads in the Cepu Field area as a contribution from Corporate Social Responsibility (CSR).Keywords: sulphur, innovation, paving block, CSR, sulphur paving
Procedia PDF Downloads 7517099 Towards an Enhanced Compartmental Model for Profiling Malware Dynamics
Authors: Jessemyn Modiini, Timothy Lynar, Elena Sitnikova
Abstract:
We present a novel enhanced compartmental model for malware spread analysis in cyber security. This paper applies cyber security data features to epidemiological compartmental models to model the infectious potential of malware. Compartmental models are most efficient for calculating the infectious potential of a disease. In this paper, we discuss and profile epidemiologically relevant data features from a Domain Name System (DNS) dataset. We then apply these features to epidemiological compartmental models to network traffic features. This paper demonstrates how epidemiological principles can be applied to the novel analysis of key cybersecurity behaviours and trends and provides insight into threat modelling above that of kill-chain analysis. In applying deterministic compartmental models to a cyber security use case, the authors analyse the deficiencies and provide an enhanced stochastic model for cyber epidemiology. This enhanced compartmental model (SUEICRN model) is contrasted with the traditional SEIR model to demonstrate its efficacy.Keywords: cybersecurity, epidemiology, cyber epidemiology, malware
Procedia PDF Downloads 10717098 Hidden Oscillations in the Mathematical Model of the Optical Binary Phase Shift Keying (BPSK) Costas Loop
Authors: N. V. Kuznetsov, O. A. Kuznetsova, G. A. Leonov, M. V. Yuldashev, R. V. Yuldashev
Abstract:
Nonlinear analysis of the phase locked loop (PLL)-based circuits is a challenging task. Thus, the simulation is widely used for their study. In this work, we consider a mathematical model of the optical Costas loop and demonstrate the limitations of simulation approach related to the existence of so-called hidden oscillations in the phase space of the model.Keywords: optical Costas loop, mathematical model, simulation, hidden oscillation
Procedia PDF Downloads 44017097 Reference Model for the Implementation of an E-Commerce Solution in Peruvian SMEs in the Retail Sector
Authors: Julio Kauss, Miguel Cadillo, David Mauricio
Abstract:
E-commerce is a business model that allows companies to optimize the processes of buying, selling, transferring goods and exchanging services through computer networks or the Internet. In Peru, the electronic commerce is used infrequently. This situation is due, in part to the fact that there is no model that allows companies to implement an e-commerce solution, which means that most SMEs do not have adequate knowledge to adapt to electronic commerce. In this work, a reference model is proposed for the implementation of an e-commerce solution in Peruvian SMEs in the retail sector. It consists of five phases: Business Analysis, Business Modeling, Implementation, Post Implementation and Results. The present model was validated in a SME of the Peruvian retail sector through the implementation of an electronic commerce platform, through which the company increased its sales through the delivery channel by 10% in the first month of deployment. This result showed that the model is easy to implement, is economical and agile. In addition, it allowed the company to increase its business offer, adapt to e-commerce and improve customer loyalty.Keywords: e-commerce, retail, SMEs, reference model
Procedia PDF Downloads 32017096 Kinetic Façade Design Using 3D Scanning to Convert Physical Models into Digital Models
Authors: Do-Jin Jang, Sung-Ah Kim
Abstract:
In designing a kinetic façade, it is hard for the designer to make digital models due to its complex geometry with motion. This paper aims to present a methodology of converting a point cloud of a physical model into a single digital model with a certain topology and motion. The method uses a Microsoft Kinect sensor, and color markers were defined and applied to three paper folding-inspired designs. Although the resulted digital model cannot represent the whole folding range of the physical model, the method supports the designer to conduct a performance-oriented design process with the rough physical model in the reduced folding range.Keywords: design media, kinetic facades, tangible user interface, 3D scanning
Procedia PDF Downloads 41317095 A Large Language Model-Driven Method for Automated Building Energy Model Generation
Authors: Yake Zhang, Peng Xu
Abstract:
The development of building energy models (BEM) required for architectural design and analysis is a time-consuming and complex process, demanding a deep understanding and proficient use of simulation software. To streamline the generation of complex building energy models, this study proposes an automated method for generating building energy models using a large language model and the BEM library aimed at improving the efficiency of model generation. This method leverages a large language model to parse user-specified requirements for target building models, extracting key features such as building location, window-to-wall ratio, and thermal performance of the building envelope. The BEM library is utilized to retrieve energy models that match the target building’s characteristics, serving as reference information for the large language model to enhance the accuracy and relevance of the generated model, allowing for the creation of a building energy model that adapts to the user’s modeling requirements. This study enables the automatic creation of building energy models based on natural language inputs, reducing the professional expertise required for model development while significantly decreasing the time and complexity of manual configuration. In summary, this study provides an efficient and intelligent solution for building energy analysis and simulation, demonstrating the potential of a large language model in the field of building simulation and performance modeling.Keywords: artificial intelligence, building energy modelling, building simulation, large language model
Procedia PDF Downloads 2517094 An Improved Model of Estimation Global Solar Irradiation from in situ Data: Case of Oran Algeria Region
Authors: Houcine Naim, Abdelatif Hassini, Noureddine Benabadji, Alex Van Den Bossche
Abstract:
In this paper, two models to estimate the overall monthly average daily radiation on a horizontal surface were applied to the site of Oran (35.38 ° N, 0.37 °W). We present a comparison between the first one is a regression equation of the Angstrom type and the second model is developed by the present authors some modifications were suggested using as input parameters: the astronomical parameters as (latitude, longitude, and altitude) and meteorological parameters as (relative humidity). The comparisons are made using the mean bias error (MBE), root mean square error (RMSE), mean percentage error (MPE), and mean absolute bias error (MABE). This comparison shows that the second model is closer to the experimental values that the model of Angstrom.Keywords: meteorology, global radiation, Angstrom model, Oran
Procedia PDF Downloads 23117093 Expert Review on Conceptual Design Model of Assistive Courseware for Low Vision (AC4LV) Learners
Authors: Nurulnadwan Aziz, Ariffin Abdul Mutalib, Siti Mahfuzah Sarif
Abstract:
This paper reports an ongoing project regarding the development of Conceptual Design Model of Assistive Courseware for Low Vision (AC4LV) learners. Having developed the intended model, it has to be validated prior to producing it as guidance for the developers to develop an AC4LV. This study requires two phases of validation process which are through expert review and prototyping method. This paper presents a part of the validation process which is findings from experts review on Conceptual Design Model of AC4LV which has been carried out through a questionnaire. Results from 12 international and local experts from various respectable fields in Human-Computer Interaction (HCI) were discussed and justified. In a nutshell, reviewed Conceptual Design Model of AC4LV was formed. Future works of this study are to validate the reviewed model through prototyping method prior to testing it to the targeted users.Keywords: assistive courseware, conceptual design model, expert review, low vision learners
Procedia PDF Downloads 54617092 Life Cycle Assessment of Biogas Energy Production from a Small-Scale Wastewater Treatment Plant in Central Mexico
Authors: Joel Bonales, Venecia Solorzano, Carlos Garcia
Abstract:
A great percentage of the wastewater generated in developing countries don’t receive any treatment, which leads to numerous environmental impacts. In response to this, a paradigm change in the current wastewater treatment model based on large scale plants towards a small and medium scale based model has been proposed. Nevertheless, small scale wastewater treatment (SS-WTTP) with novel technologies such as anaerobic digesters, as well as the utilization of derivative co-products such as biogas, still presents diverse environmental impacts which must be assessed. This study consisted in a Life Cycle Assessment (LCA) performed to a SS-WWTP which treats wastewater from a small commercial block in the city of Morelia, Mexico. The treatment performed in the SS-WWTP consists in anaerobic and aerobic digesters with a daily capacity of 5,040 L. Two different scenarios were analyzed: the current plant conditions and a hypothetical energy use of biogas obtained in situ. Furthermore, two different allocation criteria were applied: full impact allocation to the system’s main product (treated water) and substitution credits for replacing Mexican grid electricity (biogas) and clean water pumping (treated water). The results showed that the analyzed plant had bigger impacts than what has been reported in the bibliography in the basis of wastewater volume treated, which may imply that this plant is currently operating inefficiently. The evaluated impacts appeared to be focused in the aerobic digestion and electric generation phases due to the plant’s particular configuration. Additional findings prove that the allocation criteria applied is crucial for the interpretation of impacts and that that the energy use of the biogas obtained in this plant can help mitigate associated climate change impacts. It is concluded that SS-WTTP is a environmentally sound alternative for wastewater treatment from a systemic perspective. However, this type of studies must be careful in the selection of the allocation criteria and replaced products, since these factors have a great influence in the results of the assessment.Keywords: biogas, life cycle assessment, small scale treatment, wastewater treatment
Procedia PDF Downloads 12417091 Programmable Shields in Space
Authors: Tapas Kumar Sinha, Joseph Mathew
Abstract:
At the moment earth is in grave danger due to threats of global warming. The temperature of the earth has risen by almost 20C. Glaciers in the Arctic have started to melt. It would be foolhardy to think that this is a small effect and in time it would go away. Global warming is caused by a number of factors. However, one sure and simple way to totally eliminate this problem is to put programmable shields in space. Just as an umbrella blocks sunlight, a programmable shield in space will block sun rays from reaching the earth as in a solar eclipse and cause cooling in the penumbral region just as it happens during an eclipse.Keywords: glaciers, green house, global warming space, satellites
Procedia PDF Downloads 59917090 Application of Model Tree in the Prediction of TBM Rate of Penetration with Synthetic Minority Oversampling Technique
Authors: Ehsan Mehryaar
Abstract:
The rate of penetration is (RoP) one of the vital factors in the cost and time of tunnel boring projects; therefore, predicting it can lead to a substantial increase in the efficiency of the project. RoP is heavily dependent geological properties of the project site and TBM properties. In this study, 151-point data from Queen’s water tunnel is collected, which includes unconfined compression strength, peak slope index, angle with weak planes, and distance between planes of weaknesses. Since the size of the data is small, it was observed that it is imbalanced. To solve that problem synthetic minority oversampling technique is utilized. The model based on the model tree is proposed, where each leaf consists of a support vector machine model. Proposed model performance is then compared to existing empirical equations in the literature.Keywords: Model tree, SMOTE, rate of penetration, TBM(tunnel boring machine), SVM
Procedia PDF Downloads 17417089 The Effect of Post Spinal Hypotension on Cerebral Oxygenation Using Near-Infrared Spectroscopy and Neonatal Outcomes in Full Term Parturient Undergoing Lower Segment Caesarean Section: A Prospective Observational Study
Authors: Shailendra Kumar, Lokesh Kashyap, Puneet Khanna, Nishant Patel, Rakesh Kumar, Arshad Ayub, Kelika Prakash, Yudhyavir Singh, Krithikabrindha V.
Abstract:
Introduction: Spinal anesthesia is considered a standard anesthesia technique for caesarean delivery. The incidence of spinal hypotension during caesarean delivery is 70 -80%. Spinal hypotension may cause cerebral hypoperfusion in the mother, but physiologically cerebral autoregulatory mechanisms accordingly prevent cerebral hypoxia. Cerebral blood flow remains constant in the 50-150 mmHg of Cerebral Perfusion Pressure (CPP) range. Near-infrared spectroscopy (NIRS) is a non-invasive technology that is used to detect Cerebral Desaturation Events (CDEs) immediately compared to other conventional intraoperative monitoring techniques. Objective: The primary aim of the study is to correlate the change in cerebral oxygen saturation using NIRS with respect to a fall in mean blood pressure after spinal anaesthesia and to find out the effects of spinal hypotension on neonatal APGAR score, neonatal acid-base variations, and presence of Postoperative Delirium (POD). Methodology: NIRS sensors were attached to the forehead of all the patients, and their baseline readings of cerebral oxygenation on the right and left frontal regions and mean blood pressure were noted. Subarachnoid block was given with hyperbaric 0.5% bupivacaine plus fentanyl, the dose being determined by the individual anaesthesiologist. Co-loading of IV crystalloid solutions was given to the patient. Blood pressure reading and cerebral saturation were recorded every 1 minute till 30min. Hypotension was a fall in MAP less than 20% of the baseline values. Patients going for hypotension were treated with an IV Bolus of phenylephrine/ephedrine. Umbilical cord blood samples were taken for blood gas analysis, and neonatal APGAR was noted by a neonatologist. Study design: A prospective observational study conducted in a population of Thirty ASA 2 and 3 parturients scheduled for lower segment caesarean section (LSCS). Results: Mean fall in regional cerebral saturation is 28.48 ± 14.7% with respect to the mean fall in blood pressure 38.92 ± 8.44 mm Hg. The correlation coefficient between fall in saturation and fall in mean blood pressure is 0.057, and p-value {0.7} after subarachnoid block. A fall in regional cerebral saturation occurred 2±1 min before a fall in mean blood pressure. Twenty-nine out of thirty patients required vasopressors during hypotension. The first dose of vasopressor requirement is needed at 6.02±2 min after the block. The mean APGAR score was 7.86 and 9.74 at 1 and 5 min of birth, respectively, and the mean umbilical arterial pH of 7.3±0.1. According to DRS-98 (Delirium Rating Scale), the mean delirium rating score on postoperative day 1 and day 2 were 0.1 and 0.7, respectively. Discussion: There was a fall in regional cerebral oxygen saturation, which started before with respect to a significant fall in mean blood pressure readings but was statistically not significant. Maximal fall in blood pressure requiring vasopressors occurs within 10 min of SAB. Neonatal APGAR scores and acid-base variations were in the normal range with maternal hypotension, and there was no incidence of postoperative delirium in patients with post-spinal hypotension.Keywords: cerebral oxygenation, LSCS, NIRS, spinal hypotension
Procedia PDF Downloads 6917088 An Agent-Based Modeling and Simulation of Human Muscle
Authors: Sina Saadati, Mohammadreza Razzazi
Abstract:
In this article, we have tried to present an agent-based model of human muscle. A suitable model of muscle is necessary for the analysis of mankind's movements. It can be used by clinical researchers who study the influence of motion sicknesses, like Parkinson's disease. It is also useful in the development of a prosthesis that receives the electromyography signals and generates force as a reaction. Since we have focused on computational efficiency in this research, the model can compute the calculations very fast. As far as it concerns prostheses, the model can be known as a charge-efficient method. In this paper, we are about to illustrate an agent-based model. Then, we will use it to simulate the human gait cycle. This method can also be done reversely in the analysis of gait in motion sicknesses.Keywords: agent-based modeling and simulation, human muscle, gait cycle, motion sickness
Procedia PDF Downloads 11417087 Parameter Estimation of Gumbel Distribution with Maximum-Likelihood Based on Broyden Fletcher Goldfarb Shanno Quasi-Newton
Authors: Dewi Retno Sari Saputro, Purnami Widyaningsih, Hendrika Handayani
Abstract:
Extreme data on an observation can occur due to unusual circumstances in the observation. The data can provide important information that can’t be provided by other data so that its existence needs to be further investigated. The method for obtaining extreme data is one of them using maxima block method. The distribution of extreme data sets taken with the maxima block method is called the distribution of extreme values. Distribution of extreme values is Gumbel distribution with two parameters. The parameter estimation of Gumbel distribution with maximum likelihood method (ML) is difficult to determine its exact value so that it is necessary to solve the approach. The purpose of this study was to determine the parameter estimation of Gumbel distribution with quasi-Newton BFGS method. The quasi-Newton BFGS method is a numerical method used for nonlinear function optimization without constraint so that the method can be used for parameter estimation from Gumbel distribution whose distribution function is in the form of exponential doubel function. The quasi-New BFGS method is a development of the Newton method. The Newton method uses the second derivative to calculate the parameter value changes on each iteration. Newton's method is then modified with the addition of a step length to provide a guarantee of convergence when the second derivative requires complex calculations. In the quasi-Newton BFGS method, Newton's method is modified by updating both derivatives on each iteration. The parameter estimation of the Gumbel distribution by a numerical approach using the quasi-Newton BFGS method is done by calculating the parameter values that make the distribution function maximum. In this method, we need gradient vector and hessian matrix. This research is a theory research and application by studying several journals and textbooks. The results of this study obtained the quasi-Newton BFGS algorithm and estimation of Gumbel distribution parameters. The estimation method is then applied to daily rainfall data in Purworejo District to estimate the distribution parameters. This indicates that the high rainfall that occurred in Purworejo District decreased its intensity and the range of rainfall that occurred decreased.Keywords: parameter estimation, Gumbel distribution, maximum likelihood, broyden fletcher goldfarb shanno (BFGS)quasi newton
Procedia PDF Downloads 32317086 Multinomial Dirichlet Gaussian Process Model for Classification of Multidimensional Data
Authors: Wanhyun Cho, Soonja Kang, Sanggoon Kim, Soonyoung Park
Abstract:
We present probabilistic multinomial Dirichlet classification model for multidimensional data and Gaussian process priors. Here, we have considered an efficient computational method that can be used to obtain the approximate posteriors for latent variables and parameters needed to define the multiclass Gaussian process classification model. We first investigated the process of inducing a posterior distribution for various parameters and latent function by using the variational Bayesian approximations and important sampling method, and next we derived a predictive distribution of latent function needed to classify new samples. The proposed model is applied to classify the synthetic multivariate dataset in order to verify the performance of our model. Experiment result shows that our model is more accurate than the other approximation methods.Keywords: multinomial dirichlet classification model, Gaussian process priors, variational Bayesian approximation, importance sampling, approximate posterior distribution, marginal likelihood evidence
Procedia PDF Downloads 443