Search results for: co-authorship network analysis
27900 Model-Driven and Data-Driven Approaches for Crop Yield Prediction: Analysis and Comparison
Authors: Xiangtuo Chen, Paul-Henry Cournéde
Abstract:
Crop yield prediction is a paramount issue in agriculture. The main idea of this paper is to find out efficient way to predict the yield of corn based meteorological records. The prediction models used in this paper can be classified into model-driven approaches and data-driven approaches, according to the different modeling methodologies. The model-driven approaches are based on crop mechanistic modeling. They describe crop growth in interaction with their environment as dynamical systems. But the calibration process of the dynamic system comes up with much difficulty, because it turns out to be a multidimensional non-convex optimization problem. An original contribution of this paper is to propose a statistical methodology, Multi-Scenarios Parameters Estimation (MSPE), for the parametrization of potentially complex mechanistic models from a new type of datasets (climatic data, final yield in many situations). It is tested with CORNFLO, a crop model for maize growth. On the other hand, the data-driven approach for yield prediction is free of the complex biophysical process. But it has some strict requirements about the dataset. A second contribution of the paper is the comparison of these model-driven methods with classical data-driven methods. For this purpose, we consider two classes of regression methods, methods derived from linear regression (Ridge and Lasso Regression, Principal Components Regression or Partial Least Squares Regression) and machine learning methods (Random Forest, k-Nearest Neighbor, Artificial Neural Network and SVM regression). The dataset consists of 720 records of corn yield at county scale provided by the United States Department of Agriculture (USDA) and the associated climatic data. A 5-folds cross-validation process and two accuracy metrics: root mean square error of prediction(RMSEP), mean absolute error of prediction(MAEP) were used to evaluate the crop prediction capacity. The results show that among the data-driven approaches, Random Forest is the most robust and generally achieves the best prediction error (MAEP 4.27%). It also outperforms our model-driven approach (MAEP 6.11%). However, the method to calibrate the mechanistic model from dataset easy to access offers several side-perspectives. The mechanistic model can potentially help to underline the stresses suffered by the crop or to identify the biological parameters of interest for breeding purposes. For this reason, an interesting perspective is to combine these two types of approaches.Keywords: crop yield prediction, crop model, sensitivity analysis, paramater estimation, particle swarm optimization, random forest
Procedia PDF Downloads 23327899 The Impact of Blended Learning on Developing the students' Writing Skills and the Perception of Instructors and Students: Hawassa University in Focus
Authors: Mulu G. Gencha, Gebremedhin Simon, Menna Olango
Abstract:
This study was conducted at Hawassa University (HwU) in the Southern Nation Nationalities Peoples Regional State (SNNPRS) of Ethiopia. The prime concern of this study was to examine the writing performances of experimental and control group students, perception of experimental group students, and subject instructors. The course was blended learning (BL). Blended learning is a hybrid of classroom and on-line learning. Participants were eighty students from the School of Computer Science. Forty students attended the BL delivery involved using Face-to-Face (FTF) and campus-based online instruction. All instructors, fifty, of School of Language and Communication Studies along with 10 FGD members participated in the study. The experimental group went to the computer lab two times a week for four months, March-June, 2012, using the local area network (LAN), and software (MOODLE) writing program. On the other hand, the control group, forty students, took the FTF writing course five times a week for four months in similar academic calendar. The three instruments, the attitude questionnaire, tests and FGD were designed to identify views of students, instructors, and FGD participants on BL. At the end of the study, students’ final course scores were evaluated. Data were analyzed using independent samples t-tests. A statistically, significant difference was found between the FTF and BL (p<0.05). The analysis showed that the BL group was more successful than the conventional group. Besides, both instructors and students had positive attitude towards BL. The final section of the thesis showed the potential benefits and challenges, considering the pedagogical implications for the BL, and recommended possible avenues for further works.Keywords: blended learning, computer attitudes, computer usefulness, computer liking, computer confidence, computer phobia
Procedia PDF Downloads 41227898 Influence of Atmospheric Pollutants on Child Respiratory Disease in Cartagena De Indias, Colombia
Authors: Jose A. Alvarez Aldegunde, Adrian Fernandez Sanchez, Matthew D. Menden, Bernardo Vila Rodriguez
Abstract:
Up to five statistical pre-processings have been carried out considering the pollutant records of the stations present in Cartagena de Indias, Colombia, also taking into account the childhood asthma incidence surveys conducted in hospitals in the city by the Health Ministry of Colombia for this study. These pre-processings have consisted of different techniques such as the determination of the quality of data collection, determination of the quality of the registration network, identification and debugging of errors in data collection, completion of missing data and purified data, as well as the improvement of the time scale of records. The characterization of the quality of the data has been conducted by means of density analysis of the pollutant registration stations using ArcGis Software and through mass balance techniques, making it possible to determine inconsistencies in the records relating the registration data between stations following the linear regression. The results obtained in this process have highlighted the positive quality in the pollutant registration process. Consequently, debugging of errors has allowed us to identify certain data as statistically non-significant in the incidence and series of contamination. This data, together with certain missing records in the series recorded by the measuring stations, have been completed by statistical imputation equations. Following the application of these prior processes, the basic series of incidence data for respiratory disease and pollutant records have allowed the characterization of the influence of pollutants on respiratory diseases such as, for example, childhood asthma. This characterization has been carried out using statistical correlation methods, including visual correlation, simple linear regression correlation and spectral analysis with PAST Software which identifies maximum periodicity cycles and minimums under the formula of the Lomb periodgram. In relation to part of the results obtained, up to eleven maximums and minimums considered contemporary between the incidence records and the particles have been identified taking into account the visual comparison. The spectral analyses that have been performed on the incidence and the PM2.5 have returned a series of similar maximum periods in both registers, which are at a maximum during a period of one year and another every 25 days (0.9 and 0.07 years). The bivariate analysis has managed to characterize the variable "Daily Vehicular Flow" in the ninth position of importance of a total of 55 variables. However, the statistical correlation has not obtained a favorable result, having obtained a low value of the R2 coefficient. The series of analyses conducted has demonstrated the importance of the influence of pollutants such as PM2.5 in the development of childhood asthma in Cartagena. The quantification of the influence of the variables has been able to determine that there is a 56% probability of dependence between PM2.5 and childhood respiratory asthma in Cartagena. Considering this justification, the study could be completed through the application of the BenMap Software, throwing a series of spatial results of interpolated values of the pollutant contamination records that exceeded the established legal limits (represented by homogeneous units up to the neighborhood level) and results of the impact on the exacerbation of pediatric asthma. As a final result, an economic estimate (in Colombian Pesos) of the monthly and individual savings derived from the percentage reduction of the influence of pollutants in relation to visits to the Hospital Emergency Room due to asthma exacerbation in pediatric patients has been granted.Keywords: Asthma Incidence, BenMap, PM2.5, Statistical Analysis
Procedia PDF Downloads 11827897 PatchMix: Learning Transferable Semi-Supervised Representation by Predicting Patches
Authors: Arpit Rai
Abstract:
In this work, we propose PatchMix, a semi-supervised method for pre-training visual representations. PatchMix mixes patches of two images and then solves an auxiliary task of predicting the label of each patch in the mixed image. Our experiments on the CIFAR-10, 100 and the SVHN dataset show that the representations learned by this method encodes useful information for transfer to new tasks and outperform the baseline Residual Network encoders by on CIFAR 10 by 12% on ResNet 101 and 2% on ResNet-56, by 4% on CIFAR-100 on ResNet101 and by 6% on SVHN dataset on the ResNet-101 baseline model.Keywords: self-supervised learning, representation learning, computer vision, generalization
Procedia PDF Downloads 9027896 A Survey on Concurrency Control Methods in Distributed Database
Authors: Seyed Mohsen Jameii
Abstract:
In the last years, remarkable improvements have been made in the ability of distributed database systems performance. A distributed database is composed of some sites which are connected to each other through network connections. In this system, if good harmonization is not made between different transactions, it may result in database incoherence. Nowadays, because of the complexity of many sites and their connection methods, it is difficult to extend different models in distributed database serially. The principle goal of concurrency control in distributed database is to ensure not interfering in accessibility of common database by different sites. Different concurrency control algorithms have been suggested to use in distributed database systems. In this paper, some available methods have been introduced and compared for concurrency control in distributed database.Keywords: distributed database, two phase locking protocol, transaction, concurrency
Procedia PDF Downloads 35427895 Structural Model on Organizational Climate, Leadership Behavior and Organizational Commitment: Work Engagement of Private Secondary School Teachers in Davao City
Authors: Genevaive Melendres
Abstract:
School administrators face the reality of teachers losing their engagement, or schools losing the teachers. This study is then conducted to identify a structural model that best predict work engagement of private secondary teachers in Davao City. Ninety-three teachers from four sectarian schools and 56 teachers from four non-sectarian schools were involved in the completion of four survey instruments namely Organizational Climate Questionnaire, Leader Behavior Descriptive Questionnaire, Organizational Commitment Scales, and Utrecht Work Engagement Scales. Data were analyzed using frequency distribution, mean, standardized deviation, t-test for independent sample, Pearson r, stepwise multiple regression analysis, and structural equation modeling. Results show that schools have high level of organizational climate dimensions; leaders oftentimes show work-oriented and people-oriented behavior; teachers have high normative commitment and they are very often engaged at their work. Teachers from non-sectarian schools have higher organizational commitment than those from sectarian schools. Organizational climate and leadership behavior are positively related to and predict work engagement whereas commitment did not show any relationship. This study underscores the relative effects of three variables on the work engagement of teachers. After testing network of relationships and evaluating several models, a best-fitting model was found between leadership behavior and work engagement. The noteworthy findings suggest that principals pay attention and consistently evaluate their behavior for this best predicts the work engagement of the teachers. The study provides value to administrators who take decisions and create conditions in which teachers derive fulfillment.Keywords: leadership behavior, organizational climate, organizational commitment, private secondary school teachers, structural model on work engagement
Procedia PDF Downloads 27527894 Influence of Error Correction Codes on the Quality of Optical Broadband Connections
Authors: Mouna Hemdi, Jamel bel Hadj Tahar
Abstract:
The increasing development of multimedia applications requiring the simultaneous transport of several different services contributes to the evolution of the need for very high-speed network. In this paper, we propose an effective solution to achieve the very high speed while retaining elements of the optical transmission channel. So our study focuses on error correcting codes that aim for quality improvement on duty. We present a comparison of the quality of service for single channels and integrating the code BCH, RS and LDPC in order to find the best code in the different conditions of the transmission.Keywords: code error correction, high speed broadband, optical transmission, information systems security
Procedia PDF Downloads 39427893 Measurement and Analysis of Human Hand Kinematics
Authors: Tamara Grujic, Mirjana Bonkovic
Abstract:
Measurements and quantitative analysis of kinematic parameters of human hand movements have an important role in different areas such as hand function rehabilitation, modeling of multi-digits robotic hands, and the development of machine-man interfaces. In this paper the assessment and evaluation of the reach-to-grasp movement by using computerized and robot-assisted method is described. Experiment involved the measurements of hand positions of seven healthy subjects during grasping three objects of different shapes and sizes. Results showed that three dominant phases of reach-to-grasp movements could be clearly identified.Keywords: human hand, kinematics, measurement and analysis, reach-to-grasp movement
Procedia PDF Downloads 46527892 Underwater Image Enhancement and Reconstruction Using CNN and the MultiUNet Model
Authors: Snehal G. Teli, R. J. Shelke
Abstract:
CNN and MultiUNet models are the framework for the proposed method for enhancing and reconstructing underwater images. Multiscale merging of features and regeneration are both performed by the MultiUNet. CNN collects relevant features. Extensive tests on benchmark datasets show that the proposed strategy performs better than the latest methods. As a result of this work, underwater images can be represented and interpreted in a number of underwater applications with greater clarity. This strategy will advance underwater exploration and marine research by enhancing real-time underwater image processing systems, underwater robotic vision, and underwater surveillance.Keywords: convolutional neural network, image enhancement, machine learning, multiunet, underwater images
Procedia PDF Downloads 8027891 Comprehensive Profiling and Characterization of Untargeted Extracellular Metabolites in Fermentation Processes: Insights and Advances in Analysis and Identification
Authors: Marianna Ciaccia, Gennaro Agrimi, Isabella Pisano, Maurizio Bettiga, Silvia Rapacioli, Giulia Mensa, Monica Marzagalli
Abstract:
Objective: Untargeted metabolomic analysis of extracellular metabolites is a powerful approach that focuses on comprehensively profiling in the extracellular space. In this study, we applied extracellular metabolomic analysis to investigate the metabolism of two probiotic microorganisms with health benefits that extend far beyond the digestive tract and the immune system. Methods: Analytical techniques employed in extracellular metabolomic analysis encompass various technologies, including mass spectrometry (MS), which enables the identification of metabolites present in the fermentation media, as well as the comparison of metabolic profiles under different experimental conditions. Multivariate statistical analysis techniques like principal component analysis (PCA) or partial least squares-discriminant analysis (PLS-DA) play a crucial role in uncovering metabolic signatures and understanding the dynamics of metabolic networks. Results: Different types of supernatants from fermentation processes, such as dairy-free, not dairy-free media and media with no cells or pasteurized, were subjected to metabolite profiling, which contained a complex mixture of metabolites, including substrates, intermediates, and end-products. This profiling provided insights into the metabolic activity of the microorganisms. The integration of advanced software tools has facilitated the identification and characterization of metabolites in different fermentation conditions and microorganism strains. Conclusions: In conclusion, untargeted extracellular metabolomic analysis, combined with software tools, allowed the study of the metabolites consumed and produced during the fermentation processes of probiotic microorganisms. Ongoing advancements in data analysis methods will further enhance the application of extracellular metabolomic analysis in fermentation research, leading to improved bioproduction and the advancement of sustainable manufacturing processes.Keywords: biotechnology, metabolomics, lactic bacteria, probiotics, postbiotics
Procedia PDF Downloads 7227890 SiC Merged PiN and Schottky (MPS) Power Diodes Electrothermal Modeling in SPICE
Abstract:
This paper sets out a behavioral macro-model of a Merged PiN and Schottky (MPS) diode based on silicon carbide (SiC). This model holds good for both static and dynamic electrothermal simulations for industrial applications. Its parameters have been worked out from datasheets curves by drawing on the optimization method: Simulated Annealing (SA) for the SiC MPS diodes made available in the industry. The model also adopts the Analog Behavioral Model (ABM) of PSPICE in which it has been implemented. The thermal behavior of the devices was also taken into consideration by making use of Foster’ canonical network as figured out from electro-thermal measurement provided by the manufacturer of the device.Keywords: SiC MPS diode, electro-thermal, SPICE model, behavioral macro-model
Procedia PDF Downloads 40827889 Implementation of Invisible Digital Watermarking
Authors: V. Monisha, D. Sindhuja, M. Sowmiya
Abstract:
Over the decade, the applications about multimedia have been developed rapidly. The advancement in the communication field at the faster pace, it is necessary to protect the data during transmission. Thus, security of multimedia contents becomes a vital issue, and it is a need for protecting the digital content against malfunctions. Digital watermarking becomes the solution for the copyright protection and authentication of data in the network. In multimedia applications, embedded watermarks should be robust, and imperceptible. For improving robustness, the discrete wavelet transform is used. Both encoding and extraction algorithm can be done using MATLAB R2012a. In this Discrete wavelet transform (DWT) domain of digital image, watermarking algorithm is used, and hardware implementation can be done on Xilinx based FPGA.Keywords: digital watermarking, DWT, robustness, FPGA
Procedia PDF Downloads 41527888 Structural Reliability of Existing Structures: A Case Study
Authors: Z. Sakka, I. Assakkaf, T. Al-Yaqoub, J. Parol
Abstract:
A reliability-based methodology for the analysis assessment and evaluation of reinforced concrete structural elements of concrete structures is presented herein. The results of the reliability analysis and assessment for structural elements are verified by the results obtained from the deterministic methods. The analysis outcomes of reliability-based analysis are compared against the safety limits of the required reliability index β according to international standards and codes. The methodology is based on probabilistic analysis using reliability concepts and statistics of the main random variables that are relevant to the subject matter, and for which they are to be used in the performance-function equation(s) related to the structural elements under study. These methodology techniques can result in reliability index β, which is commonly known as the reliability index or reliability measure value that can be utilized to assess and evaluate the safety, human risk, and functionality of the structural component. Also, these methods can result in revised partial safety factor values for certain target reliability indices that can be used for the purpose of redesigning the reinforced concrete elements of the building and in which they could assist in considering some other remedial actions to improve the safety and functionality of the member.Keywords: structural reliability, concrete structures, FORM, Monte Carlo simulation
Procedia PDF Downloads 52027887 Hybrid Materials Obtained via Sol-Gel Way, by the Action of Teraethylorthosilicate with 1, 3, 4-Thiadiazole 2,5-Bifunctional Compounds
Authors: Afifa Hafidh, Fathi Touati, Ahmed Hichem Hamzaoui, Sayda Somrani
Abstract:
The objective of the present study has been to synthesize and to characterize silica hybrid materials using sol-gel technic and to investigate their properties. Silica materials were successfully fabricated using various bi-functional 1,3,4-thiadiazoles and tetraethoxysilane (TEOS) as co-precursors via a facile one-pot sol-gel pathway. TEOS was introduced at room temperature with 1,3,4-thiadiazole 2,5-difunctiunal adducts, in ethanol as solvent and using HCl acid as catalyst. The sol-gel process lead to the formation of monolithic, coloured and transparent gels. TEOS was used as a principal network forming agent. The incorporation of 1,3,4-thiadiazole molecules was realized by attachment of these later onto a silica matrix. This allowed covalent linkage between organic and inorganic phases and lead to the formation of Si-N and Si-S bonds. The prepared hybrid materials were characterized by Fourier transform infrared, NMR ²⁹Si and ¹³C, scanning electron microscopy and nitrogen absorption-desorption measurements. The optic and magnetic properties of hybrids are studied respectively by ultra violet-visible spectroscopy and electron paramagnetic resonance. It was shown in this work, that heterocyclic moieties were successfully attached in the hybrid skeleton. The formation of the Si-network composed of cyclic units (Q3 structures) connected by oxygen bridges (Q4 structures) was proved by ²⁹Si NMR spectroscopy. The Brunauer-Elmet-Teller nitrogen adsorption-desorption method shows that all the prepared xerogels have isotherms type IV and are mesoporous solids. The specific surface area and pore volume of these materials are important. The obtained results show that all materials are paramagnetic semiconductors. The data obtained by Nuclear magnetic resonance ²⁹Si and Fourier transform infrared spectroscopy, show that Si-OH and Si-NH groups existing in silica hybrids can participate in adsorption interactions. The obtained materials containing reactive centers could exhibit adsorption properties of metal ions due to the presence of OH and NH functionality in the mesoporous frame work. Our design of a simple method to prepare hybrid materials may give interest of the development of mesoporous hybrid systems and their use within the domain of environment in the future.Keywords: hybrid materials, sol-gel process, 1, 3, 4-thiadaizole, TEOS
Procedia PDF Downloads 18327886 Sentiment Analysis of Consumers’ Perceptions on Social Media about the Main Mobile Providers in Jamaica
Authors: Sherrene Bogle, Verlia Bogle, Tyrone Anderson
Abstract:
In recent years, organizations have become increasingly interested in the possibility of analyzing social media as a means of gaining meaningful feedback about their products and services. The aspect based sentiment analysis approach is used to predict the sentiment for Twitter datasets for Digicel and Lime, the main mobile companies in Jamaica, using supervised learning classification techniques. The results indicate an average of 82.2 percent accuracy in classifying tweets when comparing three separate classification algorithms against the purported baseline of 70 percent and an average root mean squared error of 0.31. These results indicate that the analysis of sentiment on social media in order to gain customer feedback can be a viable solution for mobile companies looking to improve business performance.Keywords: machine learning, sentiment analysis, social media, supervised learning
Procedia PDF Downloads 44727885 Detection of Image Blur and Its Restoration for Image Enhancement
Authors: M. V. Chidananda Murthy, M. Z. Kurian, H. S. Guruprasad
Abstract:
Image restoration in the process of communication is one of the emerging fields in the image processing. The motion analysis processing is the simplest case to detect motion in an image. Applications of motion analysis widely spread in many areas such as surveillance, remote sensing, film industry, navigation of autonomous vehicles, etc. The scene may contain multiple moving objects, by using motion analysis techniques the blur caused by the movement of the objects can be enhanced by filling-in occluded regions and reconstruction of transparent objects, and it also removes the motion blurring. This paper presents the design and comparison of various motion detection and enhancement filters. Median filter, Linear image deconvolution, Inverse filter, Pseudoinverse filter, Wiener filter, Lucy Richardson filter and Blind deconvolution filters are used to remove the blur. In this work, we have considered different types and different amount of blur for the analysis. Mean Square Error (MSE) and Peak Signal to Noise Ration (PSNR) are used to evaluate the performance of the filters. The designed system has been implemented in Matlab software and tested for synthetic and real-time images.Keywords: image enhancement, motion analysis, motion detection, motion estimation
Procedia PDF Downloads 29027884 Endothelial Dysfunction in Non-Alcoholic Fatty Liver Disease: An Updated Meta-Analysis
Authors: Anit S. Malhotra, Ajay Duseja, Neelam Chadha
Abstract:
Endothelial dysfunction is a precursor to atherosclerosis, and flow-mediated dilatation (FMD) in the brachial artery is the commonest method to evaluate endothelial function in humans. Non-alcoholic fatty liver disease (NAFLD) is one of the most common liver disorders encountered in clinical practice. An earlier meta-analysis had quantitatively assessed the degree of endothelial dysfunction using FMD. However, the largest study investigating the relation of FMD with NAFLD was published after that meta-analysis. In addition, that meta-analysis did not include some studies, including one from our centre. Therefore, an updating the previous meta-analysis was considered important. We searched PubMed, Cochrane Library, Embase, Scopus, SCI, Google Scholar, conference proceedings, and references of included studies till June 2017 to identify observational studies evaluating endothelial function using FMD in patients with non-alcoholic fatty liver disease. Data was analyzed using MedCalc. Fourteen studies were found eligible for inclusion in the meta-analysis. Patients with NAFLD had lower brachial artery FMD as compared to controls, standardized mean difference (random effects model) being –1.279%; 95% confidence interval (CI), –1.478 to –0.914. The effect size became smaller after addition of the recent study with the largest sample size was included compared with the earlier meta-analysis. In conclusion, patients with NAFLD had low FMD values indicating that they are at a higher risk of cardiovascular disease although our results suggest the effect size is not as large as reported previously.Keywords: endothelial dysfunction, flow-mediated dilatation, meta-analysis, non-alcoholic fatty liver disease
Procedia PDF Downloads 19027883 Throughput of Point Coordination Function (PCF)
Authors: Faisel Eltuhami Alzaalik, Omar Imhemed Alramli, Ahmed Mohamed Elaieb
Abstract:
The IEEE 802.11 defines two modes of MAC, distributed coordination function (DCF) and point coordination function (PCF) mode. The first sub-layer of the MAC is the distributed coordination function (DCF). A contention algorithm is used via DCF to provide access to all traffic. The point coordination function (PCF) is the second sub-layer used to provide contention-free service. PCF is upper DCF and it uses features of DCF to establish guarantee access of its users. Some papers and researches that have been published in this technology were reviewed in this paper, as well as talking briefly about the distributed coordination function (DCF) technology. The simulation of the PCF function have been applied by using a simulation program called network simulator (NS2) and have been found out the throughput of a transmitter system by using this function.Keywords: DCF, PCF, throughput, NS2
Procedia PDF Downloads 57927882 Exploring the Impact of Input Sequence Lengths on Long Short-Term Memory-Based Streamflow Prediction in Flashy Catchments
Authors: Farzad Hosseini Hossein Abadi, Cristina Prieto Sierra, Cesar Álvarez Díaz
Abstract:
Predicting streamflow accurately in flashy catchments prone to floods is a major research and operational challenge in hydrological modeling. Recent advancements in deep learning, particularly Long Short-Term Memory (LSTM) networks, have shown to be promising in achieving accurate hydrological predictions at daily and hourly time scales. In this work, a multi-timescale LSTM (MTS-LSTM) network was applied to the context of regional hydrological predictions at an hourly time scale in flashy catchments. The case study includes 40 catchments allocated in the Basque Country, north of Spain. We explore the impact of hyperparameters on the performance of streamflow predictions given by regional deep learning models through systematic hyperparameter tuning - where optimal regional values for different catchments are identified. The results show that predictions are highly accurate, with Nash-Sutcliffe (NSE) and Kling-Gupta (KGE) metrics values as high as 0.98 and 0.97, respectively. A principal component analysis reveals that a hyperparameter related to the length of the input sequence contributes most significantly to the prediction performance. The findings suggest that input sequence lengths have a crucial impact on the model prediction performance. Moreover, employing catchment-scale analysis reveals distinct sequence lengths for individual basins, highlighting the necessity of customizing this hyperparameter based on each catchment’s characteristics. This aligns with well known “uniqueness of the place” paradigm. In prior research, tuning the length of the input sequence of LSTMs has received limited focus in the field of streamflow prediction. Initially it was set to 365 days to capture a full annual water cycle. Later, performing limited systematic hyper-tuning using grid search, revealed a modification to 270 days. However, despite the significance of this hyperparameter in hydrological predictions, usually studies have overlooked its tuning and fixed it to 365 days. This study, employing a simultaneous systematic hyperparameter tuning approach, emphasizes the critical role of input sequence length as an influential hyperparameter in configuring LSTMs for regional streamflow prediction. Proper tuning of this hyperparameter is essential for achieving accurate hourly predictions using deep learning models.Keywords: LSTMs, streamflow, hyperparameters, hydrology
Procedia PDF Downloads 7427881 Adjusting Electricity Demand Data to Account for the Impact of Loadshedding in Forecasting Models
Authors: Migael van Zyl, Stefanie Visser, Awelani Phaswana
Abstract:
The electricity landscape in South Africa is characterized by frequent occurrences of loadshedding, a measure implemented by Eskom to manage electricity generation shortages by curtailing demand. Loadshedding, classified into stages ranging from 1 to 8 based on severity, involves the systematic rotation of power cuts across municipalities according to predefined schedules. However, this practice introduces distortions in recorded electricity demand, posing challenges to accurate forecasting essential for budgeting, network planning, and generation scheduling. Addressing this challenge requires the development of a methodology to quantify the impact of loadshedding and integrate it back into metered electricity demand data. Fortunately, comprehensive records of loadshedding impacts are maintained in a database, enabling the alignment of Loadshedding effects with hourly demand data. This adjustment ensures that forecasts accurately reflect true demand patterns, independent of loadshedding's influence, thereby enhancing the reliability of electricity supply management in South Africa. This paper presents a methodology for determining the hourly impact of load scheduling and subsequently adjusting historical demand data to account for it. Furthermore, two forecasting models are developed: one utilizing the original dataset and the other using the adjusted data. A comparative analysis is conducted to evaluate forecast accuracy improvements resulting from the adjustment process. By implementing this methodology, stakeholders can make more informed decisions regarding electricity infrastructure investments, resource allocation, and operational planning, contributing to the overall stability and efficiency of South Africa's electricity supply system.Keywords: electricity demand forecasting, load shedding, demand side management, data science
Procedia PDF Downloads 6327880 Public Values in Service Innovation Management: Case Study in Elderly Care in Danish Municipality
Authors: Christian T. Lystbaek
Abstract:
Background: The importance of innovation management has traditionally been ascribed to private production companies, however, there is an increasing interest in public services innovation management. One of the major theoretical challenges arising from this situation is to understand public values justifying public services innovation management. However, there is not single and stable definition of public value in the literature. The research question guiding this paper is: What is the supposed added value operating in the public sphere? Methodology: The study takes an action research strategy. This is highly contextualized methodology, which is enacted within a particular set of social relations into which on expects to integrate the results. As such, this research strategy is particularly well suited for its potential to generate results that can be applied by managers. The aim of action research is to produce proposals with a creative dimension capable of compelling actors to act in a new and pertinent way in relation to the situations they encounter. The context of the study is a workshop on public services innovation within elderly care. The workshop brought together different actors, such as managers, personnel and two groups of users-citizens (elderly clients and their relatives). The process was designed as an extension of the co-construction methods inherent in action research. Scenario methods and focus groups were applied to generate dialogue. The main strength of these techniques is to gather and exploit as much data as possible by exposing the discourse of justification used by the actors to explain or justify their points of view when interacting with others on a given subject. The approach does not directly interrogate the actors on their values, but allows their values to emerge through debate and dialogue. Findings: The public values related to public services innovation management in elderly care were identified in two steps. In the first step, identification of values, values were identified in the discussions. Through continuous analysis of the data, a network of interrelated values was developed. In the second step, tracking group consensus, we then ascertained the degree to which the meaning attributed to the value was common to the participants, classifying the degree of consensus as high, intermediate or low. High consensus corresponds to strong convergence in meaning, intermediate to generally shared meanings between participants, and low to divergences regarding the meaning between participants. Only values with high or intermediate degree of consensus were retained in the analysis. Conclusion: The study shows that the fundamental criterion for justifying public services innovation management is the capacity for actors to enact public values in their work. In the workshop, we identified two categories of public values, intrinsic value and behavioural values, and a list of more specific values.Keywords: public services innovation management, public value, co-creation, action research
Procedia PDF Downloads 28127879 Detection Efficient Enterprises via Data Envelopment Analysis
Authors: S. Turkan
Abstract:
In this paper, the Turkey’s Top 500 Industrial Enterprises data in 2014 were analyzed by data envelopment analysis. Data envelopment analysis is used to detect efficient decision-making units such as universities, hospitals, schools etc. by using inputs and outputs. The decision-making units in this study are enterprises. To detect efficient enterprises, some financial ratios are determined as inputs and outputs. For this reason, financial indicators related to productivity of enterprises are considered. The efficient foreign weighted owned capital enterprises are detected via super efficiency model. According to the results, it is said that Mercedes-Benz is the most efficient foreign weighted owned capital enterprise in Turkey.Keywords: data envelopment analysis, super efficiency, logistic regression, financial ratios
Procedia PDF Downloads 32827878 Neuron Dynamics of Single-Compartment Traub Model for Hardware Implementations
Authors: J. C. Moctezuma, V. Breña-Medina, Jose Luis Nunez-Yanez, Joseph P. McGeehan
Abstract:
In this work we make a bifurcation analysis for a single compartment representation of Traub model, one of the most important conductance-based models. The analysis focus in two principal parameters: current and leakage conductance. Study of stable and unstable solutions are explored; also Hop-bifurcation and frequency interpretation when current varies is examined. This study allows having control of neuron dynamics and neuron response when these parameters change. Analysis like this is particularly important for several applications such as: tuning parameters in learning process, neuron excitability tests, measure bursting properties of the neuron, etc. Finally, a hardware implementation results were developed to corroborate these results.Keywords: Traub model, Pinsky-Rinzel model, Hopf bifurcation, single-compartment models, bifurcation analysis, neuron modeling
Procedia PDF Downloads 32427877 Identifying Strategies for Improving Railway Services in Bangladesh
Authors: Armana Sabiha Huq, Tahmina Rahman Chowdhury
Abstract:
In this paper, based on the stated preference experiment, the service quality of Bangladesh Railway has been assessed, and particular importance has been given to investigate if there exists a relationship between service quality and safety. For investigation purposes, environmental and organizational factors were assumed to determine the safety performance of the railway. Data collected from the survey has been analyzed by importance-performance analysis (IPA). In this paper, a modification of the well-known importance-performance analysis (IPA) has been done by adopting the importance of the weights determined through a structural equation modeling (SEM) approach and by plotting the gap between importance and performance on a visual graph. It has been found that there exists a relationship between safety and serviceability to some extent. Limited resources are an important factor to improve the safety and serviceability condition of the BD railway. Moreover, it is observed that the limited resources available to monitor and improve the safety performance of railway.Keywords: importance-performance analysis, GAP-IPA, SEM, serviceability, safety, factor analysis
Procedia PDF Downloads 14427876 Blockchain Security in MANETs
Authors: Nada Mouchfiq, Ahmed Habbani, Chaimae Benjbara
Abstract:
The security aspect of the IoT occupies a place of great importance especially after the evolution that has known this field lastly because it must take into account the transformations and the new applications .Blockchain is a new technology dedicated to the data sharing. However, this does not work the same way in the different systems with different operating principles. This article will discuss network security using the Blockchain to facilitate the sending of messages and information, enabling the use of new processes and enabling autonomous coordination of devices. To do this, we will discuss proposed solutions to ensure a high level of security in these networks in the work of other researchers. Finally, our article will propose a method of security more adapted to our needs as a team working in the ad hoc networks, this method is based on the principle of the Blockchain and that we named ”MPR Blockchain”.Keywords: Ad hocs networks, blockchain, MPR, security
Procedia PDF Downloads 18727875 Comparative Analysis of Data Gathering Protocols with Multiple Mobile Elements for Wireless Sensor Network
Authors: Bhat Geetalaxmi Jairam, D. V. Ashoka
Abstract:
Wireless Sensor Networks are used in many applications to collect sensed data from different sources. Sensed data has to be delivered through sensors wireless interface using multi-hop communication towards the sink. The data collection in wireless sensor networks consumes energy. Energy consumption is the major constraints in WSN .Reducing the energy consumption while increasing the amount of generated data is a great challenge. In this paper, we have implemented two data gathering protocols with multiple mobile sinks/elements to collect data from sensor nodes. First, is Energy-Efficient Data Gathering with Tour Length-Constrained Mobile Elements in Wireless Sensor Networks (EEDG), in which mobile sinks uses vehicle routing protocol to collect data. Second is An Intelligent Agent-based Routing Structure for Mobile Sinks in WSNs (IAR), in which mobile sinks uses prim’s algorithm to collect data. Authors have implemented concepts which are common to both protocols like deployment of mobile sinks, generating visiting schedule, collecting data from the cluster member. Authors have compared the performance of both protocols by taking statistics based on performance parameters like Delay, Packet Drop, Packet Delivery Ratio, Energy Available, Control Overhead. Authors have concluded this paper by proving EEDG is more efficient than IAR protocol but with few limitations which include unaddressed issues likes Redundancy removal, Idle listening, Mobile Sink’s pause/wait state at the node. In future work, we plan to concentrate more on these limitations to avail a new energy efficient protocol which will help in improving the life time of the WSN.Keywords: aggregation, consumption, data gathering, efficiency
Procedia PDF Downloads 50027874 Reverse Logistics Information Management Using Ontological Approach
Authors: F. Lhafiane, A. Elbyed, M. Bouchoum
Abstract:
Reverse Logistics (RL) Process is considered as complex and dynamic network that involves many stakeholders such as: suppliers, manufactures, warehouse, retails, and costumers, this complexity is inherent in such process due to lack of perfect knowledge or conflicting information. Ontologies, on the other hand, can be considered as an approach to overcome the problem of sharing knowledge and communication among the various reverse logistics partners. In this paper, we propose a semantic representation based on hybrid architecture for building the Ontologies in an ascendant way, this method facilitates the semantic reconciliation between the heterogeneous information systems (ICT) that support reverse logistics Processes and product data.Keywords: Reverse Logistics, information management, heterogeneity, ontologies, semantic web
Procedia PDF Downloads 49427873 Arc Interruption Design for DC High Current/Low SC Fuses via Simulation
Authors: Ali Kadivar, Kaveh Niayesh
Abstract:
This report summarizes a simulation-based approach to estimate the current interruption behavior of a fuse element utilized in a DC network protecting battery banks under different stresses. Due to internal resistance of the battries, the short circuit current in very close to the nominal current, and it makes the fuse designation tricky. The base configuration considered in this report consists of five fuse units in parallel. The simulations are performed using a multi-physics software package, COMSOL® 5.6, and the necessary material parameters have been calculated using two other software packages.The first phase of the simulation starts with the heating of the fuse elements resulted from the current flow through the fusing element. In this phase, the heat transfer between the metallic strip and the adjacent materials results in melting and evaporation of the filler and housing before the aluminum strip is evaporated and the current flow in the evaporated strip is cut-off, or an arc is eventually initiated. The initiated arc starts to expand, so the entire metallic strip is ablated, and a long arc of around 20 mm is created within the first 3 milliseconds after arc initiation (v_elongation = 6.6 m/s. The final stage of the simulation is related to the arc simulation and its interaction with the external circuitry. Because of the strong ablation of the filler material and venting of the arc caused by the melting and evaporation of the filler and housing before an arc initiates, the arc is assumed to burn in almost pure ablated material. To be able to precisely model this arc, one more step related to the derivation of the transport coefficients of the plasma in ablated urethane was necessary. The results indicate that an arc current interruption, in this case, will not be achieved within the first tens of milliseconds. In a further study, considering two series elements, the arc was interrupted within few milliseconds. A very important aspect in this context is the potential impact of many broken strips parallel to the one where the arc occurs. The generated arcing voltage is also applied to the other broken strips connected in parallel with arcing path. As the gap between the other strips is very small, a large voltage of a few hundred volts generated during the current interruption may eventually lead to a breakdown of another gap. As two arcs in parallel are not stable, one of the arcs will extinguish, and the total current will be carried by one single arc again. This process may be repeated several times if the generated voltage is very large. The ultimate result would be that the current interruption may be delayed.Keywords: DC network, high current / low SC fuses, FEM simulation, paralle fuses
Procedia PDF Downloads 6927872 Mobile Cloud Computing: How to Improve
Authors: Abdullah Aljumah, Tariq Ahamad
Abstract:
The simplest possible human-computer interaction is mobile cloud computing as it emerges and makes the use of all modern-day human-oriented technology. The main aim of this idea is the QoS (quality of service) by using user-friendly and reliable software over the global network in order to make it economical by reducing cost, reliable, and increase the main storage. Since we studied and went through almost all the existing related work in this area and we came up with some challenges that will rise or might be rising for some basic areas in mobile cloud computing and mostly stogie and security area. In this research article, we suggest some recommendation for mobile cloud computing and for its security that will help in building more powerful tools to handle all this pressure.Keywords: Cloud Computing, MCC, SAAS, computer interaction
Procedia PDF Downloads 38327871 Nanocellulose Reinforced Biocomposites Based on Wheat Plasticized Starch for Food Packaging
Authors: Belen Montero, Carmen Ramirez, Maite Rico, Rebeca Bouza, Irene Derungs
Abstract:
Starch is a promising polymer for producing biocomposite materials because it is renewable, completely biodegradable and easily available at a low cost. Thermoplastic starches (TPS) can be obtained after the disruption and plasticization of native starch with a plasticizer. In this work, the solvent casting method was used to obtain TPS films from wheat starch plasticized with glycerol and reinforced with nanocellulose (CNC). X-ray diffraction analysis was used to follow the evolution of the crystallinity. The native wheat starch granules have shown a profile corresponding to A-type crystal structures typical for cereal starches. When TPS films are analyzed a high amorphous halo centered on 19º is obtained, indicating the plasticization process is completed. SEM imaging was made in order to analyse the morphology. The image from the raw wheat starch granules shows a bimodal granule size distribution with some granules in large round disk-shape forms (A-type) and the others as smaller spherical particles (B-type). The image from the neat TPS surface shows a continuous surface. No starch aggregates or swollen granules can be seen so, the plasticization process is complete. In the surfaces of reinforced TPS films aggregates are seen as the CNC concentration in the matrix increases. The CNC influence on the mechanical properties of TPS films has been studied by dynamic mechanical analysis. A direct relation exists between the storage modulus values, E’, and the CNC content in reinforced TPS films: higher is the content of nanocellulose in the composite, higher is the value of E’. This reinforcement effect can be explained by the appearance of a strong and crystalline nanoparticle-TPS interphase. Thermal stability of films was analysed by TGA. It has not observed any influence on the behaviour related to the thermal degradation of films with the incorporation of the CNC. Finally, the resistance to the water absorption films was analysed following the standard UNE-EN ISO 1998:483. The percentage of water absorbed by the samples at each time was calculated. The addition of 5 wt % of CNC to the TPS matrix leads to a significant improvement in the moisture resistance of the starch based material decreasing their diffusivity. It has been associated to the formation of a nanocrystal network that prevents swelling of the starch and therefore water absorption and to the high crystallinity of cellulose compared to starch. As a conclusion, the wheat film reinforced with 5 wt % of cellulose nanocrystals seems to be a good alternative for short-life applications into the packaging industry, because of its greatest rigidity, thermal stability and moisture sorption resistance.Keywords: biocomposites, nanocellulose, starch, wheat
Procedia PDF Downloads 213