Search results for: vector calculus
447 Forecasting the Future Implications of ChatGPT Usage in Education Based on AI Algorithms
Authors: Yakubu Bala Mohammed, Nadire Chavus, Mohammed Bulama
Abstract:
Generative Pre-trained Transformer (ChatGPT) represents an artificial intelligence (AI) tool capable of swiftly generating comprehensive responses to prompts and follow-up inquiries. This emerging AI tool was introduced in November 2022 by OpenAI firm, an American AI research laboratory, utilizing substantial language models. This present study aims to delve into the potential future consequences of ChatGPT usage in education using AI-based algorithms. The paper will bring forth the likely potential risks of ChatGBT utilization, such as academic integrity concerns, unfair learning assessments, excessive reliance on AI, and dissemination of inaccurate information using four machine learning algorithms: eXtreme-Gradient Boosting (XGBoost), Support vector machine (SVM), Emotional artificial neural network (EANN), and Random forest (RF) would be used to analyze the study collected data due to their robustness. Finally, the findings of the study will assist education stakeholders in understanding the future implications of ChatGPT usage in education and propose solutions and directions for upcoming studies.Keywords: machine learning, ChatGPT, education, learning, implications
Procedia PDF Downloads 232446 Modified Form of Margin Based Angular Softmax Loss for Speaker Verification
Authors: Jamshaid ul Rahman, Akhter Ali, Adnan Manzoor
Abstract:
Learning-based systems have received increasing interest in recent years; recognition structures, including end-to-end speak recognition, are one of the hot topics in this area. A famous work on end-to-end speaker verification by using Angular Softmax Loss gained significant importance and is considered useful to directly trains a discriminative model instead of the traditional adopted i-vector approach. The margin-based strategy in angular softmax is beneficial to learn discriminative speaker embeddings where the random selection of margin values is a big issue in additive angular margin and multiplicative angular margin. As a better solution in this matter, we present an alternative approach by introducing a bit similar form of an additive parameter that was originally introduced for face recognition, and it has a capacity to adjust automatically with the corresponding margin values and is applicable to learn more discriminative features than the Softmax. Experiments are conducted on the part of Fisher dataset, where it observed that the additive parameter with angular softmax to train the front-end and probabilistic linear discriminant analysis (PLDA) in the back-end boosts the performance of the structure.Keywords: additive parameter, angular softmax, speaker verification, PLDA
Procedia PDF Downloads 103445 Data Mining Approach for Commercial Data Classification and Migration in Hybrid Storage Systems
Authors: Mais Haj Qasem, Maen M. Al Assaf, Ali Rodan
Abstract:
Parallel hybrid storage systems consist of a hierarchy of different storage devices that vary in terms of data reading speed performance. As we ascend in the hierarchy, data reading speed becomes faster. Thus, migrating the application’ important data that will be accessed in the near future to the uppermost level will reduce the application I/O waiting time; hence, reducing its execution elapsed time. In this research, we implement trace-driven two-levels parallel hybrid storage system prototype that consists of HDDs and SSDs. The prototype uses data mining techniques to classify application’ data in order to determine its near future data accesses in parallel with the its on-demand request. The important data (i.e. the data that the application will access in the near future) are continuously migrated to the uppermost level of the hierarchy. Our simulation results show that our data migration approach integrated with data mining techniques reduces the application execution elapsed time when using variety of traces in at least to 22%.Keywords: hybrid storage system, data mining, recurrent neural network, support vector machine
Procedia PDF Downloads 308444 Biosynthesis of L-Xylose from Xylitol Using a Dual Enzyme Cascade in Escherichia coli
Authors: Mesfin Angaw Tesfay
Abstract:
L-xylose is an important intermediate in the pharmaceutical industry, playing a key role in the production of various antiviral and anticancer drugs. Despite its significance, L-xylose is a rare and costly sugar with limited availability in nature. In recent years, enzymatic production methods have garnered considerable attention due to their benefits over conventional chemical synthesis. In this research, a dual enzyme cascade system was developed to synthesize L-xylose from an inexpensive substrate, xylitol. The study involved cloning and co-expressing two key genes: the L-fucose isomerase (L-fucI) gene from Escherichia coli K-12 and the xylitol-4-dehydrogenase (xdh) gene from Pantoea ananatis ATCC 43072 in Escherichia coli. The resulting recombinant cells, engineered with the PET28a-xdh/L-fucI vector, were able to effectively convert xylitol to L-xylose. The system showed optimal performance at 40°C and a pH of 10.0. Moreover, Zn²⁺ (7.5 mM) enhanced the catalytic activity by 1.34 times. This approach yielded 52.2 g/L of L-xylose from an initial 80 g/L xylitol concentration, with a 65% conversion efficiency and a productivity rate of 1.86. The study highlights a practical method for producing L-xylose from xylitol through a co-expression system carrying the L-fucI and xdh genes.Keywords: l-fucose isomerase, xylitol-4-dehydrogenase, l-xylose, xylitol, co-expression
Procedia PDF Downloads 25443 Vibration-Based Data-Driven Model for Road Health Monitoring
Authors: Guru Prakash, Revanth Dugalam
Abstract:
A road’s condition often deteriorates due to harsh loading such as overload due to trucks, and severe environmental conditions such as heavy rain, snow load, and cyclic loading. In absence of proper maintenance planning, this results in potholes, wide cracks, bumps, and increased roughness of roads. In this paper, a data-driven model will be developed to detect these damages using vibration and image signals. The key idea of the proposed methodology is that the road anomaly manifests in these signals, which can be detected by training a machine learning algorithm. The use of various machine learning techniques such as the support vector machine and Radom Forest method will be investigated. The proposed model will first be trained and tested with artificially simulated data, and the model architecture will be finalized by comparing the accuracies of various models. Once a model is fixed, the field study will be performed, and data will be collected. The field data will be used to validate the proposed model and to predict the future road’s health condition. The proposed will help to automate the road condition monitoring process, repair cost estimation, and maintenance planning process.Keywords: SVM, data-driven, road health monitoring, pot-hole
Procedia PDF Downloads 86442 Statistical Analysis of the Impact of Maritime Transport Gross Domestic Product (GDP) on Nigeria’s Economy
Authors: Kehinde Peter Oyeduntan, Kayode Oshinubi
Abstract:
Nigeria is referred as the ‘Giant of Africa’ due to high population, land mass and large economy. However, it still trails far behind many smaller economies in the continent in terms of maritime operations. As we have seen that the maritime industry is the spark plug for national growth, because it houses the most crucial infrastructure that generates wealth for a nation, it is worrisome that a nation with six seaports lag in maritime activities. In this research, we have studied how the Gross Domestic Product (GDP) of the maritime transport influences the Nigerian economy. To do this, we applied Simple Linear Regression (SLR), Support Vector Machine (SVM), Polynomial Regression Model (PRM), Generalized Additive Model (GAM) and Generalized Linear Mixed Model (GLMM) to model the relationship between the nation’s Total GDP (TGDP) and the Maritime Transport GDP (MGDP) using a time series data of 20 years. The result showed that the MGDP is statistically significant to the Nigerian economy. Amongst the statistical tool applied, the PRM of order 4 describes the relationship better when compared to other methods. The recommendations presented in this study will guide policy makers and help improve the economy of Nigeria in terms of its GDP.Keywords: maritime transport, economy, GDP, regression, port
Procedia PDF Downloads 154441 Isolation and Classification of Red Blood Cells in Anemic Microscopic Images
Authors: Jameela Ali Alkrimi, Abdul Rahim Ahmad, Azizah Suliman, Loay E. George
Abstract:
Red blood cells (RBCs) are among the most commonly and intensively studied type of blood cells in cell biology. The lack of RBCs is a condition characterized by lower than normal hemoglobin level; this condition is referred to as 'anemia'. In this study, a software was developed to isolate RBCs by using a machine learning approach to classify anemic RBCs in microscopic images. Several features of RBCs were extracted using image processing algorithms, including principal component analysis (PCA). With the proposed method, RBCs were isolated in 34 second from an image containing 18 to 27 cells. We also proposed that PCA could be performed to increase the speed and efficiency of classification. Our classifier algorithm yielded accuracy rates of 100%, 99.99%, and 96.50% for K-nearest neighbor (K-NN) algorithm, support vector machine (SVM), and neural network ANN, respectively. Classification was evaluated in highly sensitivity, specificity, and kappa statistical parameters. In conclusion, the classification results were obtained for a short time period with more efficient when PCA was used.Keywords: red blood cells, pre-processing image algorithms, classification algorithms, principal component analysis PCA, confusion matrix, kappa statistical parameters, ROC
Procedia PDF Downloads 405440 Assessing the Role of Human Mobility on Malaria Transmission in South Sudan
Authors: A. Y. Mukhtar, J. B. Munyakazi, R. Ouifki
Abstract:
Over the past few decades, the unprecedented increase in mobility has raised considerable concern about the relationship between mobility and vector-borne diseases and malaria in particular. Thus, one can claim that human mobility is one of the contributing factors to the resurgence of malaria. To assess human mobility on malaria burden among hosts, we formulate a movement-based model on a network of patches. We then extend human multi-group SEIAR deterministic epidemic models into a system of stochastic differential equations (SDEs). Our quantitative stochastic model which is expressed in terms of average rates of movement between compartments is fitted to time-series data (weekly malaria data of 2011 for each patch) using the maximum likelihood approach. Using the metapopulation (multi-group) model, we compute and analyze the basic reproduction number. The result shows that human movement is sufficient to preserve malaria disease firmness in the patches with the low transmission. With these results, we concluded that the sensitivity of malaria to the human mobility is turning to be greatly important over the implications of future malaria control in South Sudan.Keywords: basic reproduction number, malaria, maximum likelihood, movement, stochastic model
Procedia PDF Downloads 134439 The Survey of Phlebotomine Sandfly (Diptera: Psychodidae) of Al-Asaba Area in the Northwest Region of the Libya
Authors: Asherf El-Abaied, Elsadik Anan, Badereddin Annajar, Mustafa Saieh, Abudalnaser El-Buni
Abstract:
Zoonotic Cutaneous Leishmaniasis (ZCL) has been endemic in the Northwestern region of Libya for over nine decades. Survey of sandfly fauna in the region revealed that 13 species have been recorded with various distribution and abundance patterns. Phlebotomus papatasi proved to be the main vector of the disease in many areas. To identify sandfly species present in the Al-Asaba town and determine their spatial and seasonal abundance. An epidemiological analysis of the data obtained from the recorded cases was also carried out. Sand flies collected from various sites using sticky traps and CDC miniature light traps during the period from March-November 2006. Recorded ZCL cases were collected from the local Primary Health Care Department and analysed using SPSS statistical package. Ten species of sandflies were identified, seven belong to the genus Phlebotomus and three belong to the genus Sergentomyia. P. papatasi was the most abundant species with peak season recorded in September. The prevalence of the disease was low however; notable increase of ZCL cases in last three years has been indicated.Keywords: Cutaneous leishmaniasis, Phlebotomus papatasi, sandfly fauna, Libya
Procedia PDF Downloads 301438 Social Media Mining with R. Twitter Analyses
Authors: Diana Codat
Abstract:
Tweets' analysis is part of text mining. Each document is a written text. It's possible to apply the usual text search techniques, in particular by switching to the bag-of-words representation. But the tweets induce peculiarities. Some may enrich the analysis. Thus, their length is calibrated (at least as far as public messages are concerned), special characters make it possible to identify authors (@) and themes (#), the tweet and retweet mechanisms make it possible to follow the diffusion of the information. Conversely, other characteristics may disrupt the analyzes. Because space is limited, authors often use abbreviations, emoticons to express feelings, and they do not pay much attention to spelling. All this creates noise that can complicate the task. The tweets carry a lot of potentially interesting information. Their exploitation is one of the main axes of the analysis of the social networks. We show how to access Twitter-related messages. We will initiate a study of the properties of the tweets, and we will follow up on the exploitation of the content of the messages. We will work under R with the package 'twitteR'. The study of tweets is a strong focus of analysis of social networks because Twitter has become an important vector of communication. This example shows that it is easy to initiate an analysis from data extracted directly online. The data preparation phase is of great importance.Keywords: data mining, language R, social networks, Twitter
Procedia PDF Downloads 184437 Use the Null Space to Create Starting Point for Stochastic Programming
Authors: Ghussoun Al-Jeiroudi
Abstract:
Stochastic programming is one of the powerful technique which is used to solve real-life problems. Hence, the data of real-life problems is subject to significant uncertainty. Uncertainty is well studied and modeled by stochastic programming. Each day, problems become bigger and bigger and the need for a tool, which does deal with large scale problems, increase. Interior point method is a perfect tool to solve such problems. Interior point method is widely employed to solve the programs, which arise from stochastic programming. It is an iterative technique, so it is required a starting point. Well design starting point plays an important role in improving the convergence speed. In this paper, we propose a starting point for interior point method for multistage stochastic programming. Usually, the optimal solution of stage k+1 is used as starting point for the stage k. This point has the advantage of being close to the solution of the current program. However, it has a disadvantage; it is not in the feasible region of the current program. So, we suggest to take this point and modifying it. That is by adding to it a vector in the null space of the matrix of the unchanged constraints because the solution will change only in the null space of this matrix.Keywords: interior point methods, stochastic programming, null space, starting points
Procedia PDF Downloads 418436 Traffic Density Measurement by Automatic Detection of the Vehicles Using Gradient Vectors from Aerial Images
Authors: Saman Ghaffarian, Ilgin Gökaşar
Abstract:
This paper presents a new automatic vehicle detection method from very high resolution aerial images to measure traffic density. The proposed method starts by extracting road regions from image using road vector data. Then, the road image is divided into equal sections considering resolution of the images. Gradient vectors of the road image are computed from edge map of the corresponding image. Gradient vectors on the each boundary of the sections are divided where the gradient vectors significantly change their directions. Finally, number of vehicles in each section is carried out by calculating the standard deviation of the gradient vectors in each group and accepting the group as vehicle that has standard deviation above predefined threshold value. The proposed method was tested in four very high resolution aerial images acquired from Istanbul, Turkey which illustrate roads and vehicles with diverse characteristics. The results show the reliability of the proposed method in detecting vehicles by producing 86% overall F1 accuracy value.Keywords: aerial images, intelligent transportation systems, traffic density measurement, vehicle detection
Procedia PDF Downloads 379435 Optimizing Machine Learning Algorithms for Defect Characterization and Elimination in Liquids Manufacturing
Authors: Tolulope Aremu
Abstract:
The key process steps to produce liquid detergent products will introduce potential defects, such as formulation, mixing, filling, and packaging, which might compromise product quality, consumer safety, and operational efficiency. Real-time identification and characterization of such defects are of prime importance for maintaining high standards and reducing waste and costs. Usually, defect detection is performed by human inspection or rule-based systems, which is very time-consuming, inconsistent, and error-prone. The present study overcomes these limitations in dealing with optimization in defect characterization within the process for making liquid detergents using Machine Learning algorithms. Performance testing of various machine learning models was carried out: Support Vector Machine, Decision Trees, Random Forest, and Convolutional Neural Network on defect detection and classification of those defects like wrong viscosity, color deviations, improper filling of a bottle, packaging anomalies. These algorithms have significantly benefited from a variety of optimization techniques, including hyperparameter tuning and ensemble learning, in order to greatly improve detection accuracy while minimizing false positives. Equipped with a rich dataset of defect types and production parameters consisting of more than 100,000 samples, our study further includes information from real-time sensor data, imaging technologies, and historic production records. The results are that optimized machine learning models significantly improve defect detection compared to traditional methods. Take, for instance, the CNNs, which run at 98% and 96% accuracy in detecting packaging anomaly detection and bottle filling inconsistency, respectively, by fine-tuning the model with real-time imaging data, through which there was a reduction in false positives of about 30%. The optimized SVM model on detecting formulation defects gave 94% in viscosity variation detection and color variation. These values of performance metrics correspond to a giant leap in defect detection accuracy compared to the usual 80% level achieved up to now by rule-based systems. Moreover, this optimization with models can hasten defect characterization, allowing for detection time to be below 15 seconds from an average of 3 minutes using manual inspections with real-time processing of data. With this, the reduction in time will be combined with a 25% reduction in production downtime because of proactive defect identification, which can save millions annually in recall and rework costs. Integrating real-time machine learning-driven monitoring drives predictive maintenance and corrective measures for a 20% improvement in overall production efficiency. Therefore, the optimization of machine learning algorithms in defect characterization optimum scalability and efficiency for liquid detergent companies gives improved operational performance to higher levels of product quality. In general, this method could be conducted in several industries within the Fast moving consumer Goods industry, which would lead to an improved quality control process.Keywords: liquid detergent manufacturing, defect detection, machine learning, support vector machines, convolutional neural networks, defect characterization, predictive maintenance, quality control, fast-moving consumer goods
Procedia PDF Downloads 18434 Synthetic Aperture Radar Remote Sensing Classification Using the Bag of Visual Words Model to Land Cover Studies
Authors: Reza Mohammadi, Mahmod R. Sahebi, Mehrnoosh Omati, Milad Vahidi
Abstract:
Classification of high resolution polarimetric Synthetic Aperture Radar (PolSAR) images plays an important role in land cover and land use management. Recently, classification algorithms based on Bag of Visual Words (BOVW) model have attracted significant interest among scholars and researchers in and out of the field of remote sensing. In this paper, BOVW model with pixel based low-level features has been implemented to classify a subset of San Francisco bay PolSAR image, acquired by RADARSAR 2 in C-band. We have used segment-based decision-making strategy and compared the result with the result of traditional Support Vector Machine (SVM) classifier. 90.95% overall accuracy of the classification with the proposed algorithm has shown that the proposed algorithm is comparable with the state-of-the-art methods. In addition to increase in the classification accuracy, the proposed method has decreased undesirable speckle effect of SAR images.Keywords: Bag of Visual Words (BOVW), classification, feature extraction, land cover management, Polarimetric Synthetic Aperture Radar (PolSAR)
Procedia PDF Downloads 209433 Development and Characterization of Double Liposomes Based Dual Drug Delivery System for H. Pylori Targeting
Authors: Ashish Kumar Jain, Deepak Mishra
Abstract:
The objective of the present investigation was to prepare and evaluate a vesicular dual drug delivery system for effective management of mucosal ulcer. Inner encapsulating and Double liposomes were prepared by glass bead and reverse phase evaporation method respectively. The formulation consisted of inner liposomes bearing Ranitidine Bismuth Citrate (RBC) and outer liposomes encapsulating Amoxicillin trihydrate (AMOX). The optimized inner liposomes and double liposomes were extensively characterized for vesicle size, morphology, zeta potential, vesicles count, entrapment efficiency and in vitro drug release. In vitro, the double liposomes demonstrated a sustained release of AMOX and RBC viz 91.4±1.8% and 77.2±2.1% respectively at the end of 72 hr. Furthermore binding specificity and targeting propensity toward H. pylori (SKP-56) was confirmed by agglutination and in situ adherence assay. Reduction of the absolute alcohol induced ulcerogenic index from 3.01 ± 0.25 to 0.31 ± 0.09 and 100% H. pylori clearance rate was observed. These results suggested that double liposomes are potential vector for the development of dual drug delivery for effective treatment of H. pylori-associated peptic ulcer.Keywords: double liposomes, H. pylori targeting, PE liposomes, glass-beads method, peptic ulcers
Procedia PDF Downloads 449432 Spatial Patterns and Temporal Evolution of Octopus Abundance in the Mauritanian Zone
Authors: Dedah Ahmed Babou, Nicolas Bez
Abstract:
The Min-Max autocorrelation factor (MAF) approach makes it possible to express in a space formed by spatially independent factors, spatiotemporal observations. These factors are ordered in decreasing order of spatial autocorrelation. The starting observations are thus expressed in the space formed by these factors according to temporal coordinates. Each vector of temporal coefficients expresses the temporal evolution of the weight of the corresponding factor. Applying this approach has enabled us to achieve the following results: (i) Define a spatially orthogonal space in which the projections of the raw data are determined; (ii) Define a limit threshold for the factors with the strongest structures in order to analyze the weight, and the temporal evolution of these different structures (iii) Study the correlation between the temporal evolution of the persistent spatial structures and that of the observed average abundance (iv) Propose prototypes of campaigns reflecting a high vs. low abundance (v) Propose a classification of campaigns that highlights seasonal and/or temporal similarities. These results were obtained by analyzing the octopus yield during the scientific campaigns of the oceanographic vessel Al Awam during the period 1989-2017 in the Mauritanian exclusive economic zone.Keywords: spatiotemporal , autocorrelation, kriging, variogram, Octopus vulgaris
Procedia PDF Downloads 147431 Automated Ultrasound Carotid Artery Image Segmentation Using Curvelet Threshold Decomposition
Authors: Latha Subbiah, Dhanalakshmi Samiappan
Abstract:
In this paper, we propose denoising Common Carotid Artery (CCA) B mode ultrasound images by a decomposition approach to curvelet thresholding and automatic segmentation of the intima media thickness and adventitia boundary. By decomposition, the local geometry of the image, its direction of gradients are well preserved. The components are combined into a single vector valued function, thus removes noise patches. Double threshold is applied to inherently remove speckle noise in the image. The denoised image is segmented by active contour without specifying seed points. Combined with level set theory, they provide sub regions with continuous boundaries. The deformable contours match to the shapes and motion of objects in the images. A curve or a surface under constraints is developed from the image with the goal that it is pulled into the necessary features of the image. Region based and boundary based information are integrated to achieve the contour. The method treats the multiplicative speckle noise in objective and subjective quality measurements and thus leads to better-segmented results. The proposed denoising method gives better performance metrics compared with other state of art denoising algorithms.Keywords: curvelet, decomposition, levelset, ultrasound
Procedia PDF Downloads 340430 An Automated System for the Detection of Citrus Greening Disease Based on Visual Descriptors
Authors: Sidra Naeem, Ayesha Naeem, Sahar Rahim, Nadia Nawaz Qadri
Abstract:
Citrus greening is a bacterial disease that causes considerable damage to citrus fruits worldwide. Efficient method for this disease detection must be carried out to minimize the production loss. This paper presents a pattern recognition system that comprises three stages for the detection of citrus greening from Orange leaves: segmentation, feature extraction and classification. Image segmentation is accomplished by adaptive thresholding. The feature extraction stage comprises of three visual descriptors i.e. shape, color and texture. From shape feature we have used asymmetry index, from color feature we have used histogram of Cb component from YCbCr domain and from texture feature we have used local binary pattern. Classification was done using support vector machines and k nearest neighbors. The best performances of the system is Accuracy = 88.02% and AUROC = 90.1% was achieved by automatic segmented images. Our experiments validate that: (1). Segmentation is an imperative preprocessing step for computer assisted diagnosis of citrus greening, and (2). The combination of shape, color and texture features form a complementary set towards the identification of citrus greening disease.Keywords: citrus greening, pattern recognition, feature extraction, classification
Procedia PDF Downloads 184429 Introduction of Robust Multivariate Process Capability Indices
Authors: Behrooz Khalilloo, Hamid Shahriari, Emad Roghanian
Abstract:
Process capability indices (PCIs) are important concepts of statistical quality control and measure the capability of processes and how much processes are meeting certain specifications. An important issue in statistical quality control is parameter estimation. Under the assumption of multivariate normality, the distribution parameters, mean vector and variance-covariance matrix must be estimated, when they are unknown. Classic estimation methods like method of moment estimation (MME) or maximum likelihood estimation (MLE) makes good estimation of the population parameters when data are not contaminated. But when outliers exist in the data, MME and MLE make weak estimators of the population parameters. So we need some estimators which have good estimation in the presence of outliers. In this work robust M-estimators for estimating these parameters are used and based on robust parameter estimators, robust process capability indices are introduced. The performances of these robust estimators in the presence of outliers and their effects on process capability indices are evaluated by real and simulated multivariate data. The results indicate that the proposed robust capability indices perform much better than the existing process capability indices.Keywords: multivariate process capability indices, robust M-estimator, outlier, multivariate quality control, statistical quality control
Procedia PDF Downloads 283428 Effect of Fiscal Policy on Growth in India
Authors: Parma Chakravartti
Abstract:
The impact of government spending and taxation on economic growth has remained a central issue of fiscal policy analysis. There is a wide range of opinions over the strength of fiscal policy’s effect on macroeconomic variables. It can be argued that the impact of fiscal policy depends on the structure and economic condition of the economy. This study makes an attempt to examine the effect of fiscal policy shocks on growth in India using the structural vector autoregressive model (SVAR), considering data from 1950 to 2019. The study finds that government spending is an important instrument of growth in India, where the share of revenue expenditure to capital expenditure plays a key role. The optimum composition of total expenditure is important for growth and it is not necessarily true that capital expenditure multiplier is more than revenue expenditure multiplier. The study also finds that the impact of public economic activities on private economic activities for both consumption expenditure and gross capital formation of government crowds in private consumption expenditure and private gross capital formation, respectively, thus indicating that government expenditure complements private expenditure in India.Keywords: government spending, fiscal policy, multiplier, growth
Procedia PDF Downloads 133427 Competition, Stability, and Economic Growth: A Causality Approach
Authors: Mahvish Anwaar
Abstract:
Research Question: In this paper, we explore the causal relationship between banking competition, banking stability, and economic growth. Research Findings: The unbalanced panel data starting from 2000 to 2018 is collected to analyze the causality among banking competition, banking stability, and economic growth. The main focus of the study is to check the direction of causality among selected variables. The results of the study support the demand following, supply leading, feedback, and neutrality hypothesis conditional to different measures of banking competition, banking stability, and economic growth. Theoretical Implication: Jayakumar, Pradhan, Dash, Maradana, and Gaurav (2018) proposed a theoretical model of the causal relationship between banking competition, banking stability, and economic growth by using different indicators. So, we empirically test the proposed indicators in our study. This study makes a contribution to the literature by showing the defined relationship between developing and developed countries. Policy Implications: The study covers various policy implications regarding investors to analyze how to properly manage their finances, and government agencies will take help from the present study to find the best and most suitable policies by examining how the economy can grow concerning its finances.Keywords: competition, stability, economic growth, vector auto-regression, granger causality
Procedia PDF Downloads 63426 A Study of ZY3 Satellite Digital Elevation Model Verification and Refinement with Shuttle Radar Topography Mission
Authors: Bo Wang
Abstract:
As the first high-resolution civil optical satellite, ZY-3 satellite is able to obtain high-resolution multi-view images with three linear array sensors. The images can be used to generate Digital Elevation Models (DEM) through dense matching of stereo images. However, due to the clouds, forest, water and buildings covered on the images, there are some problems in the dense matching results such as outliers and areas failed to be matched (matching holes). This paper introduced an algorithm to verify the accuracy of DEM that generated by ZY-3 satellite with Shuttle Radar Topography Mission (SRTM). Since the accuracy of SRTM (Internal accuracy: 5 m; External accuracy: 15 m) is relatively uniform in the worldwide, it may be used to improve the accuracy of ZY-3 DEM. Based on the analysis of mass DEM and SRTM data, the processing can be divided into two aspects. The registration of ZY-3 DEM and SRTM can be firstly performed using the conjugate line features and area features matched between these two datasets. Then the ZY-3 DEM can be refined by eliminating the matching outliers and filling the matching holes. The matching outliers can be eliminated based on the statistics on Local Vector Binning (LVB). The matching holes can be filled by the elevation interpolated from SRTM. Some works are also conducted for the accuracy statistics of the ZY-3 DEM.Keywords: ZY-3 satellite imagery, DEM, SRTM, refinement
Procedia PDF Downloads 343425 Performance Estimation of Two Port Multiple-Input and Multiple-Output Antenna for Wireless Local Area Network Applications
Authors: Radha Tomar, Satish K. Jain, Manish Panchal, P. S. Rathore
Abstract:
In the presented work, inset fed microstrip patch antenna (IFMPA) based two port MIMO Antenna system has been proposed, which is suitable for wireless local area network (WLAN) applications. IFMPA has been designed, optimized for 2.4 GHz and applied for MIMO formation. The optimized parameters of the proposed IFMPA have been used for fabrication of antenna and two port MIMO in a laboratory. Fabrication of the designed MIMO antenna has been done and tested experimentally for performance parameters like Envelope Correlation Coefficient (ECC), Mean Effective Gain (MEG), Directive Gain (DG), Channel Capacity Loss (CCL), Multiplexing Efficiency (ME) etc and results are compared with simulated parameters extracted with simulated S parameters to validate the results. The simulated and experimentally measured plots and numerical values of these MIMO performance parameters resembles very much with each other. This shows the success of MIMO antenna design methodology.Keywords: multiple-input and multiple-output, wireless local area network, vector network analyzer, envelope correlation coefficient
Procedia PDF Downloads 55424 Epileptic Seizure Onset Detection via Energy and Neural Synchronization Decision Fusion
Authors: Marwa Qaraqe, Muhammad Ismail, Erchin Serpedin
Abstract:
This paper presents a novel architecture for a patient-specific epileptic seizure onset detector using scalp electroencephalography (EEG). The proposed architecture is based on the decision fusion calculated from energy and neural synchronization related features. Specifically, one level of the detector calculates the condition number (CN) of an EEG matrix to evaluate the amount of neural synchronization present within the EEG channels. On a parallel level, the detector evaluates the energy contained in four EEG frequency subbands. The information is then fed into two independent (parallel) classification units based on support vector machines to determine the onset of a seizure event. The decisions from the two classifiers are then combined together according to two fusion techniques to determine a global decision. Experimental results demonstrate that the detector based on the AND fusion technique outperforms existing detectors with a sensitivity of 100%, detection latency of 3 seconds, while it achieves a 2:76 false alarm rate per hour. The OR fusion technique achieves a sensitivity of 100%, and significantly improves delay latency (0:17 seconds), yet it achieves 12 false alarms per hour.Keywords: epilepsy, EEG, seizure onset, electroencephalography, neuron, detection
Procedia PDF Downloads 477423 Tibyan Automated Arabic Correction Using Machine-Learning in Detecting Syntactical Mistakes
Authors: Ashwag O. Maghraby, Nida N. Khan, Hosnia A. Ahmed, Ghufran N. Brohi, Hind F. Assouli, Jawaher S. Melibari
Abstract:
The Arabic language is one of the most important languages. Learning it is so important for many people around the world because of its religious and economic importance and the real challenge lies in practicing it without grammatical or syntactical mistakes. This research focused on detecting and correcting the syntactic mistakes of Arabic syntax according to their position in the sentence and focused on two of the main syntactical rules in Arabic: Dual and Plural. It analyzes each sentence in the text, using Stanford CoreNLP morphological analyzer and machine-learning approach in order to detect the syntactical mistakes and then correct it. A prototype of the proposed system was implemented and evaluated. It uses support vector machine (SVM) algorithm to detect Arabic grammatical errors and correct them using the rule-based approach. The prototype system has a far accuracy 81%. In general, it shows a set of useful grammatical suggestions that the user may forget about while writing due to lack of familiarity with grammar or as a result of the speed of writing such as alerting the user when using a plural term to indicate one person.Keywords: Arabic language acquisition and learning, natural language processing, morphological analyzer, part-of-speech
Procedia PDF Downloads 152422 Plasma Properties Effect on Fluorescent Tube Plasma Antenna Performance
Authors: A. N. Dagang, E. I. Ismail, Z. Zakaria
Abstract:
This paper presents the analysis on the performance of monopole antenna with fluorescent tubes. In this research, the simulation and experimental approach is conducted. The fluorescent tube with different length and size is designed using Computer Simulation Technology (CST) software and the characteristics of antenna parameter are simulated throughout the software. CST was used to simulate antenna parameters such as return loss, resonant frequency, gain and directivity. Vector Network Analyzer (VNA) was used to measure the return loss of plasma antenna in order to validate the simulation results. In the simulation and experiment, the supply frequency is set starting from 1 GHz to 10 GHz. The results show that the return loss of plasma antenna changes when size of fluorescent tubes is varied, correspond to the different plasma properties. It shows that different values of plasma properties such as plasma frequency and collision frequency gives difference result of return loss, gain and directivity. For the gain, the values range from 2.14 dB to 2.36 dB. The return loss of plasma antenna offers higher value range from -22.187 dB to -32.903 dB. The higher the values of plasma frequency and collision frequency, the higher return loss can be obtained. The values obtained are comparative to the conventional type of metal antenna.Keywords: plasma antenna, fluorescent tube, CST, plasma parameters
Procedia PDF Downloads 387421 Multi-Granularity Feature Extraction and Optimization for Pathological Speech Intelligibility Evaluation
Authors: Chunying Fang, Haifeng Li, Lin Ma, Mancai Zhang
Abstract:
Speech intelligibility assessment is an important measure to evaluate the functional outcomes of surgical and non-surgical treatment, speech therapy and rehabilitation. The assessment of pathological speech plays an important role in assisting the experts. Pathological speech usually is non-stationary and mutational, in this paper, we describe a multi-granularity combined feature schemes, and which is optimized by hierarchical visual method. First of all, the difference granularity level pathological features are extracted which are BAFS (Basic acoustics feature set), local spectral characteristics MSCC (Mel s-transform cepstrum coefficients) and nonlinear dynamic characteristics based on chaotic analysis. Latterly, radar chart and F-score are proposed to optimize the features by the hierarchical visual fusion. The feature set could be optimized from 526 to 96-dimensions.The experimental results denote that new features by support vector machine (SVM) has the best performance, with a recognition rate of 84.4% on NKI-CCRT corpus. The proposed method is thus approved to be effective and reliable for pathological speech intelligibility evaluation.Keywords: pathological speech, multi-granularity feature, MSCC (Mel s-transform cepstrum coefficients), F-score, radar chart
Procedia PDF Downloads 283420 Data Modeling and Calibration of In-Line Pultrusion and Laser Ablation Machine Processes
Authors: David F. Nettleton, Christian Wasiak, Jonas Dorissen, David Gillen, Alexandr Tretyak, Elodie Bugnicourt, Alejandro Rosales
Abstract:
In this work, preliminary results are given for the modeling and calibration of two inline processes, pultrusion, and laser ablation, using machine learning techniques. The end product of the processes is the core of a medical guidewire, manufactured to comply with a user specification of diameter and flexibility. An ensemble approach is followed which requires training several models. Two state of the art machine learning algorithms are benchmarked: Kernel Recursive Least Squares (KRLS) and Support Vector Regression (SVR). The final objective is to build a precise digital model of the pultrusion and laser ablation process in order to calibrate the resulting diameter and flexibility of a medical guidewire, which is the end product while taking into account the friction on the forming die. The result is an ensemble of models, whose output is within a strict required tolerance and which covers the required range of diameter and flexibility of the guidewire end product. The modeling and automatic calibration of complex in-line industrial processes is a key aspect of the Industry 4.0 movement for cyber-physical systems.Keywords: calibration, data modeling, industrial processes, machine learning
Procedia PDF Downloads 298419 Coupled Spacecraft Orbital and Attitude Modeling and Simulation in Multi-Complex Modes
Authors: Amr Abdel Azim Ali, G. A. Elsheikh, Moutaz Hegazy
Abstract:
This paper presents verification of a modeling and simulation for a Spacecraft (SC) attitude and orbit control system. Detailed formulation of coupled SC orbital and attitude equations of motion is performed in order to achieve accepted accuracy to meet the requirements of multitargets tracking and orbit correction complex modes. Correction of the target parameter based on the estimated state vector during shooting time to enhance pointing accuracy is considered. Time-optimal nonlinear feedback control technique was used in order to take full advantage of the maximum torques that the controller can deliver. This simulation provides options for visualizing SC trajectory and attitude in a 3D environment by including an interface with V-Realm Builder and VR Sink in Simulink/MATLAB. Verification data confirms the simulation results, ensuring that the model and the proposed control law can be used successfully for large and fast tracking and is robust enough to keep the pointing accuracy within the desired limits with considerable uncertainty in inertia and control torque.Keywords: attitude and orbit control, time-optimal nonlinear feedback control, modeling and simulation, pointing accuracy, maximum torques
Procedia PDF Downloads 331418 Comparison of Food Products Contaminated by DDTs in South Africa and Mozambique
Authors: Lesa A. Thompson, Yoshinori Ikenaka, Victor Wepener, Mayumi Ishizuka
Abstract:
One method for controlling malaria in endemic regions is the killing of vector mosquitoes using pesticides such as DDT in indoor residual spraying (IRS). This study was carried out to investigate the presence of and human health risk due to DDT and its metabolites (collectively, DDTs) contaminating human food sources in areas where DDT is used for IRS. Free-range chicken products (meat and eggs) were collected from homesteads in KwaZulu-Natal Province in the northeast of South Africa, and fish meat samples from Maputo Bay in neighbouring Mozambique. Samples were analysed for DDTs (o,p’-DDT, p,p’-DDT, o,p’-DDD, p,p’-DDD, o,p’-DDE and p,p’-DDE) using a gas chromatograph with electron capture detector (GC-ECD). DDTs were detected in all food types, with the predominant congener being p,p’-DDE. The presence of p,p’-DDT confirmed recent release of DDT into the environment. By using concentration levels detected in foods and national consumption levels, the risk to human health through consumption of such food products was calculated. In order of risk level, these were: chicken eggs > chicken meat > fish meat. Human risk (carcinogenic) values greater than one suggest there is an increased health risk through consumption of these foods.Keywords: DDT, food contamination, human health risk, Mozambique, South Africa
Procedia PDF Downloads 342