Search results for: unknown input observer
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3139

Search results for: unknown input observer

1909 Reversible and Adaptive Watermarking for MRI Medical Images

Authors: Nisar Ahmed Memon

Abstract:

A new medical image watermarking scheme delivering high embedding capacity is presented in this paper. Integer Wavelet Transform (IWT), Companding technique and adaptive thresholding are used in this scheme. The proposed scheme implants, recovers the hidden information and restores the input image to its pristine state at the receiving end. Magnetic Resonance Imaging (MRI) images are used for experimental purposes. The scheme first segment the MRI medical image into non-overlapping blocks and then inserts watermark into wavelet coefficients having a high frequency of each block. The scheme uses block-based watermarking adopting iterative optimization of threshold for companding in order to avoid the histogram pre and post processing. Results show that proposed scheme performs better than other reversible medical image watermarking schemes available in literature for MRI medical images.

Keywords: adaptive thresholding, companding technique, data authentication, reversible watermarking

Procedia PDF Downloads 288
1908 Unsupervised Learning with Self-Organizing Maps for Named Entity Recognition in the CONLL2003 Dataset

Authors: Assel Jaxylykova, Alexnder Pak

Abstract:

This study utilized a Self-Organizing Map (SOM) for unsupervised learning on the CONLL-2003 dataset for Named Entity Recognition (NER). The process involved encoding words into 300-dimensional vectors using FastText. These vectors were input into a SOM grid, where training adjusted node weights to minimize distances. The SOM provided a topological representation for identifying and clustering named entities, demonstrating its efficacy without labeled examples. Results showed an F1-measure of 0.86, highlighting SOM's viability. Although some methods achieve higher F1 measures, SOM eliminates the need for labeled data, offering a scalable and efficient alternative. The SOM's ability to uncover hidden patterns provides insights that could enhance existing supervised methods. Further investigation into potential limitations and optimization strategies is suggested to maximize benefits.

Keywords: named entity recognition, natural language processing, self-organizing map, CONLL-2003, semantics

Procedia PDF Downloads 30
1907 Web Proxy Detection via Bipartite Graphs and One-Mode Projections

Authors: Zhipeng Chen, Peng Zhang, Qingyun Liu, Li Guo

Abstract:

With the Internet becoming the dominant channel for business and life, many IPs are increasingly masked using web proxies for illegal purposes such as propagating malware, impersonate phishing pages to steal sensitive data or redirect victims to other malicious targets. Moreover, as Internet traffic continues to grow in size and complexity, it has become an increasingly challenging task to detect the proxy service due to their dynamic update and high anonymity. In this paper, we present an approach based on behavioral graph analysis to study the behavior similarity of web proxy users. Specifically, we use bipartite graphs to model host communications from network traffic and build one-mode projections of bipartite graphs for discovering social-behavior similarity of web proxy users. Based on the similarity matrices of end-users from the derived one-mode projection graphs, we apply a simple yet effective spectral clustering algorithm to discover the inherent web proxy users behavior clusters. The web proxy URL may vary from time to time. Still, the inherent interest would not. So, based on the intuition, by dint of our private tools implemented by WebDriver, we examine whether the top URLs visited by the web proxy users are web proxies. Our experiment results based on real datasets show that the behavior clusters not only reduce the number of URLs analysis but also provide an effective way to detect the web proxies, especially for the unknown web proxies.

Keywords: bipartite graph, one-mode projection, clustering, web proxy detection

Procedia PDF Downloads 238
1906 A Longitudinal Study of Social Engagement in Classroom in Children with Autism Spectrum Disorder

Authors: Cecile Garry, Katia Rovira, Julie Brisson

Abstract:

Autism Spectrum Disorder (ASD) is defined by a qualitative and quantitative impairment of social interaction. Indeed early intervention programs, such as the Early Start Denver Model (ESDM), aimed at encouraging the development of social skills. In classroom, the children need to be socially engaged to learn. Early intervention programs can thus be implemented in kindergarten schools. In these schools, ASD children have more opportunities to interact with their peers or adults than in elementary schools. However, the preschool children with ASD are less socially engaged than their typically developing peers in the classroom. They initiate, respond and maintain less the social interactions. In addition, they produce more responses than initiations. When they interact, the non verbal communication is more used than verbal or symbolic communication forms and they are more engaged with adults than with peers. Nevertheless, communicative patterns may vary according to the clinical profiles of ASD children. Indeed, the ASD children with better cognitive skills interact more with their peers and use more symbolic communication than the ASD children with a low cognitive level. ASD children with the less severe symptoms use more the verbal communication than ASD children with the more severe symptoms. Small groups and structured activities encourage coordinated joint engagement episodes in ASD children. Our goal is to evaluate ASD children’s social engagement development in class, with their peers or adults, during dyadic or group activities. Participants were 19 preschool children with ASD aged from 3 to 6 years old that benefited of an early intervention in special kindergarten schools. Severity of ASD symptoms was measured with the CARS at the beginning of the follow-up. Classroom situations of interaction were recorded during 10 minutes (5 minutes of dyadic interaction and 5 minutes of a group activity), every 2 months, during 10 months. Social engagement behaviors of children, including initiations, responses and imitation, directed to a peer or an adult, were then coded. The Observer software (Noldus) that allows to annotate behaviors was the coding system used. A double coding was conducted and revealed a good inter judges fidelity. Results show that ASD children were more often and longer socially engaged in dyadic than in groups situations. They were also more engaged with adults than with peers. Children with the less severe symptoms of ASD were more socially engaged in groups situations than children with the more severe symptoms of ASD. Then, ASD children with the less severe symptoms of ASD were more engaged with their peers than ASD children with the more severe symptoms of ASD. However, the engagement frequency increased during the 10 month of follow-up but only for ASD children with the more severe symptoms at the beginning. To conclude, these results highlighted the necessity of individualizing early intervention programs according to the clinical profile of the child.

Keywords: autism spectrum disorder, preschool children, developmental psychology, early interventions, social interactions

Procedia PDF Downloads 154
1905 The Application of a Hybrid Neural Network for Recognition of a Handwritten Kazakh Text

Authors: Almagul Assainova , Dariya Abykenova, Liudmila Goncharenko, Sergey Sybachin, Saule Rakhimova, Abay Aman

Abstract:

The recognition of a handwritten Kazakh text is a relevant objective today for the digitization of materials. The study presents a model of a hybrid neural network for handwriting recognition, which includes a convolutional neural network and a multi-layer perceptron. Each network includes 1024 input neurons and 42 output neurons. The model is implemented in the program, written in the Python programming language using the EMNIST database, NumPy, Keras, and Tensorflow modules. The neural network training of such specific letters of the Kazakh alphabet as ә, ғ, қ, ң, ө, ұ, ү, h, і was conducted. The neural network model and the program created on its basis can be used in electronic document management systems to digitize the Kazakh text.

Keywords: handwriting recognition system, image recognition, Kazakh font, machine learning, neural networks

Procedia PDF Downloads 253
1904 Modelling Railway Noise Over Large Areas, Assisted by GIS

Authors: Conrad Weber

Abstract:

The modelling of railway noise over large projects areas can be very time consuming in terms of preparing the noise models and calculation time. An open-source GIS program has been utilised to assist with the modelling of operational noise levels for 675km of railway corridor. A range of GIS algorithms were utilised to break up the noise model area into manageable calculation sizes. GIS was utilised to prepare and filter a range of noise modelling inputs, including building files, land uses and ground terrain. A spreadsheet was utilised to manage the accuracy of key input parameters, including train speeds, train types, curve corrections, bridge corrections and engine notch settings. GIS was utilised to present the final noise modelling results. This paper explains the noise modelling process and how the spreadsheet and GIS were utilised to accurately model this massive project efficiently.

Keywords: noise, modeling, GIS, rail

Procedia PDF Downloads 115
1903 Design of Incident Information System in IoT Virtualization Platform

Authors: Amon Olimov, Umarov Jamshid, Dae-Ho Kim, Chol-U Lee, Ryum-Duck Oh

Abstract:

This paper proposes IoT virtualization platform based incident information system. IoT information based environment is the platform that was developed for the purpose of collecting a variety of data by managing regionally scattered IoT devices easily and conveniently in addition to analyzing data collected from roads. Moreover, this paper configured the platform for the purpose of providing incident information based on sensed data. It also provides the same input/output interface as UNIX and Linux by means of matching IoT devices with the directory of file system and also the files. In addition, it has a variety of approaches as to the devices. Thus, it can be applied to not only incident information but also other platforms. This paper proposes the incident information system that identifies and provides various data in real time as to urgent matters on roads based on the existing USN/M2M and IoT visualization platform.

Keywords: incident information system, IoT, virtualization platform, USN, M2M

Procedia PDF Downloads 347
1902 A Static Android Malware Detection Based on Actual Used Permissions Combination and API Calls

Authors: Xiaoqing Wang, Junfeng Wang, Xiaolan Zhu

Abstract:

Android operating system has been recognized by most application developers because of its good open-source and compatibility, which enriches the categories of applications greatly. However, it has become the target of malware attackers due to the lack of strict security supervision mechanisms, which leads to the rapid growth of malware, thus bringing serious safety hazards to users. Therefore, it is critical to detect Android malware effectively. Generally, the permissions declared in the AndroidManifest.xml can reflect the function and behavior of the application to a large extent. Since current Android system has not any restrictions to the number of permissions that an application can request, developers tend to apply more than actually needed permissions in order to ensure the successful running of the application, which results in the abuse of permissions. However, some traditional detection methods only consider the requested permissions and ignore whether it is actually used, which leads to incorrect identification of some malwares. Therefore, a machine learning detection method based on the actually used permissions combination and API calls was put forward in this paper. Meanwhile, several experiments are conducted to evaluate our methodology. The result shows that it can detect unknown malware effectively with higher true positive rate and accuracy while maintaining a low false positive rate. Consequently, the AdaboostM1 (J48) classification algorithm based on information gain feature selection algorithm has the best detection result, which can achieve an accuracy of 99.8%, a true positive rate of 99.6% and a lowest false positive rate of 0.

Keywords: android, API Calls, machine learning, permissions combination

Procedia PDF Downloads 323
1901 Assessment Environmental and Economic of Yerba Mate as a Feed Additive on Feedlot Lamb

Authors: Danny Alexander R. Moreno, Gustavo L. Sartorello, Yuli Andrea P. Bermudez, Richard R. Lobo, Ives Claudio S. Bueno, Augusto H. Gameiro

Abstract:

Meat production is a significant sector for Brazil's economy; however, the agricultural segment has suffered censure regarding the negative impacts on the environment, which consequently results in climate change. Therefore, it is essential the implementation of nutritional strategies that can improve the environmental performance of livestock. This research aimed to estimate the environmental impact and profitability of the use of yerba mate extract (Ilex paraguariensis) as an additive in the feeding of feedlot lamb. Thirty-six castrated male lambs (average weight of 23.90 ± 3.67 kg and average age of 75 days) were randomly assigned to four experimental diets with different levels of inclusion of yerba mate extract (0, 1, 2, and 4 %) based on dry matter. The animals were confined for fifty-three days and fed with 60:40 corn silage to concentrate ratio. As an indicator of environmental impact, the carbon footprint (CF) was measured as kg of CO₂ equivalent (CO₂-eq) per kg of body weight produced (BWP). The greenhouse gas (GHG) emissions such as methane (CH₄) generated from enteric fermentation, were calculated using the sulfur hexafluoride gas tracer (SF₆) technique; while the CH₄, nitrous oxide (N₂O - emissions generated by feces and urine), and carbon dioxide (CO₂ - emissions generated by concentrate and silage processing) were estimated using the Intergovernmental Panel on Climate Change (IPCC) methodology. To estimate profitability, the gross margin was used, which is the total revenue minus the total cost; the latter is composed of the purchase of animals and food. The boundaries of this study considered only the lamb fattening system. The enteric CH₄ emission from the lamb was the largest source of on-farm GHG emissions (47%-50%), followed by CH₄ and N₂O emissions from manure (10%-20%) and CO₂ emission from the concentrate, silage, and fossil energy (17%-5%). The treatment that generated the least environmental impact was the group with 4% of yerba mate extract (YME), which showed a 3% reduction in total GHG emissions in relation to the control (1462.5 and 1505.5 kg CO₂-eq, respectively). However, the scenario with 1% YME showed an increase in emissions of 7% compared to the control group. In relation to CF, the treatment with 4% YME had the lowest value (4.1 kg CO₂-eq/kg LW) compared with the other groups. Nevertheless, although the 4% YME inclusion scenario showed the lowest CF, the gross margin decreased by 36% compared to the control group (0% YME), due to the cost of YME as a food additive. The results showed that the extract has the potential for use in reducing GHG. However, the cost of implementing this input as a mitigation strategy increased the production cost. Therefore, it is important to develop political strategies that help reduce the acquisition costs of input that contribute to the search for the environmental and economic benefit of the livestock sector.

Keywords: meat production, natural additives, profitability, sheep

Procedia PDF Downloads 129
1900 MPC of Single Phase Inverter for PV System

Authors: Irtaza M. Syed, Kaamran Raahemifar

Abstract:

This paper presents a model predictive control (MPC) of a utility interactive (UI) single phase inverter (SPI) for a photovoltaic (PV) system at residential/distribution level. The proposed model uses single-phase phase locked loop (PLL) to synchronize SPI with the grid and performs MPC control in a dq reference frame. SPI model consists of boost converter (BC), maximum power point tracking (MPPT) control, and a full bridge (FB) voltage source inverter (VSI). No PI regulators to tune and carrier and modulating waves are required to produce switching sequence. Instead, the operational model of VSI is used to synthesize sinusoidal current and track the reference. Model is validated using a three kW PV system at the input of UI-SPI in Matlab/Simulink. Implementation and results demonstrate simplicity and accuracy, as well as reliability of the model.

Keywords: phase locked loop, voltage source inverter, single phase inverter, model predictive control, Matlab/Simulink

Procedia PDF Downloads 524
1899 Emerging Cyber Threats and Cognitive Vulnerabilities: Cyberterrorism

Authors: Oludare Isaac Abiodun, Esther Omolara Abiodun

Abstract:

The purpose of this paper is to demonstrate that cyberterrorism is existing and poses a threat to computer security and national security. Nowadays, people have become excitedly dependent upon computers, phones, the Internet, and the Internet of things systems to share information, communicate, conduct a search, etc. However, these network systems are at risk from a different source that is known and unknown. These network systems risk being caused by some malicious individuals, groups, organizations, or governments, they take advantage of vulnerabilities in the computer system to hawk sensitive information from people, organizations, or governments. In doing so, they are engaging themselves in computer threats, crime, and terrorism, thereby making the use of computers insecure for others. The threat of cyberterrorism is of various forms and ranges from one country to another country. These threats include disrupting communications and information, stealing data, destroying data, leaking, and breaching data, interfering with messages and networks, and in some cases, demanding financial rewards for stolen data. Hence, this study identifies many ways that cyberterrorists utilize the Internet as a tool to advance their malicious mission, which negatively affects computer security and safety. One could identify causes for disparate anomaly behaviors and the theoretical, ideological, and current forms of the likelihood of cyberterrorism. Therefore, for a countermeasure, this paper proposes the use of previous and current computer security models as found in the literature to help in countering cyberterrorism

Keywords: cyberterrorism, computer security, information, internet, terrorism, threat, digital forensic solution

Procedia PDF Downloads 90
1898 Wavelength Conversion of Dispersion Managed Solitons at 100 Gbps through Semiconductor Optical Amplifier

Authors: Kadam Bhambri, Neena Gupta

Abstract:

All optical wavelength conversion is essential in present day optical networks for transparent interoperability, contention resolution, and wavelength routing. The incorporation of all optical wavelength convertors leads to better utilization of the network resources and hence improves the efficiency of optical networks. Wavelength convertors that can work with Dispersion Managed (DM) solitons are attractive due to their superior transmission capabilities. In this paper, wavelength conversion for dispersion managed soliton signals was demonstrated at 100 Gbps through semiconductor optical amplifier and an optical filter. The wavelength conversion was achieved for a 1550 nm input signal to1555nm output signal. The output signal was measured in terms of BER, Q factor and system margin.    

Keywords: all optical wavelength conversion, dispersion managed solitons, semiconductor optical amplifier, cross gain modultation

Procedia PDF Downloads 446
1897 Assessment of the Neuroprotective Effect of Oral Hypoglycemic Agents in Patients with Acute Ischemic Stroke

Authors: A. Alhusban, M. Alqawasmeh, F. Alfawares

Abstract:

Introduction: Diabetes is a chronic health problem and a major risk factor of stroke. A number of therapeutic modalities exist for diabetes management. It’s still unknown whether the different oral hypoglycemic agents would ameliorate the detrimental effect of diabetes on stroke severity. The objective of this work is to assess the effect of pretreatment with oral hypoglycemic agents, insulin and their combination on stroke severity at presentation. Patients and Methods: Patients admitted to the King Abdullah University Hospital (KAUH)-Jordan with ischemic stroke between January 2015 and December 2016 were evaluated and their comorbid diseases, treatment on admission and their neurologic severity was assessed using the National Institute of Health Stroke Scale (NIHSS) were documented. Stroke severity was compared for non-diabetic patients and diabetic patients treated with different antidiabetic agents. Results: Data from 324 patients with acute stroke was documented. The median age of participants was 69 years. Diabetes was documented in about 50% of the patients. Multinomial regression analysis identified diabetes treatment status as an independent predictor of neurological severity of stroke (p=0.032). Patients treated with oral hypoglycemic agents had a significantly lower NIHSS as compared to nondiabetic patients and insulin treated patients (p < 0.02). The positive effect of oral hypoglycemic agents was blunted by insulin co-treatment. Insulin did not alter the severity of stroke as compared to non-diabetics. Conclusion: Oral hypoglycemic agents may reduce the severity of neurologic deficit of ischemic stroke and may have neuroprotective effect.

Keywords: diabetes, stroke, neuroprotection, oral hypoglycemic agents

Procedia PDF Downloads 154
1896 Marginalized Two-Part Joint Models for Generalized Gamma Family of Distributions

Authors: Mohadeseh Shojaei Shahrokhabadi, Ding-Geng (Din) Chen

Abstract:

Positive continuous outcomes with a substantial number of zero values and incomplete longitudinal follow-up are quite common in medical cost data. To jointly model semi-continuous longitudinal cost data and survival data and to provide marginalized covariate effect estimates, a marginalized two-part joint model (MTJM) has been developed for outcome variables with lognormal distributions. In this paper, we propose MTJM models for outcome variables from a generalized gamma (GG) family of distributions. The GG distribution constitutes a general family that includes approximately all of the most frequently used distributions like the Gamma, Exponential, Weibull, and Log Normal. In the proposed MTJM-GG model, the conditional mean from a conventional two-part model with a three-parameter GG distribution is parameterized to provide the marginal interpretation for regression coefficients. In addition, MTJM-gamma and MTJM-Weibull are developed as special cases of MTJM-GG. To illustrate the applicability of the MTJM-GG, we applied the model to a set of real electronic health record data recently collected in Iran, and we provided SAS code for application. The simulation results showed that when the outcome distribution is unknown or misspecified, which is usually the case in real data sets, the MTJM-GG consistently outperforms other models. The GG family of distribution facilitates estimating a model with improved fit over the MTJM-gamma, standard Weibull, or Log-Normal distributions.

Keywords: marginalized two-part model, zero-inflated, right-skewed, semi-continuous, generalized gamma

Procedia PDF Downloads 168
1895 Effects of Aging on Ultra: Triathlon Performance

Authors: Richard S. Jatau, Kankanala Venkateswarlu, Bulus Kpame

Abstract:

The purpose of this critical review is to find out what is known and what is unknown about the effects of aging on endurance performance, especially on ultra- triathlon performance. It has been shown that among master’s athlete’s peak levels of performance decreased by 50% by age 50 it has also been clearly revealed that age associated atrophy, weakness and fatigability cannot be halted, although year round athletic training can slow down this age associated decline. Studies have further revealed that 30% to 50% decrease in skeletal muscle mass between ages 40 and 80 years, which is accompanied by an equal or even greater decline in strength and power and an increase in muscle weakness and fatigability. Studies on ultra- triathlon athletes revealed that 30 to 39 year old showed fastest time, with athletes in younger and older age groups were slower. It appears that the length of the endurance performance appears to influence age related endurance performance decline in short distance triathlons. A significant decline seems to start at the age of 40 to 50 years, whereas in long distance triathlons this decline seems to start after the age of 65 years. However, it is not clear whether this decline is related in any way to the training methods used, the duration of training, or the frequency of training. It’s also not clear whether the triathlon athletes experience more injuries due to long hours of training. It’s also not clear whether these athletes used performance enhancing drugs to enhance their performance. It’s not also clear whiles there has been tremendous increase in the number of athletes specializing in triathlon. On the basis of our experience and available research evidence we have provided answers to some of these questions. We concluded that aging associated decline in ultra–endurance performance is inevitable although it can be slowed down.

Keywords: aging, triathlon, atrophy, endurance

Procedia PDF Downloads 368
1894 Application of Deep Learning Algorithms in Agriculture: Early Detection of Crop Diseases

Authors: Manaranjan Pradhan, Shailaja Grover, U. Dinesh Kumar

Abstract:

Farming community in India, as well as other parts of the world, is one of the highly stressed communities due to reasons such as increasing input costs (cost of seeds, fertilizers, pesticide), droughts, reduced revenue leading to farmer suicides. Lack of integrated farm advisory system in India adds to the farmers problems. Farmers need right information during the early stages of crop’s lifecycle to prevent damage and loss in revenue. In this paper, we use deep learning techniques to develop an early warning system for detection of crop diseases using images taken by farmers using their smart phone. The research work leads to building a smart assistant using analytics and big data which could help the farmers with early diagnosis of the crop diseases and corrective actions. The classical approach for crop disease management has been to identify diseases at crop level. Recently, ImageNet Classification using the convolutional neural network (CNN) has been successfully used to identify diseases at individual plant level. Our model uses convolution filters, max pooling, dense layers and dropouts (to avoid overfitting). The models are built for binary classification (healthy or not healthy) and multi class classification (identifying which disease). Transfer learning is used to modify the weights of parameters learnt through ImageNet dataset and apply them on crop diseases, which reduces number of epochs to learn. One shot learning is used to learn from very few images, while data augmentation techniques are used to improve accuracy with images taken from farms by using techniques such as rotation, zoom, shift and blurred images. Models built using combination of these techniques are more robust for deploying in the real world. Our model is validated using tomato crop. In India, tomato is affected by 10 different diseases. Our model achieves an accuracy of more than 95% in correctly classifying the diseases. The main contribution of our research is to create a personal assistant for farmers for managing plant disease, although the model was validated using tomato crop, it can be easily extended to other crops. The advancement of technology in computing and availability of large data has made possible the success of deep learning applications in computer vision, natural language processing, image recognition, etc. With these robust models and huge smartphone penetration, feasibility of implementation of these models is high resulting in timely advise to the farmers and thus increasing the farmers' income and reducing the input costs.

Keywords: analytics in agriculture, CNN, crop disease detection, data augmentation, image recognition, one shot learning, transfer learning

Procedia PDF Downloads 112
1893 Knowledge Creation Environment in the Iranian Universities: A Case Study

Authors: Mahdi Shaghaghi, Amir Ghaebi, Fariba Ahmadi

Abstract:

Purpose: The main purpose of the present research is to analyze the knowledge creation environment at a Iranian University (Alzahra University) as a typical University in Iran, using a combination of the i-System and Ba models. This study is necessary for understanding the determinants of knowledge creation at Alzahra University as a typical University in Iran. Methodology: To carry out the present research, which is an applied study in terms of purpose, a descriptive survey method was used. In this study, a combination of the i-System and Ba models has been used to analyze the knowledge creation environment at Alzahra University. i-System consists of 5 constructs including intervention (input), intelligence (process), involvement (process), imagination (process), and integration (output). The Ba environment has three pillars, namely the infrastructure, the agent, and the information. The integration of these two models resulted in 11 constructs which were as follows: intervention (input), infrastructure-intelligence, agent-intelligence, information-intelligence (process); infrastructure-involvement, agent-involvement, information-involvement (process); infrastructure-imagination, agent-imagination, information-imagination (process); and integration (output). These 11 constructs were incorporated into a 52-statement questionnaire and the validity and reliability of the questionnaire were examined and confirmed. The statistical population included the faculty members of Alzahra University (344 people). A total of 181 participants were selected through the stratified random sampling technique. The descriptive statistics, binomial test, regression analysis, and structural equation modeling (SEM) methods were also utilized to analyze the data. Findings: The research findings indicated that among the 11 research constructs, the levels of intervention, information-intelligence, infrastructure-involvement, and agent-imagination constructs were average and not acceptable. The levels of infrastructure-intelligence and information-imagination constructs ranged from average to low. The levels of agent-intelligence and information-involvement constructs were also completely average. The level of infrastructure-imagination construct was average to high and thus was considered acceptable. The levels of agent-involvement and integration constructs were above average and were in a highly acceptable condition. Furthermore, the regression analysis results indicated that only two constructs, viz. the information-imagination and agent-involvement constructs, positively and significantly correlate with the integration construct. The results of the structural equation modeling also revealed that the intervention, intelligence, and involvement constructs are related to the integration construct with the complete mediation of imagination. Discussion and conclusion: The present research suggests that knowledge creation at Alzahra University relatively complies with the combination of the i-System and Ba models. Unlike this model, the intervention, intelligence, and involvement constructs are not directly related to the integration construct and this seems to have three implications: 1) the information sources are not frequently used to assess and identify the research biases; 2) problem finding is probably of less concern at the end of studies and at the time of assessment and validation; 3) the involvement of others has a smaller role in the summarization, assessment, and validation of the research.

Keywords: i-System, Ba model , knowledge creation , knowledge management, knowledge creation environment, Iranian Universities

Procedia PDF Downloads 97
1892 Incidence, Pattern and Risk Factors of Congenial Heart Diseases in Neonates in a Tertiary Care Hospital, Egyptian Study

Authors: Gehan Hussein, Hams Ahmad, Baher Matta, Yasmeen Mansi, Mohamad Fawzi

Abstract:

Background: Congenital heart disease (CHD) is a common problem worldwide with variable incidence in different countries. The exact etiology is unknown, suggested to be multifactorial. We aimed to study the incidence of various CHD in a neonatal intensive care unit (NICU) in a tertiary care hospital in Egypt and the possible associations with variable risk factors. Methods: Prospective study was conducted over a period of one year (2013 /2014) at NICU KasrAlAini School of Medicine, Cairo University. Questionnaire about possible maternal and/or paternal risk factors for CHD, clinical examination, bedside echocardiography were done. Cases were classified into groups: group 1 without CHD and group 2 with CHD. Results: from 723 neonates admitted to NICU, 180 cases were proved to have CHD, 58 % of them were males. patent ductus arteriosus(PDA) was the most common CHD (70%), followed by an atrial septal defect (ASD8%), while Fallot tetralogy and single ventricle were the least common (0.45 %) for each. CHD was found in 30 % of consanguineous parents Maternal age ≥ 35 years at the time of conception was associated with increased incidence of PDA (p= 0.45 %). Maternal diabetes and insulin intake were significantly associated with cases of CHD (p=0.02 &0.001 respectively), maternal hypertension and hypothyroidism were both associated with VSD, but the difference did not reach statistical significance (P=0.36 &0.44respectively). Maternal passive smoking was significantly associated with PDA (p=0.03). Conclusion: The most frequent CHD in the studied population was PDA, followed by ASD. Maternal conditions as diabetes was associated with VSD occurrence.

Keywords: NICU, risk factors, congenital heart disease, echocardiography

Procedia PDF Downloads 180
1891 The Carbon Trading Price and Trading Volume Forecast in Shanghai City by BP Neural Network

Authors: Liu Zhiyuan, Sun Zongdi

Abstract:

In this paper, the BP neural network model is established to predict the carbon trading price and carbon trading volume in Shanghai City. First of all, we find the data of carbon trading price and carbon trading volume in Shanghai City from September 30, 2015 to December 23, 2016. The carbon trading price and trading volume data were processed to get the average value of each 5, 10, 20, 30, and 60 carbon trading price and trading volume. Then, these data are used as input of BP neural network model. Finally, after the training of BP neural network, the prediction values of Shanghai carbon trading price and trading volume are obtained, and the model is tested.

Keywords: Carbon trading price, carbon trading volume, BP neural network model, Shanghai City

Procedia PDF Downloads 344
1890 3D Object Model Reconstruction Based on Polywogs Wavelet Network Parametrization

Authors: Mohamed Othmani, Yassine Khlifi

Abstract:

This paper presents a technique for compact three dimensional (3D) object model reconstruction using wavelet networks. It consists to transform an input surface vertices into signals,and uses wavelet network parameters for signal approximations. To prove this, we use a wavelet network architecture founded on several mother wavelet families. POLYnomials WindOwed with Gaussians (POLYWOG) wavelet families are used to maximize the probability to select the best wavelets which ensure the good generalization of the network. To achieve a better reconstruction, the network is trained several iterations to optimize the wavelet network parameters until the error criterion is small enough. Experimental results will shown that our proposed technique can effectively reconstruct an irregular 3D object models when using the optimized wavelet network parameters. We will prove that an accurateness reconstruction depends on the best choice of the mother wavelets.

Keywords: 3d object, optimization, parametrization, polywog wavelets, reconstruction, wavelet networks

Procedia PDF Downloads 277
1889 Effect of Mica Content in Sand on Site Response Analyses

Authors: Volkan Isbuga, Joman M. Mahmood, Ali Firat Cabalar

Abstract:

This study presents the site response analysis of mica-sand mixtures available in certain parts of the world including Izmir, a highly populated city and located in a seismically active region in western part of Turkey. We performed site response analyses by employing SHAKE, an equivalent linear approach, for the micaceous soil deposits consisting of layers with different amount of mica contents and thicknesses. Dynamic behavior of micaceous sands such as shear modulus reduction and damping ratio curves are input for the ground response analyses. Micaceous sands exhibit a unique dynamic response under a scenario earthquake with a magnitude of Mw=6. Results showed that higher amount of mica caused higher spectral accelerations.

Keywords: micaceous sands, site response, equivalent linear approach, SHAKE

Procedia PDF Downloads 325
1888 A Low-Power, Low-Noise and High Linearity 60 GHz LNA for WPAN Applications

Authors: Noha Al Majid, Said Mazer, Moulhime El Bekkali, Catherine Algani, Mahmoud Mehdi

Abstract:

A low noise figure (NF) and high linearity V-band Low Noise Amplifier (LNA) is reported in this article. The LNA compromises a three-stage cascode configuration. This LNA will be used as a part of a WPAN (Wireless Personal Area Network) receiver in the millimeter-wave band at 60 GHz. It is designed according to the MMIC technology (Monolithic Microwave Integrated Circuit) in PH 15 process from UMS foundry and uses a 0.15 μm GaAs PHEMT (Pseudomorphic High Electron Mobility Transistor). The particularity of this LNA compared to other LNAs in literature is its very low noise figure which is equal to 1 dB and its high linearity (IIP3 is about 22 dB). The LNA consumes 0.24 Watts, achieving a high gain which is about 23 dB, an input return loss better than -10 dB and an output return loss better than -8 dB.

Keywords: low noise amplifier, V-band, MMIC technology, LNA, amplifier, cascode, pseudomorphic high electron mobility transistor (PHEMT), high linearity

Procedia PDF Downloads 503
1887 A Geographical Information System Supported Method for Determining Urban Transformation Areas in the Scope of Disaster Risks in Kocaeli

Authors: Tayfun Salihoğlu

Abstract:

Following the Law No: 6306 on Transformation of Disaster Risk Areas, urban transformation in Turkey found its legal basis. In the best practices all over the World, the urban transformation was shaped as part of comprehensive social programs through the discourses of renewing the economic, social and physical degraded parts of the city, producing spaces resistant to earthquakes and other possible disasters and creating a livable environment. In Turkish practice, a contradictory process is observed. In this study, it is aimed to develop a method for better understanding of the urban space in terms of disaster risks in order to constitute a basis for decisions in Kocaeli Urban Transformation Master Plan, which is being prepared by Kocaeli Metropolitan Municipality. The spatial unit used in the study is the 50x50 meter grids. In order to reflect the multidimensionality of urban transformation, three basic components that have spatial data in Kocaeli were identified. These components were named as 'Problems in Built-up Areas', 'Disaster Risks arising from Geological Conditions of the Ground and Problems of Buildings', and 'Inadequacy of Urban Services'. Each component was weighted and scored for each grid. In order to delimitate urban transformation zones Optimized Outlier Analysis (Local Moran I) in the ArcGIS 10.6.1 was conducted to test the type of distribution (clustered or scattered) and its significance on the grids by assuming the weighted total score of the grid as Input Features. As a result of this analysis, it was found that the weighted total scores were not significantly clustering at all grids in urban space. The grids which the input feature is clustered significantly were exported as the new database to use in further mappings. Total Score Map reflects the significant clusters in terms of weighted total scores of 'Problems in Built-up Areas', 'Disaster Risks arising from Geological Conditions of the Ground and Problems of Buildings' and 'Inadequacy of Urban Services'. Resulting grids with the highest scores are the most likely candidates for urban transformation in this citywide study. To categorize urban space in terms of urban transformation, Grouping Analysis in ArcGIS 10.6.1 was conducted to data that includes each component scores in significantly clustered grids. Due to Pseudo Statistics and Box Plots, 6 groups with the highest F stats were extracted. As a result of the mapping of the groups, it can be said that 6 groups can be interpreted in a more meaningful manner in relation to the urban space. The method presented in this study can be magnified due to the availability of more spatial data. By integrating with other data to be obtained during the planning process, this method can contribute to the continuation of research and decision-making processes of urban transformation master plans on a more consistent basis.

Keywords: urban transformation, GIS, disaster risk assessment, Kocaeli

Procedia PDF Downloads 115
1886 Dynamic Soil Structure Interaction in Buildings

Authors: Shreya Thusoo, Karan Modi, Ankit Kumar Jha, Rajesh Kumar

Abstract:

Since the evolution of computational tools and simulation software, there has been considerable increase in research on Soil Structure Interaction (SSI) to decrease the computational time and increase accuracy in the results. To aid the designer with a proper understanding of the response of structure in different soil types, the presented paper compares the deformation, shear stress, acceleration and other parameters of multi-storey building for a specific input ground motion using Response-spectrum Analysis (RSA) method. The response of all the models of different heights have been compared in different soil types. Finite Element Simulation software, ANSYS, has been used for all the computational purposes. Overall, higher response is observed with SSI, while it increases with decreasing stiffness of soil.

Keywords: soil-structure interaction, response spectrum, analysis, finite element method, multi-storey buildings

Procedia PDF Downloads 472
1885 Comparative Analysis of Water-Based Alumina Nanoparticles with Water-Based Cupric Nanoparticles Past an Exponentially Accelerated Vertical Radiative Riga Plate with Heat Transfer

Authors: Kanayo Kenneth Asogwa

Abstract:

The influence of the flow of nanoparticles in nanofluids across a vertical surface is significant, and its application in medical sciences, engineering, pharmaceutical, and food industries is enormous & widely published. However, the comparative examination of alumina nanoparticles with cupric nanoparticles past a rapid progressive Riga plate remains unknown. Thus, this report investigates water-based alumina and cupric nanoparticles passing through an exponentially accelerated Riga plate. Nanofluids containing copper (II) oxide (CuO) and aluminum oxide (Al2O3) nanoparticles are considered. The Laplace transform technique is used to solve the partial differential equations guiding the flow. The effect of various factors on skin friction coefficient, Nusselt number, velocity and temperature profiles is investigated and reported in tabular and graphical form. The upsurge of Modified Hartmann number and radiative impact improves copper (II) oxide nanofluid compared to aluminum oxide nanofluid due to Lorentz force and since CuO is a better heat conductor. At the same time, heat absorption and reactive species favor a slight decline in Alumina nanofluid than Cupric nanofluid in the thermal and velocity fields. The higher density of Cupric nanofluid is enhanced by increasing nanoparticle volume fraction over Alumina nanofluid with a decline in velocity distribution.

Keywords: alumina, cupric, nanoparticles, water-based

Procedia PDF Downloads 196
1884 Enhancement Performance of Desalination System Using Humidification and Dehumidification Processes

Authors: Zeinab Syed Abdel Rehim

Abstract:

Water shortage is considered as one of the huge problems the world encounter now. Water desalination is considered as one of the more suitable methods governments can use to substitute the increased need for potable water. The humidification-dehumidification process for water desalination is viewed as a promising technique for small capacity production plants. The process has several attraction features which include the use of sustainable energy sources, low technology, and low-temperature dehumidification. A pilot experimental set-up plant was constructed with the conventional HVAC components such as air blower that supplies air to an air duct inside which air preheater, steam injector and cooling coil of a small refrigeration unit are placed. The present work evaluates the characteristics of humidification-dehumidification process for water desalination as a function of air flow rate, total power input and air inlet temperature in order to study the optimum conditions required to produce distilled water.

Keywords: condensation, dehumidification, evaporation, humidification, water desalination

Procedia PDF Downloads 236
1883 Procedural Protocol for Dual Energy Computed Tomography (DECT) Inversion

Authors: Rezvan Ravanfar Haghighi, S. Chatterjee, Pratik Kumar, V. C. Vani, Priya Jagia, Sanjiv Sharma, Susama Rani Mandal, R. Lakshmy

Abstract:

The dual energy computed tomography (DECT) aims at noting the HU(V) values for the sample at two different voltages V=V1, V2 and thus obtain the electron densities (ρe) and effective atomic number (Zeff) of the substance. In the present paper, we aim to obtain a numerical algorithm by which (ρe, Zeff) can be obtained from the HU(100) and HU(140) data, where V=100, 140 kVp. The idea is to use this inversion method to characterize and distinguish between the lipid and fibrous coronary artery plaques.With the idea to develop the inversion algorithm for low Zeff materials, as is the case with non calcified coronary artery plaque, we prepare aqueous samples whose calculated values of (ρe, Zeff) lie in the range (2.65×1023≤ ρe≤ 3.64×1023 per cc ) and (6.80≤ Zeff ≤ 8.90). We fill the phantom with these known samples and experimentally determine HU(100) and HU(140) for the same pixels. Knowing that the HU(V) values are related to the attenuation coefficient of the system, we present an algorithm by which the (ρe, Zeff) is calibrated with respect to (HU(100), HU(140)). The calibration is done with a known set of 20 samples; its accuracy is checked with a different set of 23 known samples. We find that the calibration gives the ρe with an accuracy of ± 4% while Zeff is found within ±1% of the actual value, the confidence being 95%.In this inversion method (ρe, Zeff) of the scanned sample can be found by eliminating the effects of the CT machine and also by ensuring that the determination of the two unknowns (ρe, Zeff) does not interfere with each other. It is found that this algorithm can be used for prediction of chemical characteristic (ρe, Zeff) of unknown scanned materials with 95% confidence level, by inversion of the DECT data.

Keywords: chemical composition, dual-energy computed tomography, inversion algorithm

Procedia PDF Downloads 432
1882 Entropy-Based Multichannel Stationary Measure for Characterization of Non-Stationary Patterns

Authors: J. D. Martínez-Vargas, C. Castro-Hoyos, G. Castellanos-Dominguez

Abstract:

In this work, we propose a novel approach for measuring the stationarity level of a multichannel time-series. This measure is based on a stationarity definition over time-varying spectrum, and it is aimed to quantify the relation between local stationarity (single-channel) and global dynamic behavior (multichannel dynamics). To assess the proposed approach validity, we use a well known EEG-BCI database, that was constructed for separate between motor/imagery tasks. Thus, based on the statement that imagination of movements implies an increase on the EEG dynamics, we use as discriminant features the proposed measure computed over an estimation of the non-stationary components of input time-series. As measure of separability we use a t-student test, and the obtained results evidence that such measure is able to accurately detect the brain areas projected on the scalp where motor tasks are realized.

Keywords: stationary measure, entropy, sub-space projection, multichannel dynamics

Procedia PDF Downloads 405
1881 Using Single Decision Tree to Assess the Impact of Cutting Conditions on Vibration

Authors: S. Ghorbani, N. I. Polushin

Abstract:

Vibration during machining process is crucial since it affects cutting tool, machine, and workpiece leading to a tool wear, tool breakage, and an unacceptable surface roughness. This paper applies a nonparametric statistical method, single decision tree (SDT), to identify factors affecting on vibration in machining process. Workpiece material (AISI 1045 Steel, AA2024 Aluminum alloy, A48-class30 Gray Cast Iron), cutting tool (conventional, cutting tool with holes in toolholder, cutting tool filled up with epoxy-granite), tool overhang (41-65 mm), spindle speed (630-1000 rpm), feed rate (0.05-0.075 mm/rev) and depth of cut (0.05-0.15 mm) were used as input variables, while vibration was the output parameter. It is concluded that workpiece material is the most important parameters for natural frequency followed by cutting tool and overhang.

Keywords: cutting condition, vibration, natural frequency, decision tree, CART algorithm

Procedia PDF Downloads 331
1880 Risk Factors’ Analysis on Shanghai Carbon Trading

Authors: Zhaojun Wang, Zongdi Sun, Zhiyuan Liu

Abstract:

First of all, the carbon trading price and trading volume in Shanghai are transformed by Fourier transform, and the frequency response diagram is obtained. Then, the frequency response diagram is analyzed and the Blackman filter is designed. The Blackman filter is used to filter, and the carbon trading time domain and frequency response diagram are obtained. After wavelet analysis, the carbon trading data were processed; respectively, we got the average value for each 5 days, 10 days, 20 days, 30 days, and 60 days. Finally, the data are used as input of the Back Propagation Neural Network model for prediction.

Keywords: Shanghai carbon trading, carbon trading price, carbon trading volume, wavelet analysis, BP neural network model

Procedia PDF Downloads 385