Search results for: adaptive instance normalization
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1866

Search results for: adaptive instance normalization

1176 Alternate Methods to Visualize 2016 U.S. Presidential Election Result

Authors: Hong Beom Hur

Abstract:

Politics in America is polarized. The best illustration of this is the 2016 presidential election result map. States with megacities like California, New York, Illinois, Virginia, and others are marked blue to signify the color of the Democratic party. States located in inland and south like Texas, Florida, Tennesse, Kansas and others are marked red to signify the color of the Republican party. Such a stark difference between two colors, red and blue, combined with geolocations of each state with their borderline remarks one central message; America is divided into two colors between urban Democrats and rural Republicans. This paper seeks to defy the visualization by pointing out its limitations and search for alternative ways to visualize the 2016 election result. One such limitation is that geolocations of each state and state borderlines limit the visualization of population density. As a result, the election result map does not convey the fact that Clinton won the popular vote and only accentuates the voting patterns of urban and rural states. The paper seeks whether an alternative narrative can be observed by factoring in the population number into the size of each state and manipulating the state borderline according to the normalization. Yet another alternative narrative may be reached by factoring the size of each state by the number of the electoral college of each state by voting and visualize the number. Other alternatives will be discussed but are not implemented in visualization. Such methods include dividing the land of America into about 120 million cubes each representing a voter or by the number of whole population 300 million cubes. By exploring these alternative methods to visualize the politics of the 2016 election map, the public may be able to question whether it is possible to be free from the narrative of the divide-conquer when interpreting the election map and to look at both parties as a story of the United States of America.

Keywords: 2016 U.S. presidential election, data visualization, population scale, geo-political

Procedia PDF Downloads 119
1175 Bronchospasm Analysis Following the Implementation of a Program of Maximum Aerobic Exercise in Active Men

Authors: Sajjad Shojaeidoust, Mohsen Ghanbarzadeh, Abdolhamid Habibi

Abstract:

Exercise-induced bronchospasm (EIB) is a transitory condition of airflow obstruction that is associated with physical activities. It is noted that high ventilation can lead to an increase in the heat and reduce in the moisture in airways resistance of trachea. Also causes of pathophysiological mechanism are EIB. Accordingly, studying some parameters of pulmonary function (FVC, FEV1) among active people seems quintessential. The aim of this study was to analyze bronchospasm following the implementation of a program of maximum aerobic exercise in active men at Chamran University of Ahwaz. Method: In this quasi-experimental study, the population consisted of all students at Chamran University. Among from 55 participants, of which, 15 were randomly selected as the experimental group. In this study, the size of the maximum oxygen consumption was initially measured, and then, based on the maximum oxygen consumed, the active individuals were identified. After five minutes’ warm-up, Strand treadmill exercise test was taken (one session) and pulmonary parameters were measured at both pre- and post-tests (spirometer). After data normalization using KS and non-normality of the data, the Wilcoxon test was used to analyze the data. The significance level for all statistical surveys was considered p≤0/05. Results: The results showed that the ventilation factors and bronchospasm (FVC, FEV1) in the pre-test and post-test resulted in no significant difference among the active people (p≥0/05). Discussion and conclusion: Based on the results observed in this study, it appears that pulmonary indices in active individuals increased after aerobic test. The increase in this indicator in active people is due to increased volume and elasticity of the lungs as well. In other words, pulmonary index is affected by rib muscles. It is considered that progress over respiratory muscle strength and endurance has raised FEV1 in the active cases.

Keywords: aerobic active maximum, bronchospasm, pulmonary function, spirometer

Procedia PDF Downloads 286
1174 An Active Rectifier with Time-Domain Delay Compensation to Enhance the Power Conversion Efficiency

Authors: Shao-Ku Kao

Abstract:

This paper presents an active rectifier with time-domain delay compensation to enhance the efficiency. A delay calibration circuit is designed to convert delay time to voltage and adaptive control on/off delay in variable input voltage. This circuit is designed in 0.18 mm CMOS process. The input voltage range is from 2 V to 3.6 V with the output voltage from 1.8 V to 3.4 V. The efficiency can maintain more than 85% when the load from 50 Ω ~ 1500 Ω for 3.6 V input voltage. The maximum efficiency is 92.4 % at output power to be 38.6 mW for 3.6 V input voltage.

Keywords: wireless power transfer, active diode, delay compensation, time to voltage converter, PCE

Procedia PDF Downloads 276
1173 Complexity Leadership and Knowledge Management in Higher Education

Authors: Prabhakar Venugopal G.

Abstract:

Complex environments triggered by globalization have necessitated new paradigms of leadership – complexity leadership that encompasses multiple roles that leaders need to take upon. The success of higher education institutions depends on how well leaders can provide adaptive, administrative and enabling leadership. Complexity leadership seems all the more relevant for institutions that are knowledge-driven and thrive on knowledge creation, knowledge storage and retrieval, knowledge sharing and knowledge applications. In this paper are the elements of globalization, the opportunities and challenges that are brought forth by globalization are discussed. The complexity leadership paradigm in a knowledge-based economy and the need for such a paradigm shift for higher education institutions is presented. Further, the paper also discusses the support the leader requires in a knowledge-driven economy through knowledge management initiatives.

Keywords: globalization, complexity leadership, knowledge management

Procedia PDF Downloads 488
1172 Domain-Specific Deep Neural Network Model for Classification of Abnormalities on Chest Radiographs

Authors: Nkechinyere Joy Olawuyi, Babajide Samuel Afolabi, Bola Ibitoye

Abstract:

This study collected a preprocessed dataset of chest radiographs and formulated a deep neural network model for detecting abnormalities. It also evaluated the performance of the formulated model and implemented a prototype of the formulated model. This was with the view to developing a deep neural network model to automatically classify abnormalities in chest radiographs. In order to achieve the overall purpose of this research, a large set of chest x-ray images were sourced for and collected from the CheXpert dataset, which is an online repository of annotated chest radiographs compiled by the Machine Learning Research Group, Stanford University. The chest radiographs were preprocessed into a format that can be fed into a deep neural network. The preprocessing techniques used were standardization and normalization. The classification problem was formulated as a multi-label binary classification model, which used convolutional neural network architecture to make a decision on whether an abnormality was present or not in the chest radiographs. The classification model was evaluated using specificity, sensitivity, and Area Under Curve (AUC) score as the parameter. A prototype of the classification model was implemented using Keras Open source deep learning framework in Python Programming Language. The AUC ROC curve of the model was able to classify Atelestasis, Support devices, Pleural effusion, Pneumonia, A normal CXR (no finding), Pneumothorax, and Consolidation. However, Lung opacity and Cardiomegaly had a probability of less than 0.5 and thus were classified as absent. Precision, recall, and F1 score values were 0.78; this implies that the number of False Positive and False Negative is the same, revealing some measure of label imbalance in the dataset. The study concluded that the developed model is sufficient to classify abnormalities present in chest radiographs into present or absent.

Keywords: transfer learning, convolutional neural network, radiograph, classification, multi-label

Procedia PDF Downloads 120
1171 Biologically Inspired Small Infrared Target Detection Using Local Contrast Mechanisms

Authors: Tian Xia, Yuan Yan Tang

Abstract:

In order to obtain higher small target detection accuracy, this paper presents an effective algorithm inspired by the local contrast mechanism. The proposed method can enhance target signal and suppress background clutter simultaneously. In the first stage, a enhanced image is obtained using the proposed Weighted Laplacian of Gaussian. In the second stage, an adaptive threshold is adopted to segment the target. Experimental results on two changeling image sequences show that the proposed method can detect the bright and dark targets simultaneously, and is not sensitive to sea-sky line of the infrared image. So it is fit for IR small infrared target detection.

Keywords: small target detection, local contrast, human vision system, Laplacian of Gaussian

Procedia PDF Downloads 464
1170 Impairments Correction of Six-Port Based Millimeter-Wave Radar

Authors: Dan Ohev Zion, Alon Cohen

Abstract:

In recent years, the presence of short-range millimeter-wave radar in civil application has increased significantly. Autonomous driving, security, 3D imaging and high data rate communication systems are a few examples. The next challenge is the integration inside small form-factor devices, such as smartphones (e.g. gesture recognition). The main challenge is implementation of a truly low-power, low-complexity high-resolution radar. The most popular approach is the Frequency Modulated Continuous Wave (FMCW) radar, with an analog multiplication front-end. In this paper, we present an approach for adaptive estimation and correction of impairments of such front-end, specifically implemented using the Six-Port Device (SPD) as the multiplier element. The proposed algorithm was simulated and implemented on a 60 GHz radar lab prototype.

Keywords: radar, FMCW Radar, IQ mismatch, six port

Procedia PDF Downloads 150
1169 Vibration Control of a Flexible Structure Using MFC Actuator

Authors: Jinsiang Shaw, Jeng-Jie Huang

Abstract:

Active vibration control is good for low frequency excitation, with advantages of light weight and adaptability. This paper employs a macro-fiber composite (MFC) actuator for vibration suppression in a cantilevered beam due to its higher output force to reject the disturbance. A notch filter with an adaptive tuning algorithm, the leaky filtered-X least mean square algorithm (leaky FXLMS algorithm), is developed and applied to the system. Experimental results show that the controller and MFC actuator was very effective in attenuating the structural vibration. Furthermore, this notch filter controller was compared with the traditional skyhook controller. It was found that its performance was better, with over 88% vibration suppression near the first resonant frequency of the structure.

Keywords: macro-fiber composite, notch filter, skyhook controller, vibration suppression

Procedia PDF Downloads 460
1168 Becoming Vegan: The Theory of Planned Behavior and the Moderating Effect of Gender

Authors: Estela Díaz

Abstract:

This article aims to make three contributions. First, build on the literature on ethical decision-making literature by exploring factors that influence the intention of adopting veganism. Second, study the superiority of extended models of the Theory of Planned Behavior (TPB) for understanding the process involved in forming the intention of adopting veganism. Third, analyze the moderating effect of gender on TPB given that attitudes and behavior towards animals are gender-sensitive. No study, to our knowledge, has examined these questions. Veganism is not a diet but a political and moral stand that exclude, for moral reasons, the use of animals. Although there is a growing interest in studying veganism, it continues being overlooked in empirical research, especially within the domain of social psychology. TPB has been widely used to study a broad range of human behaviors, including moral issues. Nonetheless, TPB has rarely been applied to examine ethical decisions about animals and, even less, to veganism. Hence, the validity of TPB in predicting the intention of adopting veganism remains unanswered. A total of 476 non-vegan Spanish university students (55.6% female; the mean age was 23.26 years, SD= 6.1) responded to online and pencil-and-paper self-reported questionnaire based on previous studies. TPB extended models incorporated two background factors: ‘general attitudes towards humanlike-attributes ascribed to animals’ (AHA) (capacity for reason/emotions/suffer, moral consideration, and affect-towards-animals); and ‘general attitudes towards 11 uses of animals’ (AUA). SPSS 22 and SmartPLS 3.0 were used for statistical analyses. This study constructed a second-order reflective-formative model and took the multi-group analysis (MGA) approach to study gender effects. Six models of TPB (the standard and five competing) were tested. No a priori hypotheses were formulated. The results gave partial support to TPB. Attitudes (ATTV) (β = .207, p < .001), subjective norms (SNV) (β = .323, p < .001), and perceived control behavior (PCB) (β = .149, p < .001) had a significant direct effect on intentions (INTV). This model accounted for 27,9% of the variance in intention (R2Adj = .275) and had a small predictive relevance (Q2 = .261). However, findings from this study reveal that contrary to what TPB generally proposes, the effect of the background factors on intentions was not fully mediated by the proximal constructs of intentions. For instance, in the final model (Model#6), both factors had significant multiple indirect effect on INTV (β = .074, 95% C = .030, .126 [AHA:INTV]; β = .101, 95% C = .055, .155 [AUA:INTV]) and significant direct effect on INTV (β = .175, p < .001 [AHA:INTV]; β = .100, p = .003 [AUA:INTV]). Furthermore, the addition of direct paths from background factors to intentions improved the explained variance in intention (R2 = .324; R2Adj = .317) and the predictive relevance (Q2 = .300) over the base-model. This supports existing literature on the superiority of enhanced TPB models to predict ethical issues; which suggests that moral behavior may add additional complexity to decision-making. Regarding gender effect, MGA showed that gender only moderated the influence of AHA on ATTV (e.g., βWomen−βMen = .296, p < .001 [Model #6]). However, other observed gender differences (e.g. the explained variance of the model for intentions were always higher for men that for women, for instance, R2Women = .298; R2Men = .394 [Model #6]) deserve further considerations, especially for developing more effective communication strategies.

Keywords: veganism, Theory of Planned Behavior, background factors, gender moderation

Procedia PDF Downloads 344
1167 Achieving Social Sustainability through Architectural Designs for Physically Challenged People: Datascapes Technique

Authors: Fatemeh Zare, Kaveh Bazrafkan, Alireza Bolhari

Abstract:

Quality of life is one of the most recent issues in today's architectural world. It has numerous criteria and has diverse aspects in different nation's cultures. Social sustainability, on the other hand, is frequently a positive attitude which is manifested by integration of human beings and equity of access to fundamental amenities; for instance, transportation, hygienic systems, equal education facilities, etc. This paper demonstrates that achieving desired quality of life is through assurance of sustainable society. Choosing a sustainable approach in every day's life becomes a practical manner and solution for human life. By assuming that an architect is someone who designs people's life by his/her projects, scrutinizing the relationship between quality of life and architectural buildings would reveal hidden criteria through Datascapes technique. This would be enriched when considering this relationship with everyone's basic needs in the society. One the most impressive needs are the particular demands of physically challenged people which are directly examined and discussed.

Keywords: sustainable design, social sustainability, disabled people, datascapes technique

Procedia PDF Downloads 478
1166 Examining the Racialisation of White Workers in Rural Louisiana as a Technology of Capitalist Management and Control

Authors: Kendall Artz

Abstract:

In the 1950s, a wave of violent labor unrest shook a small town in south-western Louisiana leading to the racialisation of workers—previously considered white—as ‘mixed-race’ or, in local terms, ‘Redbone.’ This paper examines why the group known as ‘Redbones’ were marked as non-white in relation to strike violence and their opposition to capitalist expansion. Utilising archival research, historiography and oral testimony, I examine how an instance of labor unrest was reinterpreted by local law enforcement, an interstate capitalist class and the national press as calling into question the racial integrity of a group of workers who had been formerly marked as white. This explosive and largely unstudied strike provides an opportunity to better understand how racialisation operates as a technology of control, even over individuals who appear phenotypically white. The strike at Elizabeth allows a glimpse at the tactics of representatives of white supremacy when white workers do not fully embrace the ‘wages of whiteness.

Keywords: American federation of labor, labor history, Louisiana history, wages of whiteness

Procedia PDF Downloads 185
1165 Combined Localization, Beamforming, and Interference Threshold Estimation in Underlay Cognitive System

Authors: Omar Nasr, Yasser Naguib, Mohamed Hafez

Abstract:

This paper aims at providing an innovative solution for blind interference threshold estimation in an underlay cognitive network to be used in adaptive beamforming by secondary user Transmitter and Receiver. For the task of threshold estimation, blind detection of modulation and SNR are used. For the sake of beamforming several localization algorithms are compared to settle on best one for cognitive environment. Beamforming algorithms as LCMV (Linear Constraint Minimum Variance) and MVDR (Minimum Variance Distortion less) are also proposed and compared. The idea of just nulling the primary user after knowledge of its location is discussed against the idea of working under interference threshold.

Keywords: cognitive radio, underlay, beamforming, MUSIC, MVDR, LCMV, threshold estimation

Procedia PDF Downloads 578
1164 The Effect of TiO₂ Nano-Thin Films on Light Transmission and Self-Cleaning Capabilities of Glass Surface

Authors: Ahmad Alduweesh

Abstract:

Self-cleaning surfaces have become essential in various applications. For instance, in photovoltaics, they provide an easy-cost effecting way to keep the solar cells clean. Titanium dioxide (TiO₂) nanoparticles were fabricated at different thicknesses to study the effect of different thicknesses on the hydrophilicity behavior of TiO₂, eventually leading to customizing hydrophilicity levels to desired values under natural light. As a result, a remarkable increase was noticed in surface hydrophilicity after applying thermal annealing on the as-deposited TiO₂ thin-films, with contact angle dropping from around 85.4ᵒ for as-deposited thin-films down to 5.1ᵒ for one of the annealed samples. The produced thin films were exposed to the outside environment to observe the effect of dust. The transmittance of light using UV-VIS spectroscopy will be conducted on the lowest and highest thicknesses (5-40 nm); this will show whether the Titania has successfully enabled more sunlight to penetrate the glass or not. Surface characterizations, including AFM and contact angle, have been included in this test.

Keywords: physical vapor deposition, TiO₂, nano-thin films, hydrophobicity, hydrophilicity, self-cleaning surfaces

Procedia PDF Downloads 111
1163 Multiscale Connected Component Labelling and Applications to Scientific Microscopy Image Processing

Authors: Yayun Hsu, Henry Horng-Shing Lu

Abstract:

In this paper, a new method is proposed to extending the method of connected component labeling from processing binary images to multi-scale modeling of images. By using the adaptive threshold of multi-scale attributes, this approach minimizes the possibility of missing those important components with weak intensities. In addition, the computational cost of this approach remains similar to that of the typical approach of component labeling. Then, this methodology is applied to grain boundary detection and Drosophila Brain-bow neuron segmentation. These demonstrate the feasibility of the proposed approach in the analysis of challenging microscopy images for scientific discovery.

Keywords: microscopic image processing, scientific data mining, multi-scale modeling, data mining

Procedia PDF Downloads 433
1162 Evaluation of Anti-Arthritic Activity of Eulophia ochreata Lindl and Zingiber cassumunar Roxb in Freund's Complete Adjuvant Induced Arthritic Rat Model

Authors: Akshada Amit Koparde, Candrakant S. Magdum

Abstract:

Objective: To investigate the anti-arthritic activity of chloroform extract and Isolate 1 of Eulophia ochreata Lindl and dichloromethane extract and Isolate 2 of Zingiber cassumunar Roxb in adjuvant arthritic (AA) rat model induced by Freund’s complete adjuvant (FCA). Methods: Forty two healthy albino rats were selected and randomly divided into six groups. Freund’s complete adjuvant (FCA) was used to induce arthritis and then treated with chloroform extract, isolate 1 and dichloromethane extract, isolate 2 for 28 days. The various parameters like paw volume, haematological parameters (RBC, WBC, Hb and ESR), were studied. Structural elucidation of active constituents isolate 1 and isolate 2 from Eulophia ochreata Lindl and Zingiber cassumunar Roxb will be done using GCMS and H1NMR. Results: In FCA induced arthritic rats, there was significant increase in rat paw volume whereas chloroform extract and Isolate 1 of Eulophia ochreata Lindl and dichloromethane extract and Isolate 2 of Zingiber cassumunar Roxb treated groups showed strong significant reduction in paw volume. The altered haematological parameters in the arthritic rats were significantly recovered to near normal by the treatment with extracts at the dose of 200 mg/kg. Further histopathological studies revealed the anti-arthritic activity of Eulophia ochreata Lindl and Zingiber cassumunar Roxb by preventing cartilage and bone destruction of the arthritic joints of AA rats. Conclusion: Extracts and isolates of Eulophia ochreata Lindl and Zingiber cassumunar Roxb have shown anti-arthritic activity. Decrease in paw volume and normalization of haematological abnormalities in adjuvant induced arthritic rats is significantly seen in the experiment. Further histopathological studies confirmed the anti-arthritic activity of Eulophia ochreata Lindl and Zingiber cassumunar Roxb.

Keywords: arthritis, Eulophia ochreata Lindl, Freund's complete adjuvant, paw volume, Zingiber cassumunar Roxb

Procedia PDF Downloads 172
1161 Rough Neural Networks in Adapting Cellular Automata Rule for Reducing Image Noise

Authors: Yasser F. Hassan

Abstract:

The reduction or removal of noise in a color image is an essential part of image processing, whether the final information is used for human perception or for an automatic inspection and analysis. This paper describes the modeling system based on the rough neural network model to adaptive cellular automata for various image processing tasks and noise remover. In this paper, we consider the problem of object processing in colored image using rough neural networks to help deriving the rules which will be used in cellular automata for noise image. The proposed method is compared with some classical and recent methods. The results demonstrate that the new model is capable of being trained to perform many different tasks, and that the quality of these results is comparable or better than established specialized algorithms.

Keywords: rough sets, rough neural networks, cellular automata, image processing

Procedia PDF Downloads 435
1160 Exploring a Teaching Method for Elementary Students to Promote Cross-Cultural Understanding: Utilizing an American Film

Authors: Mikako Nobuhara

Abstract:

This study explores the effective methods of nurturing elementary students’ cross-cultural understanding. The delivery lecture was conducted in a private elementary school class for understanding cross-cultural differences through the film E.T. (1982). Interviews of care supporters and students were conducted, as well as student discussions were held after the class. The results were carefully observed and analyzed. Suitable findings were obtained, for instance, students’ listening skills improved; further, they deeply thought about the main character’s feelings after watching the movie. Moreover, their interest in studying English as a foreign language increased. In conclusion, more classes where students can express their opinions in front of the class need to be offered; this would enable the students to nurture their critical thinking abilities and build a sense of accomplishment when they are in elementary school. Utilizing films is one of the best ways to provide students good opportunities to engage in discussions on a specific theme. This is particularly true for elementary school students.

Keywords: cross-cultural understanding, English education, elementary schools, films

Procedia PDF Downloads 160
1159 Selection of Optimal Reduced Feature Sets of Brain Signal Analysis Using Heuristically Optimized Deep Autoencoder

Authors: Souvik Phadikar, Nidul Sinha, Rajdeep Ghosh

Abstract:

In brainwaves research using electroencephalogram (EEG) signals, finding the most relevant and effective feature set for identification of activities in the human brain is a big challenge till today because of the random nature of the signals. The feature extraction method is a key issue to solve this problem. Finding those features that prove to give distinctive pictures for different activities and similar for the same activities is very difficult, especially for the number of activities. The performance of a classifier accuracy depends on this quality of feature set. Further, more number of features result in high computational complexity and less number of features compromise with the lower performance. In this paper, a novel idea of the selection of optimal feature set using a heuristically optimized deep autoencoder is presented. Using various feature extraction methods, a vast number of features are extracted from the EEG signals and fed to the autoencoder deep neural network. The autoencoder encodes the input features into a small set of codes. To avoid the gradient vanish problem and normalization of the dataset, a meta-heuristic search algorithm is used to minimize the mean square error (MSE) between encoder input and decoder output. To reduce the feature set into a smaller one, 4 hidden layers are considered in the autoencoder network; hence it is called Heuristically Optimized Deep Autoencoder (HO-DAE). In this method, no features are rejected; all the features are combined into the response of responses of the hidden layer. The results reveal that higher accuracy can be achieved using optimal reduced features. The proposed HO-DAE is also compared with the regular autoencoder to test the performance of both. The performance of the proposed method is validated and compared with the other two methods recently reported in the literature, which reveals that the proposed method is far better than the other two methods in terms of classification accuracy.

Keywords: autoencoder, brainwave signal analysis, electroencephalogram, feature extraction, feature selection, optimization

Procedia PDF Downloads 111
1158 WebAppShield: An Approach Exploiting Machine Learning to Detect SQLi Attacks in an Application Layer in Run-time

Authors: Ahmed Abdulla Ashlam, Atta Badii, Frederic Stahl

Abstract:

In recent years, SQL injection attacks have been identified as being prevalent against web applications. They affect network security and user data, which leads to a considerable loss of money and data every year. This paper presents the use of classification algorithms in machine learning using a method to classify the login data filtering inputs into "SQLi" or "Non-SQLi,” thus increasing the reliability and accuracy of results in terms of deciding whether an operation is an attack or a valid operation. A method Web-App auto-generated twin data structure replication. Shielding against SQLi attacks (WebAppShield) that verifies all users and prevents attackers (SQLi attacks) from entering and or accessing the database, which the machine learning module predicts as "Non-SQLi" has been developed. A special login form has been developed with a special instance of data validation; this verification process secures the web application from its early stages. The system has been tested and validated, up to 99% of SQLi attacks have been prevented.

Keywords: SQL injection, attacks, web application, accuracy, database

Procedia PDF Downloads 148
1157 A Study on Impact of Corporate Social Responsibility on Rural Development

Authors: N. Amruth Raj, Suja S. Nair

Abstract:

The last six decades have borne witness to a radical change in the private sectors relationship with both the state and civil society. Firms have been increasingly called upon to adopt strategies beyond the financial aspects of their operations and consider the social and environmental impact of their business activities. In this context, many companies have modified their policies and activities and engaged into Corporate Social Responsibility (CSR) especially on Rural development in India. At the firm level, CSR is implemented through various practices, which aim to enhance the company’s social and environmental performance and may cover various topics. Examples of CSR practices are abundant in Andhra Pradesh relevant literature. For instance, in India especially at Andhra Pradesh companies like Amara Raaja requires from its suppliers to prohibit child labour, Nagarjuna Cements applies a series of programs for reducing its CO2 emissions, LANCO group of Industries addresses health and safety issues in the workplace whereas GVK works limited has adopted a series of policies for addressing human rights and environmental abuse related to its operations.

Keywords: CSR, limitations, need, objectives, rural development

Procedia PDF Downloads 255
1156 Optimized Cluster Head Selection Algorithm Based on LEACH Protocol for Wireless Sensor Networks

Authors: Wided Abidi, Tahar Ezzedine

Abstract:

Low-Energy Adaptive Clustering Hierarchy (LEACH) has been considered as one of the effective hierarchical routing algorithms that optimize energy and prolong the lifetime of network. Since the selection of Cluster Head (CH) in LEACH is carried out randomly, in this paper, we propose an approach of electing CH based on LEACH protocol. In other words, we present a formula for calculating the threshold responsible for CH election. In fact, we adopt three principle criteria: the remaining energy of node, the number of neighbors within cluster range and the distance between node and CH. Simulation results show that our proposed approach beats LEACH protocol in regards of prolonging the lifetime of network and saving residual energy.

Keywords: wireless sensors networks, LEACH protocol, cluster head election, energy efficiency

Procedia PDF Downloads 328
1155 Causality, Special Relativity and Non-existence of Material Particles of Zero Rest Mass

Authors: Mohammad Saleem, Mujahid Kamran

Abstract:

It is shown that causality, the principle that cause must precede effect, leads inter alia, to highly significant result that the velocity of a material particle cannot be even equal to that of light. Consequently, combined with special relativity, it leads to the conclusion that material particles of zero rest mass cannot exist in nature. Thus, causality, a principle without which nature would be incomprehensible, combined with special relativity, forbids the existence of material particles of zero rest mass. For instance, the neutrinos, as is now known, are material particles of non-zero rest mass. The situation changes when we consider the gauge particles. In fact, when the principle of causality was proposed, the concept of gauge particles had not yet been introduced. Now we know that photon, a gauge particle with zero rest mass does exist in nature. Therefore, principle of causality, as generally stated, is valid only for material particles. For gauge particles, in order to make the statement of causality consistent with experiment, it has to be modified: The cause should either precede or be simultaneous with the effect. Combined with special relativity, it allows gauge particles of zero rest mass.

Keywords: causality, gauge particles, material particles, special relativity

Procedia PDF Downloads 498
1154 A Comparison of Convolutional Neural Network Architectures for the Classification of Alzheimer’s Disease Patients Using MRI Scans

Authors: Tomas Premoli, Sareh Rowlands

Abstract:

In this study, we investigate the impact of various convolutional neural network (CNN) architectures on the accuracy of diagnosing Alzheimer’s disease (AD) using patient MRI scans. Alzheimer’s disease is a debilitating neurodegenerative disorder that affects millions worldwide. Early, accurate, and non-invasive diagnostic methods are required for providing optimal care and symptom management. Deep learning techniques, particularly CNNs, have shown great promise in enhancing this diagnostic process. We aim to contribute to the ongoing research in this field by comparing the effectiveness of different CNN architectures and providing insights for future studies. Our methodology involved preprocessing MRI data, implementing multiple CNN architectures, and evaluating the performance of each model. We employed intensity normalization, linear registration, and skull stripping for our preprocessing. The selected architectures included VGG, ResNet, and DenseNet models, all implemented using the Keras library. We employed transfer learning and trained models from scratch to compare their effectiveness. Our findings demonstrated significant differences in performance among the tested architectures, with DenseNet201 achieving the highest accuracy of 86.4%. Transfer learning proved to be helpful in improving model performance. We also identified potential areas for future research, such as experimenting with other architectures, optimizing hyperparameters, and employing fine-tuning strategies. By providing a comprehensive analysis of the selected CNN architectures, we offer a solid foundation for future research in Alzheimer’s disease diagnosis using deep learning techniques. Our study highlights the potential of CNNs as a valuable diagnostic tool and emphasizes the importance of ongoing research to develop more accurate and effective models.

Keywords: Alzheimer’s disease, convolutional neural networks, deep learning, medical imaging, MRI

Procedia PDF Downloads 70
1153 Energy Saving Techniques for MIMO Decoders

Authors: Zhuofan Cheng, Qiongda Hu, Mohammed El-Hajjar, Basel Halak

Abstract:

Multiple-input multiple-output (MIMO) systems can allow significantly higher data rates compared to single-antenna-aided systems. They are expected to be a prominent part of the 5G communication standard. However, these decoders suffer from high power consumption. This work presents a design technique in order to improve the energy efficiency of MIMO systems; this facilitates their use in the next generation of battery-operated communication devices such as mobile phones and tablets. The proposed optimization approach consists of the use of low complexity lattice reduction algorithm in combination with an adaptive VLSI implementation. The proposed design has been realized and verified in 65nm technology. The results show that the proposed design is significantly more energy-efficient than conventional K-best MIMO systems.

Keywords: energy, lattice reduction, MIMO, VLSI

Procedia PDF Downloads 325
1152 A New Cytoprotective Drug on the Basis of Cytisine: Phase I Clinical Trial Results

Authors: B. Yermekbayeva, A. Gulyayaev, T. Nurgozhin, C. Bektur

Abstract:

Cytisine aminophosphonate under the name "Cytafat" was approved for clinical trials in Republic of Kazakhstan as a putative liver protecting drug for the treatment of acute toxic hepatitis. A method of conducting the clinical trial is a double blind study. Total number of patients -71, aged from 16 to 56 years. Research on healthy volunteers determined the maximal tolerable doze of "Cytafat" as 200 mg/kg. Side effects when administered at high dozes (100-200 mg/kg) are tachycardia and increase of arterial blood pressure. The drug is tested in the treatment of 28 patients with a syndrome of hepatocellular failure (a poisoning with substitutes of alcohol, rat poison, or medical products). "Cytafat" was intravenously administered at a dose of 10 mg/kg in 200 ml of 5 % glucose solution once daily. The number of administrations: 1-3. In the comparison group, 23 patients were treated intravenously once a day with “Essenciale H” at a dose of 10 ml. 20 patients received a placebo (10 ml of glucose intravenously). In all cases of toxic hepatopathology the significant positive clinical effect of the testing drug distinguishable from placebo and surpassing the alternative was observed. Within a day after administration a sharp reduction of cytolitic syndrome parameters (ALT, AST, alkaline phosphatase, thymol turbidity test, GGT) was registered, a reduction of the severity of cholestatic syndrome (bilirubin decreased) was recorded, significantly decreased indices of lipid peroxidation. The following day, in all cases the positive dynamics was determined with ultrasound study (reduction of diffuse changes and events of reactive pancreatitis), hepatomegaly disappeared. Normalization of all parameters occurred in 2-3 times faster, than when using the drug "Essenciale H" and placebo. Average term of elimination of toxic hepatopathy when using the drug "Cytafat" -2,8 days, "Essenciale H" -7,2 days, and placebo -10,6 days. The new drug "Cytafat" has expressed cytoprotective properties.

Keywords: cytisine, cytoprotection, hepatopathy, hepatoprotection

Procedia PDF Downloads 360
1151 Experimental Analysis of Control in Electric Vehicle Charging Station Based Grid Tied Photovoltaic-Battery System

Authors: A. Hassoune, M. Khafallah, A. Mesbahi, T. Bouragba

Abstract:

This work presents an improved strategy of control for charging a lithium-ion battery in an electric vehicle charging station using two charger topologies i.e. single ended primary inductor converter (SEPIC) and forward converter. In terms of rapidity and accuracy, the power system consists of a topology/control diagram that would overcome the performance constraints, for instance the power instability, the battery overloading and how the energy conversion blocks would react efficiently to any kind of perturbations. Simulation results show the effectiveness of the proposed topologies operated with a power management algorithm based on voltage/peak current mode controls. In order to provide credible findings, a low power prototype is developed to test the control strategy via experimental evaluations of the converter topology and its controls.

Keywords: battery storage buffer, charging station, electric vehicle, experimental analysis, management algorithm, switches control

Procedia PDF Downloads 162
1150 A Mathematical Model to Select Shipbrokers

Authors: Y. Smirlis, G. Koronakos, S. Plitsos

Abstract:

Shipbrokers assist the ship companies in chartering or selling and buying vessels, acting as intermediates between them and the market. They facilitate deals, providing their expertise, negotiating skills, and knowledge about ship market bargains. Their role is very important as it affects the profitability and market position of a shipping company. Due to their significant contribution, the shipping companies have to employ systematic procedures to evaluate the shipbrokers’ services in order to select the best and, consequently, to achieve the best deals. Towards this, in this paper, we consider shipbrokers as financial service providers, and we formulate the problem of evaluating and selecting shipbrokers’ services as a multi-criteria decision making (MCDM) procedure. The proposed methodology comprises a first normalization step to adjust different scales and orientations of the criteria and a second step that includes the mathematical model to evaluate the performance of the shipbrokers’ services involved in the assessment. The criteria along which the shipbrokers are assessed may refer to their size and reputation, the potential efficiency of the services, the terms and conditions imposed, the expenses (e.g., commission – brokerage), the expected time to accomplish a chartering or selling/buying task, etc. and according to our modelling approach these criteria may be assigned different importance. The mathematical programming model performs a comparative assessment and estimates for the shipbrokers involved in the evaluation, a relative score that ranks the shipbrokers in terms of their potential performance. To illustrate the proposed methodology, we present a case study in which a shipping company evaluates and selects the most suitable among a number of sale and purchase (S&P) brokers. Acknowledgment: This study is supported by the OptiShip project, implemented within the framework of the National Recovery Plan and Resilience “Greece 2.0” and funded by the European Union – NextGenerationEU programme.

Keywords: shipbrokers, multi-criteria decision making, mathematical programming, service-provider selection

Procedia PDF Downloads 84
1149 Robust Diagnosis Efficiency by Bond-Graph Approach

Authors: Benazzouz Djamel, Termeche Adel, Touati Youcef, Alem Said, Ouziala Mahdi

Abstract:

This paper presents an approach which detect and isolate efficiently a fault in a system. This approach avoids false alarms, non-detections and delays in detecting faults. A study case have been proposed to show the importance of taking into consideration the uncertainties in the decision-making procedure and their effect on the degradation diagnostic performance and advantage of using Bond Graph (BG) for such degradation. The use of BG in the Linear Fractional Transformation (LFT) form allows generating robust Analytical Redundancy Relations (ARR’s), where the uncertain part of ARR’s is used to generate the residuals adaptive thresholds. The study case concerns an electromechanical system composed of a motor, a reducer and an external load. The aim of this application is to show the effectiveness of the BG-LFT approach to robust fault detection.

Keywords: bond graph, LFT, uncertainties, detection and faults isolation, ARR

Procedia PDF Downloads 303
1148 Applying Epistemology to Artificial Intelligence in the Social Arena: Exploring Fundamental Considerations

Authors: Gianni Jacucci

Abstract:

Epistemology traditionally finds its place within human research philosophies and methodologies. Artificial intelligence methods pose challenges, particularly given the unresolved relationship between AI and pivotal concepts in social arenas such as hermeneutics and accountability. We begin by examining the essential criteria governing scientific rigor in the human sciences. We revisit the three foundational philosophies underpinning qualitative research methods: empiricism, hermeneutics, and phenomenology. We elucidate the distinct attributes, merits, and vulnerabilities inherent in the methodologies they inspire. The integration of AI, e.g., deep learning algorithms, sparks an interest in evaluating these criteria against the diverse forms of AI architectures. For instance, Interpreted AI could be viewed as a hermeneutic approach, relying on a priori interpretations, while straight AI may be perceived as a descriptive phenomenological approach, processing original and uncontaminated data. This paper serves as groundwork for such explorations, offering preliminary reflections to lay the foundation and outline the initial landscape.

Keywords: artificial intelligence, deep learning, epistemology, qualitative research, methodology, hermeneutics, accountability

Procedia PDF Downloads 30
1147 Agile Software Effort Estimation Using Regression Techniques

Authors: Mikiyas Adugna

Abstract:

Effort estimation is among the activities carried out in software development processes. An accurate model of estimation leads to project success. The method of agile effort estimation is a complex task because of the dynamic nature of software development. Researchers are still conducting studies on agile effort estimation to enhance prediction accuracy. Due to these reasons, we investigated and proposed a model on LASSO and Elastic Net regression to enhance estimation accuracy. The proposed model has major components: preprocessing, train-test split, training with default parameters, and cross-validation. During the preprocessing phase, the entire dataset is normalized. After normalization, a train-test split is performed on the dataset, setting training at 80% and testing set to 20%. We chose two different phases for training the two algorithms (Elastic Net and LASSO) regression following the train-test-split. In the first phase, the two algorithms are trained using their default parameters and evaluated on the testing data. In the second phase, the grid search technique (the grid is used to search for tuning and select optimum parameters) and 5-fold cross-validation to get the final trained model. Finally, the final trained model is evaluated using the testing set. The experimental work is applied to the agile story point dataset of 21 software projects collected from six firms. The results show that both Elastic Net and LASSO regression outperformed the compared ones. Compared to the proposed algorithms, LASSO regression achieved better predictive performance and has acquired PRED (8%) and PRED (25%) results of 100.0, MMRE of 0.0491, MMER of 0.0551, MdMRE of 0.0593, MdMER of 0.063, and MSE of 0.0007. The result implies LASSO regression algorithm trained model is the most acceptable, and higher estimation performance exists in the literature.

Keywords: agile software development, effort estimation, elastic net regression, LASSO

Procedia PDF Downloads 63