Search results for: depth-wise separable convolutional neural network for light-weight GAN architecture for EDGE devices
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9532

Search results for: depth-wise separable convolutional neural network for light-weight GAN architecture for EDGE devices

8062 Mathematical Modelling and AI-Based Degradation Analysis of the Second-Life Lithium-Ion Battery Packs for Stationary Applications

Authors: Farhad Salek, Shahaboddin Resalati

Abstract:

The production of electric vehicles (EVs) featuring lithium-ion battery technology has substantially escalated over the past decade, demonstrating a steady and persistent upward trajectory. The imminent retirement of electric vehicle (EV) batteries after approximately eight years underscores the critical need for their redirection towards recycling, a task complicated by the current inadequacy of recycling infrastructures globally. A potential solution for such concerns involves extending the operational lifespan of electric vehicle (EV) batteries through their utilization in stationary energy storage systems during secondary applications. Such adoptions, however, require addressing the safety concerns associated with batteries’ knee points and thermal runaways. This paper develops an accurate mathematical model representative of the second-life battery packs from a cell-to-pack scale using an equivalent circuit model (ECM) methodology. Neural network algorithms are employed to forecast the degradation parameters based on the EV batteries' aging history to develop a degradation model. The degradation model is integrated with the ECM to reflect the impacts of the cycle aging mechanism on battery parameters during operation. The developed model is tested under real-life load profiles to evaluate the life span of the batteries in various operating conditions. The methodology and the algorithms introduced in this paper can be considered the basis for Battery Management System (BMS) design and techno-economic analysis of such technologies.

Keywords: second life battery, electric vehicles, degradation, neural network

Procedia PDF Downloads 42
8061 Predicting Oil Spills in Real-Time: A Machine Learning and AIS Data-Driven Approach

Authors: Tanmay Bisen, Aastha Shayla, Susham Biswas

Abstract:

Oil spills from tankers can cause significant harm to the environment and local communities, as well as have economic consequences. Early predictions of oil spills can help to minimize these impacts. Our proposed system uses machine learning and neural networks to predict potential oil spills by monitoring data from ship Automatic Identification Systems (AIS). The model analyzes ship movements, speeds, and changes in direction to identify patterns that deviate from the norm and could indicate a potential spill. Our approach not only identifies anomalies but also predicts spills before they occur, providing early detection and mitigation measures. This can prevent or minimize damage to the reputation of the company responsible and the country where the spill takes place. The model's performance on the MV Wakashio oil spill provides insight into its ability to detect and respond to real-world oil spills, highlighting areas for improvement and further research.

Keywords: Anomaly Detection, Oil Spill Prediction, Machine Learning, Image Processing, Graph Neural Network (GNN)

Procedia PDF Downloads 54
8060 Enabling Non-invasive Diagnosis of Thyroid Nodules with High Specificity and Sensitivity

Authors: Sai Maniveer Adapa, Sai Guptha Perla, Adithya Reddy P.

Abstract:

Thyroid nodules can often be diagnosed with ultrasound imaging, although differentiating between benign and malignant nodules can be challenging for medical professionals. This work suggests a novel approach to increase the precision of thyroid nodule identification by combining machine learning and deep learning. The new approach first extracts information from the ultrasound pictures using a deep learning method known as a convolutional autoencoder. A support vector machine, a type of machine learning model, is then trained using these features. With an accuracy of 92.52%, the support vector machine can differentiate between benign and malignant nodules. This innovative technique may decrease the need for pointless biopsies and increase the accuracy of thyroid nodule detection.

Keywords: thyroid tumor diagnosis, ultrasound images, deep learning, machine learning, convolutional auto-encoder, support vector machine

Procedia PDF Downloads 37
8059 Method and Experiment of Fabricating and Cutting the Burr for Y Shape Nanochannel

Authors: Zone-Ching Lin, Hao-Yuan Jheng, Shih-Hung Ma

Abstract:

The present paper proposes using atomic force microscopy (AFM) and the concept of specific down force energy (SDFE) to establish a method for fabricating and cutting the burr for Y shape nanochannel on silicon (Si) substrate. For fabricating Y shape nanochannel, it first makes the experimental cutting path planning for fabricating Y shape nanochannel until the fifth cutting layer. Using the constant down force by AFM and SDFE theory and following the experimental cutting path planning, the cutting depth and width of each pass of Y shape nanochannel can be predicted by simulation. The paper plans the path for cutting the burr at the edge of Y shape nanochannel. Then, it carries out cutting the burr along the Y nanochannel edge by using a smaller down force. The height of standing burr at the edge is required to be below the set value of 0.54 nm. The results of simulation and experiment of fabricating and cutting the burr for Y shape nanochannel is further compared.

Keywords: atomic force microscopy (AFM), nanochannel, specific down force energy (SDFE), Y shape, burr, silicon

Procedia PDF Downloads 391
8058 Performance and Emission Prediction in a Biodiesel Engine Fuelled with Honge Methyl Ester Using RBF Neural Networks

Authors: Shiva Kumar, G. S. Vijay, Srinivas Pai P., Shrinivasa Rao B. R.

Abstract:

In the present study RBF neural networks were used for predicting the performance and emission parameters of a biodiesel engine. Engine experiments were carried out in a 4 stroke diesel engine using blends of diesel and Honge methyl ester as the fuel. Performance parameters like BTE, BSEC, Tech and emissions from the engine were measured. These experimental results were used for ANN modeling. RBF center initialization was done by random selection and by using Clustered techniques. Network was trained by using fixed and varying widths for the RBF units. It was observed that RBF results were having a good agreement with the experimental results. Networks trained by using clustering technique gave better results than using random selection of centers in terms of reduced MRE and increased prediction accuracy. The average MRE for the performance parameters was 3.25% with the prediction accuracy of 98% and for emissions it was 10.4% with a prediction accuracy of 80%.

Keywords: radial basis function networks, emissions, performance parameters, fuzzy c means

Procedia PDF Downloads 543
8057 Improving the Performance of Deep Learning in Facial Emotion Recognition with Image Sharpening

Authors: Ksheeraj Sai Vepuri, Nada Attar

Abstract:

We as humans use words with accompanying visual and facial cues to communicate effectively. Classifying facial emotion using computer vision methodologies has been an active research area in the computer vision field. In this paper, we propose a simple method for facial expression recognition that enhances accuracy. We tested our method on the FER-2013 dataset that contains static images. Instead of using Histogram equalization to preprocess the dataset, we used Unsharp Mask to emphasize texture and details and sharpened the edges. We also used ImageDataGenerator from Keras library for data augmentation. Then we used Convolutional Neural Networks (CNN) model to classify the images into 7 different facial expressions, yielding an accuracy of 69.46% on the test set. Our results show that using image preprocessing such as the sharpening technique for a CNN model can improve the performance, even when the CNN model is relatively simple.

Keywords: facial expression recognittion, image preprocessing, deep learning, CNN

Procedia PDF Downloads 129
8056 Deep Learning Based Text to Image Synthesis for Accurate Facial Composites in Criminal Investigations

Authors: Zhao Gao, Eran Edirisinghe

Abstract:

The production of an accurate sketch of a suspect based on a verbal description obtained from a witness is an essential task for most criminal investigations. The criminal investigation system employs specifically trained professional artists to manually draw a facial image of the suspect according to the descriptions of an eyewitness for subsequent identification. Within the advancement of Deep Learning, Recurrent Neural Networks (RNN) have shown great promise in Natural Language Processing (NLP) tasks. Additionally, Generative Adversarial Networks (GAN) have also proven to be very effective in image generation. In this study, a trained GAN conditioned on textual features such as keywords automatically encoded from a verbal description of a human face using an RNN is used to generate photo-realistic facial images for criminal investigations. The intention of the proposed system is to map corresponding features into text generated from verbal descriptions. With this, it becomes possible to generate many reasonably accurate alternatives to which the witness can use to hopefully identify a suspect from. This reduces subjectivity in decision making both by the eyewitness and the artist while giving an opportunity for the witness to evaluate and reconsider decisions. Furthermore, the proposed approach benefits law enforcement agencies by reducing the time taken to physically draw each potential sketch, thus increasing response times and mitigating potentially malicious human intervention. With publically available 'CelebFaces Attributes Dataset' (CelebA) and additionally providing verbal description as training data, the proposed architecture is able to effectively produce facial structures from given text. Word Embeddings are learnt by applying the RNN architecture in order to perform semantic parsing, the output of which is fed into the GAN for synthesizing photo-realistic images. Rather than the grid search method, a metaheuristic search based on genetic algorithms is applied to evolve the network with the intent of achieving optimal hyperparameters in a fraction the time of a typical brute force approach. With the exception of the ‘CelebA’ training database, further novel test cases are supplied to the network for evaluation. Witness reports detailing criminals from Interpol or other law enforcement agencies are sampled on the network. Using the descriptions provided, samples are generated and compared with the ground truth images of a criminal in order to calculate the similarities. Two factors are used for performance evaluation: The Structural Similarity Index (SSIM) and the Peak Signal-to-Noise Ratio (PSNR). A high percentile output from this performance matrix should attribute to demonstrating the accuracy, in hope of proving that the proposed approach can be an effective tool for law enforcement agencies. The proposed approach to criminal facial image generation has potential to increase the ratio of criminal cases that can be ultimately resolved using eyewitness information gathering.

Keywords: RNN, GAN, NLP, facial composition, criminal investigation

Procedia PDF Downloads 144
8055 High Responsivity of Zirconium boride/Chromium Alloy Heterostructure for Deep and Near UV Photodetector

Authors: Sanjida Akter, Ambali Alade Odebowale, Andrey E. Miroshnichenko, Haroldo T. Hattori

Abstract:

Photodetectors (PDs) play a pivotal role in optoelectronics and optical devices, serving as fundamental components that convert light signals into electrical signals. As the field progresses, the integration of advanced materials with unique optical properties has become a focal point, paving the way for the innovation of novel PDs. This study delves into the exploration of a cutting-edge photodetector designed for deep and near ultraviolet (UV) applications. The photodetector is constructed with a composite of Zirconium Boride (ZrB2) and Chromium (Cr) alloy, deposited onto a 6H nitrogen-doped silicon carbide substrate. The determination of the optimal alloy thickness is achieved through Finite-Difference Time-Domain (FDTD) simulation, and the synthesis of the alloy is accomplished using radio frequency (RF) sputtering. Remarkably, the resulting photodetector exhibits an exceptional responsivity of 3.5 A/W under an applied voltage of -2 V, at wavelengths of 405 nm and 280 nm. This heterostructure not only exemplifies high performance but also provides a versatile platform for the development of near UV photodetectors capable of operating effectively in challenging conditions, such as environments characterized by high power and elevated temperatures. This study contributes to the expanding landscape of photodetector technology, offering a promising avenue for the advancement of optoelectronic devices in demanding applications.

Keywords: responsivity, silicon carbide, ultraviolet photodetector, zirconium boride

Procedia PDF Downloads 47
8054 Evaluation of Modern Natural Language Processing Techniques via Measuring a Company's Public Perception

Authors: Burak Oksuzoglu, Savas Yildirim, Ferhat Kutlu

Abstract:

Opinion mining (OM) is one of the natural language processing (NLP) problems to determine the polarity of opinions, mostly represented on a positive-neutral-negative axis. The data for OM is usually collected from various social media platforms. In an era where social media has considerable control over companies’ futures, it’s worth understanding social media and taking actions accordingly. OM comes to the fore here as the scale of the discussion about companies increases, and it becomes unfeasible to gauge opinion on individual levels. Thus, the companies opt to automize this process by applying machine learning (ML) approaches to their data. For the last two decades, OM or sentiment analysis (SA) has been mainly performed by applying ML classification algorithms such as support vector machines (SVM) and Naïve Bayes to a bag of n-gram representations of textual data. With the advent of deep learning and its apparent success in NLP, traditional methods have become obsolete. Transfer learning paradigm that has been commonly used in computer vision (CV) problems started to shape NLP approaches and language models (LM) lately. This gave a sudden rise to the usage of the pretrained language model (PTM), which contains language representations that are obtained by training it on the large datasets using self-supervised learning objectives. The PTMs are further fine-tuned by a specialized downstream task dataset to produce efficient models for various NLP tasks such as OM, NER (Named-Entity Recognition), Question Answering (QA), and so forth. In this study, the traditional and modern NLP approaches have been evaluated for OM by using a sizable corpus belonging to a large private company containing about 76,000 comments in Turkish: SVM with a bag of n-grams, and two chosen pre-trained models, multilingual universal sentence encoder (MUSE) and bidirectional encoder representations from transformers (BERT). The MUSE model is a multilingual model that supports 16 languages, including Turkish, and it is based on convolutional neural networks. The BERT is a monolingual model in our case and transformers-based neural networks. It uses a masked language model and next sentence prediction tasks that allow the bidirectional training of the transformers. During the training phase of the architecture, pre-processing operations such as morphological parsing, stemming, and spelling correction was not used since the experiments showed that their contribution to the model performance was found insignificant even though Turkish is a highly agglutinative and inflective language. The results show that usage of deep learning methods with pre-trained models and fine-tuning achieve about 11% improvement over SVM for OM. The BERT model achieved around 94% prediction accuracy while the MUSE model achieved around 88% and SVM did around 83%. The MUSE multilingual model shows better results than SVM, but it still performs worse than the monolingual BERT model.

Keywords: BERT, MUSE, opinion mining, pretrained language model, SVM, Turkish

Procedia PDF Downloads 125
8053 Comparative Analysis of Turbulent Plane Jets from a Sharp-Edged Orifice, a Beveled-Edge Orifice and a Radially Contoured Nozzle

Authors: Ravinesh C. Deo

Abstract:

This article investigates through experiments the flow characteristics of plane jets from sharp-edged orifice-plate, beveled-edge and radially contoured nozzle. The first two configurations exhibit saddle-backed velocity profiles while the third shows a top-hat. A vena contracta is found for the jet emanating from orifice at x/h = 3 while the contoured case displays a potential core extending to the range x/h = 5. A spurt in jet pressure on the centerline supports vena contracta for the orifice-jet. Momentum thicknesses and integral length scales elongate linearly with x although the growth of the shear-layer and large-scale eddies for the orifice are greater than the contoured case. The near-field spectrum exhibits higher frequency of the primary eddies that concur with enhanced turbulence intensity. Importantly, highly “turbulent” state of the orifice-jet prevails in the far-field where the spectra confirm more energetic secondary eddies associated with greater flapping amplitude of the orifice-jet.

Keywords: orifice, beveled-edge-orifice, radially contoured nozzle, plane jets

Procedia PDF Downloads 141
8052 Predictive Analysis of the Stock Price Market Trends with Deep Learning

Authors: Suraj Mehrotra

Abstract:

The stock market is a volatile, bustling marketplace that is a cornerstone of economics. It defines whether companies are successful or in spiral. A thorough understanding of it is important - many companies have whole divisions dedicated to analysis of both their stock and of rivaling companies. Linking the world of finance and artificial intelligence (AI), especially the stock market, has been a relatively recent development. Predicting how stocks will do considering all external factors and previous data has always been a human task. With the help of AI, however, machine learning models can help us make more complete predictions in financial trends. Taking a look at the stock market specifically, predicting the open, closing, high, and low prices for the next day is very hard to do. Machine learning makes this task a lot easier. A model that builds upon itself that takes in external factors as weights can predict trends far into the future. When used effectively, new doors can be opened up in the business and finance world, and companies can make better and more complete decisions. This paper explores the various techniques used in the prediction of stock prices, from traditional statistical methods to deep learning and neural networks based approaches, among other methods. It provides a detailed analysis of the techniques and also explores the challenges in predictive analysis. For the accuracy of the testing set, taking a look at four different models - linear regression, neural network, decision tree, and naïve Bayes - on the different stocks, Apple, Google, Tesla, Amazon, United Healthcare, Exxon Mobil, J.P. Morgan & Chase, and Johnson & Johnson, the naïve Bayes model and linear regression models worked best. For the testing set, the naïve Bayes model had the highest accuracy along with the linear regression model, followed by the neural network model and then the decision tree model. The training set had similar results except for the fact that the decision tree model was perfect with complete accuracy in its predictions, which makes sense. This means that the decision tree model likely overfitted the training set when used for the testing set.

Keywords: machine learning, testing set, artificial intelligence, stock analysis

Procedia PDF Downloads 82
8051 Reading and Writing Memories in Artificial and Human Reasoning

Authors: Ian O'Loughlin

Abstract:

Memory networks aim to integrate some of the recent successes in machine learning with a dynamic memory base that can be updated and deployed in artificial reasoning tasks. These models involve training networks to identify, update, and operate over stored elements in a large memory array in order, for example, to ably perform question and answer tasks parsing real-world and simulated discourses. This family of approaches still faces numerous challenges: the performance of these network models in simulated domains remains considerably better than in open, real-world domains, wide-context cues remain elusive in parsing words and sentences, and even moderately complex sentence structures remain problematic. This innovation, employing an array of stored and updatable ‘memory’ elements over which the system operates as it parses text input and develops responses to questions, is a compelling one for at least two reasons: first, it addresses one of the difficulties that standard machine learning techniques face, by providing a way to store a large bank of facts, offering a way forward for the kinds of long-term reasoning that, for example, recurrent neural networks trained on a corpus have difficulty performing. Second, the addition of a stored long-term memory component in artificial reasoning seems psychologically plausible; human reasoning appears replete with invocations of long-term memory, and the stored but dynamic elements in the arrays of memory networks are deeply reminiscent of the way that human memory is readily and often characterized. However, this apparent psychological plausibility is belied by a recent turn in the study of human memory in cognitive science. In recent years, the very notion that there is a stored element which enables remembering, however dynamic or reconstructive it may be, has come under deep suspicion. In the wake of constructive memory studies, amnesia and impairment studies, and studies of implicit memory—as well as following considerations from the cognitive neuroscience of memory and conceptual analyses from the philosophy of mind and cognitive science—researchers are now rejecting storage and retrieval, even in principle, and instead seeking and developing models of human memory wherein plasticity and dynamics are the rule rather than the exception. In these models, storage is entirely avoided by modeling memory using a recurrent neural network designed to fit a preconceived energy function that attains zero values only for desired memory patterns, so that these patterns are the sole stable equilibrium points in the attractor network. So although the array of long-term memory elements in memory networks seem psychologically appropriate for reasoning systems, they may actually be incurring difficulties that are theoretically analogous to those that older, storage-based models of human memory have demonstrated. The kind of emergent stability found in the attractor network models more closely fits our best understanding of human long-term memory than do the memory network arrays, despite appearances to the contrary.

Keywords: artificial reasoning, human memory, machine learning, neural networks

Procedia PDF Downloads 255
8050 Development of Temple Architecture during the Reign of Kalachuri’s of Tripuri

Authors: Shivam Dubey, Shivakant Bajpai

Abstract:

The Kalachuri dynasty of Tripuri was a significant ruling dynasty in central India that held power over a vast region for a longer period compared to renowned dynasties like the Chandellas. Their capital, Tripuri (modern-day Tewar, a small village near Jabalpur), and its surrounding area witnessed significant developments that were later disrupted by the Royal Indian Railways' construction of railway lines. However, remnants of their achievements can still be found scattered in and around Tewar. The Kalachuris made remarkable contributions in the fields of art, architecture, and iconography. The evolution of temple architecture, particularly in Baghelkhand and the Mahakoshal range after the decline of the Gupta Empire, can be attributed to the Kalachuris. There is a notable progression from early temple styles to mature architecture, with numerous examples displaying continuity between the two. One particularly unique temple style features a ground plan resembling a complete Chaitya Hall, while the elevation showcases a circular Grabhagriha with a Mandapa and a conical Shikhara adorned with a series of Gavakshas. This distinctive temple style is among the most exceptional in central India. While several studies have been conducted on the Kalachuris' architecture, there is still a need for further research, as recent discoveries have provided valuable insights into understanding their architectural achievements. This paper aims to explore the development of architecture in this region, incorporating these recent findings.

Keywords: architecture, Kalachuri, art, iconography

Procedia PDF Downloads 59
8049 Low-Surface Roughness and High Optical Quality CdS Thin Film Grown by Modified Chemical Surface Deposition Method

Authors: A. Elsayed, M. H. Dewaidar, M. Ghali

Abstract:

We report on deposition of smooth, pinhole-free, low-surface roughness ( < 4nm) and high optical quality cadmium sulfide (CdS) thin films on glass substrates using our new method based on chemical surface deposition principle. In this method, cadmium acetate and thiourea are used as reactants under special growth conditions for deposition of CdS films. X-ray diffraction (XRD) measurements were used to examine the crystal structure properties of the deposited CdS films. In addition, UV-vis transmittance and low-temperature (4K) photoluminescence (PL) measurements were performed for quantifying optical properties of the deposited films. Interestingly, we found that XRD pattern of the deposited films has dramatically changed when the growth temperature was raised during the reaction. Namely, the XRD measurements reveal a structural change of CdS film from Cubic to Hexagonal phase upon increase in the growth temperature from 75 °C to 200 °C. Furthermore, the deposited films show high optical quality as confirmed from observation of both sharp edge in the transmittance spectra and strong PL intensity at room temperature. Also, we found a strong effect of the growth conditions on the optical band gap of the deposited films; where remarkable red-shift in the absorption edge with temperature is clearly seen in both transmission and PL spectra. Such tuning of both optical band gap and crystal structure of the deposited CdS films; can be utilized for tuning the electronic bands alignments between CdS and other light harvesting materials, like CuInGaSe or CdTe, for potential improvement in the efficiency of all-solution processed solar cells devices based on these heterostructures.

Keywords: thin film, CdS, new method, optical properties

Procedia PDF Downloads 249
8048 Bring Your Own Devices (BOYD): Risks and Mitigation Strategies

Authors: Mohammed Ketel

Abstract:

This paper discusses the security issues related to Bring Your Own Devices (BYOD) programs, an increasingly popular choice for small and big businesses alike, and explores the benefits, risks, the available controls and solutions to mitigate the inherent security concerns with mobile devices, in general, and BYOD programs specifically. The paper also discusses the approaches that organizations can apply to mitigate the risks, which may include policies, diverse technologies, education, and training.

Keywords: BYOD, security, policies, standards, controls, education

Procedia PDF Downloads 272
8047 Computer Aided Diagnosis Bringing Changes in Breast Cancer Detection

Authors: Devadrita Dey Sarkar

Abstract:

Regardless of the many technologic advances in the past decade, increased training and experience, and the obvious benefits of uniform standards, the false-negative rate in screening mammography remains unacceptably high .A computer aided neural network classification of regions of suspicion (ROS) on digitized mammograms is presented in this abstract which employs features extracted by a new technique based on independent component analysis. CAD is a concept established by taking into account equally the roles of physicians and computers, whereas automated computer diagnosis is a concept based on computer algorithms only. With CAD, the performance by computers does not have to be comparable to or better than that by physicians, but needs to be complementary to that by physicians. In fact, a large number of CAD systems have been employed for assisting physicians in the early detection of breast cancers on mammograms. A CAD scheme that makes use of lateral breast images has the potential to improve the overall performance in the detection of breast lumps. Because breast lumps can be detected reliably by computer on lateral breast mammographs, radiologists’ accuracy in the detection of breast lumps would be improved by the use of CAD, and thus early diagnosis of breast cancer would become possible. In the future, many CAD schemes could be assembled as packages and implemented as a part of PACS. For example, the package for breast CAD may include the computerized detection of breast nodules, as well as the computerized classification of benign and malignant nodules. In order to assist in the differential diagnosis, it would be possible to search for and retrieve images (or lesions) with these CAD systems, which would be reliable and useful method for quantifying the similarity of a pair of images for visual comparison by radiologists.

Keywords: CAD(computer-aided design), lesions, neural network, ROS(region of suspicion)

Procedia PDF Downloads 449
8046 Parallelization of Random Accessible Progressive Streaming of Compressed 3D Models over Web

Authors: Aayushi Somani, Siba P. Samal

Abstract:

Three-dimensional (3D) meshes are data structures, which store geometric information of an object or scene, generally in the form of vertices and edges. Current technology in laser scanning and other geometric data acquisition technologies acquire high resolution sampling which leads to high resolution meshes. While high resolution meshes give better quality rendering and hence is used often, the processing, as well as storage of 3D meshes, is currently resource-intensive. At the same time, web applications for data processing have become ubiquitous owing to their accessibility. For 3D meshes, the advancement of 3D web technologies, such as WebGL, WebVR, has enabled high fidelity rendering of huge meshes. However, there exists a gap in ability to stream huge meshes to a native client and browser application due to high network latency. Also, there is an inherent delay of loading WebGL pages due to large and complex models. The focus of our work is to identify the challenges faced when such meshes are streamed into and processed on hand-held devices, owing to its limited resources. One of the solutions that are conventionally used in the graphics community to alleviate resource limitations is mesh compression. Our approach deals with a two-step approach for random accessible progressive compression and its parallel implementation. The first step includes partition of the original mesh to multiple sub-meshes, and then we invoke data parallelism on these sub-meshes for its compression. Subsequent threaded decompression logic is implemented inside the Web Browser Engine with modification of WebGL implementation in Chromium open source engine. This concept can be used to completely revolutionize the way e-commerce and Virtual Reality technology works for consumer electronic devices. These objects can be compressed in the server and can be transmitted over the network. The progressive decompression can be performed on the client device and rendered. Multiple views currently used in e-commerce sites for viewing the same product from different angles can be replaced by a single progressive model for better UX and smoother user experience. Can also be used in WebVR for commonly and most widely used activities like virtual reality shopping, watching movies and playing games. Our experiments and comparison with existing techniques show encouraging results in terms of latency (compressed size is ~10-15% of the original mesh), processing time (20-22% increase over serial implementation) and quality of user experience in web browser.

Keywords: 3D compression, 3D mesh, 3D web, chromium, client-server architecture, e-commerce, level of details, parallelization, progressive compression, WebGL, WebVR

Procedia PDF Downloads 157
8045 Natural Patterns for Sustainable Cooling in the Architecture of Residential Buildings in Iran (Hot and Dry Climate)

Authors: Elnaz Abbasian, Mohsen Faizi

Abstract:

In its thousand-year development, architecture has gained valuable patterns. Iran’s desert regions possess developed patterns of traditional architecture and outstanding skeletal features. Unfortunately increasing population and urbanization growth in the past decade as well as the lack of harmony with environment’s texture has destroyed such permanent concepts in the building’s skeleton, causing a lot of energy waste in the modern architecture. The important question is how cooling patterns of Iran’s traditional architecture can be used in a new way in the modern architecture of residential buildings? This research is library-based and documental that looks at sustainable development, analyzes the features of Iranian architecture in hot and dry climate in terms of sustainability as well as historical patterns, and makes a model for real environment. By methodological analysis of past, it intends to suggest a new pattern for residential buildings’ cooling in Iran’s hot and dry climate which is in full accordance to the ecology of the design and at the same time possesses the architectural indices of the past. In the process of cities’ physical development, ecological measures, in proportion to desert’s natural background and climate conditions, has kept the natural fences, preventing buildings from facing climate adversities. Designing and construction of buildings with this viewpoint can reduce the energy needed for maintaining and regulating environmental conditions and with the use of appropriate building technology help minimizing the consumption of fossil fuels while having permanent patterns of desert buildings’ architecture.

Keywords: sustainability concepts, sustainable development, energy climate architecture, fossil fuel, hot and dry climate, patterns of traditional sustainability for residential buildings, modern pattern of cooling

Procedia PDF Downloads 289
8044 Classification of Barley Varieties by Artificial Neural Networks

Authors: Alper Taner, Yesim Benal Oztekin, Huseyin Duran

Abstract:

In this study, an Artificial Neural Network (ANN) was developed in order to classify barley varieties. For this purpose, physical properties of barley varieties were determined and ANN techniques were used. The physical properties of 8 barley varieties grown in Turkey, namely thousand kernel weight, geometric mean diameter, sphericity, kernel volume, surface area, bulk density, true density, porosity and colour parameters of grain, were determined and it was found that these properties were statistically significant with respect to varieties. As ANN model, three models, N-l, N-2 and N-3 were constructed. The performances of these models were compared. It was determined that the best-fit model was N-1. In the N-1 model, the structure of the model was designed to be 11 input layers, 2 hidden layers and 1 output layer. Thousand kernel weight, geometric mean diameter, sphericity, kernel volume, surface area, bulk density, true density, porosity and colour parameters of grain were used as input parameter; and varieties as output parameter. R2, Root Mean Square Error and Mean Error for the N-l model were found as 99.99%, 0.00074 and 0.009%, respectively. All results obtained by the N-l model were observed to have been quite consistent with real data. By this model, it would be possible to construct automation systems for classification and cleaning in flourmills.

Keywords: physical properties, artificial neural networks, barley, classification

Procedia PDF Downloads 163
8043 The Hindu Temple: Architecture, Culture and Spirituality

Authors: Tanisha Dutta, Vinayak S. Adane

Abstract:

A Hindu temple has always been the centre of worldly knowledge, art, culture, and spiritual knowledge. The temple centers and the temple structures alike, teach the observer about all kinds of worldly systems, codes of conduct, performing and other arts etc. During the medieval period, these were the only centers of knowledge. Therefore, these spaces had the burden and responsibility of covering all the various facets of life. It is understandable therefore, that a Hindu temple is easily the confluence of intricate architecture, cultural blossoming and spiritual knowledge transmittance. The architecture of a Hindu temple supports all these in a way that they co-exist and develop a symbiotic relationship, each enhancing the manifested form of the other. This symbiosis is presented through the temples of Khajuraho, India. This paper, therefore, elaborates the finer aspects of the mentioned areas in a Hindu temple context, through the case study of the Khajuraho group of temples.

Keywords: Hindu temples' concept, symbolism, temple architecture

Procedia PDF Downloads 239
8042 System Survivability in Networks

Authors: Asma Ben Yaghlane, Mohamed Naceur Azaiez

Abstract:

We consider the problem of attacks on networks. We define the concept of system survivability in networks in the presence of intelligent threats. Our setting of the problem assumes a flow to be sent from one source node to a destination node. The attacker attempts to disable the network by preventing the flow to reach its destination while the defender attempts to identify the best path-set to use to maximize the chance of arrival of the flow to the destination node. Our concept is shown to be different from the classical concept of network reliability. We distinguish two types of network survivability related to the defender and to the attacker of the network, respectively. We prove that the defender-based-network survivability plays the role of a lower bound while the attacker-based-network survivability plays the role of an upper bound of network reliability. We also prove that both concepts almost never agree nor coincide with network reliability. Moreover, we use the shortest-path problem to determine the defender-based-network survivability and the min-cut problem to determine the attacker-based-network survivability. We extend the problem to a variety of models including the minimum-spanning-tree problem and the multiple source-/destination-network problems.

Keywords: defense/attack strategies, information, networks, reliability, survivability

Procedia PDF Downloads 374
8041 Evaluation of Video Quality Metrics and Performance Comparison on Contents Taken from Most Commonly Used Devices

Authors: Pratik Dhabal Deo, Manoj P.

Abstract:

With the increasing number of social media users, the amount of video content available has also significantly increased. Currently, the number of smartphone users is at its peak, and many are increasingly using their smartphones as their main photography and recording devices. There have been a lot of developments in the field of Video Quality Assessment (VQA) and metrics like VMAF, SSIM etc. are said to be some of the best performing metrics, but the evaluation of these metrics is dominantly done on professionally taken video contents using professional tools, lighting conditions etc. No study particularly pinpointing the performance of the metrics on the contents taken by users on very commonly available devices has been done. Datasets that contain a huge number of videos from different high-end devices make it difficult to analyze the performance of the metrics on the content from most used devices even if they contain contents taken in poor lighting conditions using lower-end devices. These devices face a lot of distortions due to various factors since the spectrum of contents recorded on these devices is huge. In this paper, we have presented an analysis of the objective VQA metrics on contents taken only from most used devices and their performance on them, focusing on full-reference metrics. To carry out this research, we created a custom dataset containing a total of 90 videos that have been taken from three most commonly used devices, and android smartphone, an IOS smartphone and a DSLR. On the videos taken on each of these devices, the six most common types of distortions that users face have been applied on addition to already existing H.264 compression based on four reference videos. These six applied distortions have three levels of degradation each. A total of the five most popular VQA metrics have been evaluated on this dataset and the highest values and the lowest values of each of the metrics on the distortions have been recorded. Finally, it is found that blur is the artifact on which most of the metrics didn’t perform well. Thus, in order to understand the results better the amount of blur in the data set has been calculated and an additional evaluation of the metrics was done using HEVC codec, which is the next version of H.264 compression, on the camera that proved to be the sharpest among the devices. The results have shown that as the resolution increases, the performance of the metrics tends to become more accurate and the best performing metric among them is VQM with very few inconsistencies and inaccurate results when the compression applied is H.264, but when the compression is applied is HEVC, SSIM and VMAF have performed significantly better.

Keywords: distortion, metrics, performance, resolution, video quality assessment

Procedia PDF Downloads 190
8040 Influence of Partially-Replaced Coarse Aggregates with Date Palm Seeds on the Concrete Properties

Authors: Fahed Alrshoudi

Abstract:

Saudi Arabia is ranked the third of the largest suppliers of Dates worldwide (about 28.5 million palm trees), producing more than 2 million tons of dates yearly. These trees produce large quantity of dates palm seeds (DPS) which can be considered literally as a waste. The date seeds are stiff, therefore, it is possible to utilize DPS as coarse aggregates in lightweight concrete for certain structural applications and to participate at reusing the waste. The use of DPS as coarse aggregate in concrete can be an alternative choice as a partial replacement of the stone aggregates (SA). This paper reports the influence of partially replaced stone aggregates with DPS on the hardened properties of concrete performance. Based on the experimental results, the DPS has the potential use as an acceptable alternative aggregates in producing structural lightweight concrete members, instead of stone aggregates.

Keywords: compressive strength, tensile Strength, date palm seeds, aggregate

Procedia PDF Downloads 116
8039 An Application Framework for Integrating Wireless Sensor and Actuator Networks for Precision Farmingas Web of Things to Cloud Interface Using PaaS

Authors: Sumaya Ismail, Aijaz Ahmad Reshi

Abstract:

The advances in sensor and embedded technologies have led to rapid developments in Wireless Sensor Networks (WSNs). Presently researchers focus on the integration of WSNs to the Internet for their pervasive availability to access these network resources as the interoperable subsystems. The recent computing technologies like cloud computing has made resource sharing as a converged infrastructure with required service interfaces for the shared resources over the Internet. This paper presents application architecture for wireless Sensor and Actuator Networks (WSANS) following web of things, which allows easy integration of each node to the Internet in order to provide them with web accessibility. The architecture enables the sensors and actuator nodes accessed and controlled using cloud interface on WWW. The application architecture was implemented using existing web and its emerging technologies. In particular, the Representational State Transfer protocol (REST) was extended for the specific requirements of the application. The Cloud computing environment has been used as a development platform for the application to assess the possibility of integrating the WSAN nodes to Cloud services. The mushroom farm environment monitoring and control using WSANs has been taken as a research use case.

Keywords: WSAN, REST, web of things, ZigBee, cloud interface, PaaS, sensor gateway

Procedia PDF Downloads 88
8038 Reflection Phase Tuning of Graphene Plasmons by Substrate Design

Authors: Xiaojie Jiang, Wei Cai, Yinxiao Xiang, Ni Zhang, Mengxin Ren, Xinzheng Zhang, Jingjun Xu

Abstract:

Reflection phase of graphene plasmons (GPs) at an abrupt interface is very important, which determines the plasmon resonance of graphene structures of deep sub-wavelength scales. However, at an abrupt graphene edge, the reflection phase is always a constant, ΦR ≈ π/4. In this work, we show that the reflection phase of GPs can be efficiently changed through substrate design. Reflection phase of graphene plasmons (GPs) at an abrupt interface is very important, which determines the plasmon resonance of graphene structures of deep sub-wavelength scales. However, at an abrupt graphene edge, the reflection phase is always a constant, ΦR ≈ π/4. In this work, we show that the reflection phase of GPs can be efficiently changed through substrate design. Specifically, the reflection phase is no longer π/4 at the interface formed by placing a graphene sheet on different substrates. Moreover, tailorable reflection phase of GPs up to 2π variation can be further achieved by scattering GPs at a junction consisting of two such dielectric interfaces with various gap width acting as a Fabry-Perot cavity. Besides, the evolution of plasmon mode in graphene ribbons based on the interface reflection phase tuning is predicted, which is expected to be observed in near-field experiments with scattering-type scanning near-field optical microscopy (s-SNOM). Our work provides another way for in-plane plasmon control, which should find applications for integrated plasmon devices design using graphene.Specifically, the reflection phase is no longer π/4 at the interface formed by placing a graphene sheet on different substrates. Moreover, tailorable reflection phase of GPs up to 2π variation can be further achieved by scattering GPs at a junction consisting of two such dielectric interfaces with various gap width acting as a Fabry-Perot cavity. Besides, the evolution of plasmon mode in graphene ribbons based on the interface reflection phase tuning is predicted, which is expected to be observed in near-field experiments with scattering-type scanning near-field optical microscopy (s-SNOM). Our work provides a new way for in-plane plasmon control, which should find applications for integrated plasmon devices design using graphene.

Keywords: graphene plasmons, reflection phase tuning, plasmon mode tuning, Fabry-Perot cavity

Procedia PDF Downloads 136
8037 Study of Elastic-Plastic Fatigue Crack in Functionally Graded Materials

Authors: Somnath Bhattacharya, Kamal Sharma, Vaibhav Sonkar

Abstract:

Composite materials emerged in the middle of the 20th century as a promising class of engineering materials providing new prospects for modern technology. Recently, a new class of composite materials known as functionally graded materials (FGMs) has drawn considerable attention of the scientific community. In general, FGMs are defined as composite materials in which the composition or microstructure or both are locally varied so that a certain variation of the local material properties is achieved. This gradual change in composition and microstructure of material is suitable to get gradient of properties and performances. FGMs are synthesized in such a way that they possess continuous spatial variations in volume fractions of their constituents to yield a predetermined composition. These variations lead to the formation of a non-homogeneous macrostructure with continuously varying mechanical and / or thermal properties in one or more than one direction. Lightweight functionally graded composites with high strength to weight and stiffness to weight ratios have been used successfully in aircraft industry and other engineering applications like in electronics industry and in thermal barrier coatings. In the present work, elastic-plastic crack growth problems (using Ramberg-Osgood Model) in an FGM plate under cyclic load has been explored by extended finite element method. Both edge and centre crack problems have been solved by taking additionally holes, inclusions and minor cracks under plane stress conditions. Both soft and hard inclusions have been implemented in the problems. The validity of linear elastic fracture mechanics theory is limited to the brittle materials. A rectangular plate of functionally graded material of length 100 mm and height 200 mm with 100% copper-nickel alloy on left side and 100% ceramic (alumina) on right side is considered in the problem. Exponential gradation in property is imparted in x-direction. A uniform traction of 100 MPa is applied to the top edge of the rectangular domain along y direction. In some problems, domain contains major crack along with minor cracks or / and holes or / and inclusions. Major crack is located the centre of the left edge or the centre of the domain. The discontinuities, such as minor cracks, holes, and inclusions are added either singly or in combination with each other. On the basis of this study, it is found that effect of minor crack in the domain’s failure crack length is minimum whereas soft inclusions have moderate effect and the effect of holes have maximum effect. It is observed that the crack growth is more before the failure in each case when hard inclusions are present in place of soft inclusions.

Keywords: elastic-plastic, fatigue crack, functionally graded materials, extended finite element method (XFEM)

Procedia PDF Downloads 374
8036 Computer-Aided Diagnosis of Polycystic Kidney Disease Using ANN

Authors: G. Anjan Babu, G. Sumana, M. Rajasekhar

Abstract:

Many inherited diseases and non-hereditary disorders are common in the development of renal cystic diseases. Polycystic kidney disease (PKD) is a disorder developed within the kidneys in which grouping of cysts filled with water like fluid. PKD is responsible for 5-10% of end-stage renal failure treated by dialysis or transplantation. New experimental models, application of molecular biology techniques have provided new insights into the pathogenesis of PKD. Researchers are showing keen interest for developing an automated system by applying computer aided techniques for the diagnosis of diseases. In this paper a multi-layered feed forward neural network with one hidden layer is constructed, trained and tested by applying back propagation learning rule for the diagnosis of PKD based on physical symptoms and test results of urinanalysis collected from the individual patients. The data collected from 50 patients are used to train and test the network. Among these samples, 75% of the data used for training and remaining 25% of the data are used for testing purpose. Furthermore, this trained network is used to implement for new samples. The output results in normality and abnormality of the patient.

Keywords: dialysis, hereditary, transplantation, polycystic, pathogenesis

Procedia PDF Downloads 363
8035 Quantitative Analysis of Multiprocessor Architectures for Radar Signal Processing

Authors: Deepak Kumar, Debasish Deb, Reena Mamgain

Abstract:

Radar signal processing requires high number crunching capability. Most often this is achieved using multiprocessor platform. Though multiprocessor platform provides the capability of meeting the real time computational challenges, the architecture of the same along with mapping of the algorithm on the architecture plays a vital role in efficiently using the platform. Towards this, along with standard performance metrics, few additional metrics are defined which helps in evaluating the multiprocessor platform along with the algorithm mapping. A generic multiprocessor architecture can not suit all the processing requirements. Depending on the system requirement and type of algorithms used, the most suitable architecture for the given problem is decided. In the paper, we study different architectures and quantify the different performance metrics which enables comparison of different architectures for their merit. We also carried out case study of different architectures and their efficiency depending on parallelism exploited on algorithm or data or both.

Keywords: radar signal processing, multiprocessor architecture, efficiency, load imbalance, buffer requirement, pipeline, parallel, hybrid, cluster of processors (COPs)

Procedia PDF Downloads 400
8034 Bioclimatic Devices in the Historical Rural Building: A Carried out Analysis on Some Rural Architectures in Puglia

Authors: Valentina Adduci

Abstract:

The developing research aims to define in general the criteria of environmental sustainability of rural buildings in Puglia and particularly in the manor farm. The main part of the study analyzes the relationship / dependence between the rural building and the landscape which, after many stratifications, results clearly identified and sometimes also characterized in a positive way. The location of the manor farm, in fact, is often conditioned by the infrastructural network and by the structure of the agricultural landscape. The manor farm, without the constraints due to the urban pattern’s density, was developed in accordance with a logical settlement that gives priority to the environmental aspects. These vernacular architectures are the most valuable example of how our ancestors have planned their dwellings according to nature. The 237 farms, analysis’ object, have been reported in cartography through the GIS system; a symbol has been assigned to each of them to identify the architectural typology and a different color for the historical period of construction. A datasheet template has been drawn up, and it has made possible a deeper understanding of each manor farm. This method provides a faster comparison of the most recurring characters in all the considered buildings, except for those farms which benefited from special geographical conditions, such as proximity to the road network or waterways. Below there are some of the most frequently constants derived from the statistical study of the examined buildings: southeast orientation of the main facade; placement of the sheep pen on the ground tilted and exposed to the south side; larger windowed surface on the south elevation; smaller windowed surface on the north elevation; presence of shielding vegetation near the more exposed elevations to the solar radiation; food storage’s rooms located on the ground floor or in the basement; animal shelter located in north side of the farm; presence of tanks and wells, sometimes combined with a very accurate channeling storm water system; thick layers of masonry walls, inside of which were often obtained hollow spaces to house stairwells or depots for the food storage; exclusive use of local building materials. The research aims to trace the ancient use of bioclimatic constructive techniques in the Apulian rural architecture and to define those that derive from an empirical knowledge and those that respond to an already encoded design. These constructive expedients are especially useful to obtain an effective passive cooling, to promote the natural ventilation and to built ingenious systems for the recovery and the preservation of rainwater and are still found in some of the manor farms analyzed, most of them are, today, in a serious state of neglect.

Keywords: bioclimatic devices, farmstead, rural landscape, sustainability

Procedia PDF Downloads 373
8033 Applying Multiplicative Weight Update to Skin Cancer Classifiers

Authors: Animish Jain

Abstract:

This study deals with using Multiplicative Weight Update within artificial intelligence and machine learning to create models that can diagnose skin cancer using microscopic images of cancer samples. In this study, the multiplicative weight update method is used to take the predictions of multiple models to try and acquire more accurate results. Logistic Regression, Convolutional Neural Network (CNN), and Support Vector Machine Classifier (SVMC) models are employed within the Multiplicative Weight Update system. These models are trained on pictures of skin cancer from the ISIC-Archive, to look for patterns to label unseen scans as either benign or malignant. These models are utilized in a multiplicative weight update algorithm which takes into account the precision and accuracy of each model through each successive guess to apply weights to their guess. These guesses and weights are then analyzed together to try and obtain the correct predictions. The research hypothesis for this study stated that there would be a significant difference in the accuracy of the three models and the Multiplicative Weight Update system. The SVMC model had an accuracy of 77.88%. The CNN model had an accuracy of 85.30%. The Logistic Regression model had an accuracy of 79.09%. Using Multiplicative Weight Update, the algorithm received an accuracy of 72.27%. The final conclusion that was drawn was that there was a significant difference in the accuracy of the three models and the Multiplicative Weight Update system. The conclusion was made that using a CNN model would be the best option for this problem rather than a Multiplicative Weight Update system. This is due to the possibility that Multiplicative Weight Update is not effective in a binary setting where there are only two possible classifications. In a categorical setting with multiple classes and groupings, a Multiplicative Weight Update system might become more proficient as it takes into account the strengths of multiple different models to classify images into multiple categories rather than only two categories, as shown in this study. This experimentation and computer science project can help to create better algorithms and models for the future of artificial intelligence in the medical imaging field.

Keywords: artificial intelligence, machine learning, multiplicative weight update, skin cancer

Procedia PDF Downloads 62