Search results for: e-content producing algorithm
1451 Novel Routes to the Synthesis and Functionalization of Metallic and Semiconductor Thin Film and Nanoparticles
Authors: Hanan. Al Chaghouri, Mohammad Azad Malik, P. John Thomas, Paul O’Brien
Abstract:
The process of assembling metal nanoparticles at the interface of two liquids has received a great deal of attention over the past few years due to a wide range of important applications and their unusual properties as compared to bulk materials. We present a low cost, simple and cheap synthesis of metal nanoparticles, core/shell structures and semiconductors followed by assembly of these particles between immiscible liquids. The aim of this talk is divided to three parts: Firstly, to describe the achievement of a closed loop recycling for producing cadmium sulfide as powders and/or nanostructured thin films for solar cells or other optoelectronic devices applications by using a different chain length of commercially available secondary amines of dithiocarbamato complexes. The approach can be extended to other metal sulfides such as those of Zn, Pb, Cu, or Fe and many transition metals and oxides. Secondly, to synthesis significantly cheaper magnetic particles suited for the mass market. Ni/NiO nanoparticles with ferromagnetic properties at room temperature were among the smallest and strongest magnets (5 nm) were made in solution. The applications of this work can be to produce viable storage devices and the other possibility is to disperse these nanocrystals in solution and use it to make ferrofluids which have a number of mature applications. The third part is about preparing and assembling of submicron silver, cobalt and nickel particles by using polyol methods and liquid/liquid interface, respectively. Coinage metals like gold, copper and silver are suitable for plasmonic thin film solar cells because of their low resistivity and strong interactions with visible light waves. Silver is the best choice for solar cell application since it has low absorption losses and high radiative efficiency compared to gold and copper. Assembled cobalt and nickel as films are promising for spintronic, magnetic and magneto-electronic and biomedics.Keywords: metal nanoparticles, core/shell structures and semiconductors, ferromagnetic properties, closed loop recycling, liquid/liquid interface
Procedia PDF Downloads 4621450 Profit-Based Artificial Neural Network (ANN) Trained by Migrating Birds Optimization: A Case Study in Credit Card Fraud Detection
Authors: Ashkan Zakaryazad, Ekrem Duman
Abstract:
A typical classification technique ranks the instances in a data set according to the likelihood of belonging to one (positive) class. A credit card (CC) fraud detection model ranks the transactions in terms of probability of being fraud. In fact, this approach is often criticized, because firms do not care about fraud probability but about the profitability or costliness of detecting a fraudulent transaction. The key contribution in this study is to focus on the profit maximization in the model building step. The artificial neural network proposed in this study works based on profit maximization instead of minimizing the error of prediction. Moreover, some studies have shown that the back propagation algorithm, similar to other gradient–based algorithms, usually gets trapped in local optima and swarm-based algorithms are more successful in this respect. In this study, we train our profit maximization ANN using the Migrating Birds optimization (MBO) which is introduced to literature recently.Keywords: neural network, profit-based neural network, sum of squared errors (SSE), MBO, gradient descent
Procedia PDF Downloads 4751449 Performance and Specific Emissions of an SI Engine Using Anhydrous Ethanol–Gasoline Blends in the City of Bogota
Authors: Alexander García Mariaca, Rodrigo Morillo Castaño, Juan Rolón Ríos
Abstract:
The government of Colombia has promoted the use of biofuels in the last 20 years through laws and resolutions, which regulate their use, with the objective to improve the atmospheric air quality and to promote Colombian agricultural industry. However, despite the use of blends of biofuels with fossil fuels, the air quality in large cities does not get better, this deterioration in the air is mainly caused by mobile sources that working with spark ignition internal combustion engines (SI-ICE), operating with a mixture in volume of 90 % gasoline and 10 % ethanol called E10, that for the case of Bogota represent 84 % of the fleet. Another problem is that Colombia has big cities located above 2200 masl and there are no accurate studies on the impact that the E10 mixture could cause in the emissions and performance of SI-ICE. This study aims to establish the optimal blend between gasoline ethanol in which an SI engine operates more efficiently in urban centres located at 2600 masl. The test was developed on SI engine four-stroke, single cylinder, naturally aspirated and with carburettor for the fuel supply using blends of gasoline and anhydrous ethanol in different ratios E10, E15, E20, E40, E60, E85 and E100. These tests were conducted in the city of Bogota, which is located at 2600 masl, with the engine operating at 3600 rpm and at 25, 50, 75 and 100% of load. The results show that the performance variables as engine brake torque, brake power and brake thermal efficiency decrease, while brake specific fuel consumption increases with the rise in the percentage of ethanol in the mixture. On the other hand, the specific emissions of CO2 and NOx present increases while specific emissions of CO and HC decreases compared to those produced by gasoline. From the tests, it is concluded that the SI-ICE worked more efficiently with the E40 mixture, where was obtained an increases of the brake power of 8.81 % and a reduction on brake specific fuel consumption of 2.5 %, coupled with a reduction in the specific emissions of CO2, HC and CO in 9.72, 52.88 and 76.66 % respectively compared to the results obtained with the E10 blend. This behaviour is because the E40 mixture provides the appropriate amount of the oxygen for the combustion process, which leads to better utilization of available energy in this process, thus generating a comparable power output to the E10 mixing and producing lower emissions CO and HC with the other test blends. Nevertheless, the emission of NOx increases in 106.25 %.Keywords: emissions, ethanol, gasoline, engine, performance
Procedia PDF Downloads 3241448 Degraded Document Analysis and Extraction of Original Text Document: An Approach without Optical Character Recognition
Authors: L. Hamsaveni, Navya Prakash, Suresha
Abstract:
Document Image Analysis recognizes text and graphics in documents acquired as images. An approach without Optical Character Recognition (OCR) for degraded document image analysis has been adopted in this paper. The technique involves document imaging methods such as Image Fusing and Speeded Up Robust Features (SURF) Detection to identify and extract the degraded regions from a set of document images to obtain an original document with complete information. In case, degraded document image captured is skewed, it has to be straightened (deskew) to perform further process. A special format of image storing known as YCbCr is used as a tool to convert the Grayscale image to RGB image format. The presented algorithm is tested on various types of degraded documents such as printed documents, handwritten documents, old script documents and handwritten image sketches in documents. The purpose of this research is to obtain an original document for a given set of degraded documents of the same source.Keywords: grayscale image format, image fusing, RGB image format, SURF detection, YCbCr image format
Procedia PDF Downloads 3771447 High Secure Data Hiding Using Cropping Image and Least Significant Bit Steganography
Authors: Khalid A. Al-Afandy, El-Sayyed El-Rabaie, Osama Salah, Ahmed El-Mhalaway
Abstract:
This paper presents a high secure data hiding technique using image cropping and Least Significant Bit (LSB) steganography. The predefined certain secret coordinate crops will be extracted from the cover image. The secret text message will be divided into sections. These sections quantity is equal the image crops quantity. Each section from the secret text message will embed into an image crop with a secret sequence using LSB technique. The embedding is done using the cover image color channels. Stego image is given by reassembling the image and the stego crops. The results of the technique will be compared to the other state of art techniques. Evaluation is based on visualization to detect any degradation of stego image, the difficulty of extracting the embedded data by any unauthorized viewer, Peak Signal-to-Noise Ratio of stego image (PSNR), and the embedding algorithm CPU time. Experimental results ensure that the proposed technique is more secure compared with the other traditional techniques.Keywords: steganography, stego, LSB, crop
Procedia PDF Downloads 2701446 Production of High Purity Cellulose Products from Sawdust Waste Material
Authors: Simiksha Balkissoon, Jerome Andrew, Bruce Sithole
Abstract:
Approximately half of the wood processed in the Forestry, Timber, Pulp and Paper (FTPP) sector is accumulated as waste. The concept of a “green economy” encourages industries to employ revolutionary, transformative technologies to eliminate waste generation by exploring the development of new value chains. The transition towards an almost paperless world driven by the rise of digital media has resulted in a decline in traditional paper markets, prompting the FTTP sector to reposition itself and expand its product offerings by unlocking the potential of value-adding opportunities from renewable resources such as wood to generate revenue and mitigate its environmental impact. The production of valuable products from wood waste such as sawdust has been extensively explored in recent years. Wood components such as lignin, cellulose and hemicelluloses, which can be extracted selectively by chemical processing, are suitable candidates for producing numerous high-value products. In this study, a novel approach to produce high-value cellulose products, such as dissolving wood pulp (DWP), from sawdust was developed. DWP is a high purity cellulose product used in several applications such as pharmaceutical, textile, food, paint and coatings industries. The proposed approach demonstrates the potential to eliminate several complex processing stages, such as pulping and bleaching, which are associated with traditional commercial processes to produce high purity cellulose products such as DWP, making it less chemically energy and water-intensive. The developed process followed the path of experimentally designed lab tests evaluating typical processing conditions such as residence time, chemical concentrations, liquid-to-solid ratios and temperature, followed by the application of suitable purification steps. Characterization of the product from the initial stage was conducted using commercially available DWP grades as reference materials. The chemical characteristics of the products thus far have shown similar properties to commercial products, making the proposed process a promising and viable option for the production of DWP from sawdust.Keywords: biomass, cellulose, chemical treatment, dissolving wood pulp
Procedia PDF Downloads 1881445 Occurrence and Levels of Mycotoxins in On-Farm Stored Sesame in Major-Growing Districts of Ethiopia
Authors: S. Alemayehu, F. A. Abera, K. M. Ayimut, R. Mahroof, J. Harvey, B. Subramanyam
Abstract:
The occurrence of mycotoxins in sesame seeds poses a significant threat to food safety and the economy in Ethiopia. This study aimed to determine the levels and occurrence of mycotoxins in on-farm stored sesame seeds in major-growing districts of Ethiopia. A total of 470 sesame seed samples were collected from randomly selected farmers' storage structures in five major-growing districts using purposive sampling techniques. An enzyme-linked immunosorbent assay (ELISA) was used to analyze the collected samples for the presence of four mycotoxins: total aflatoxins (AFT), ochratoxin A (OTA), total fumonisins (FUM), and deoxynivalenol (DON). The study found that all samples contained varying levels of mycotoxins, with AFT and DON being the most prevalent. AFT concentrations in detected samples ranged from 2.5 to 27.8 parts per billion (ppb), with a mean concentration of 13.8 ppb. OTA levels ranged from 5.0 ppb to 9.7 ppb, with a mean level of 7.1 ppb. Total fumonisin concentrations ranged from 300 to 1300 ppb in all samples, with a mean of 800 ppb. DON concentrations ranged from 560 to 700 ppb in the analyzed samples. The majority (96.8%) of the samples were safe from AFT, FUM, and DON mean levels when compared to the Federal Drug Administration maximum limit. AFT-OTA, DON-OTA, AFT-FUM, FUM-DON, and FUM-OTA, respectively, had co-occurrence rates of 44.0, 38.3, 33.8, 30.2, 29.8 and 26.0% for mycotoxins. On average, 37.2% of the sesame samples had fungal infection, and seed germination rates ranged from 66.8% to 91.1%. The Limmu district had higher levels of total aflatoxins, kernel infection, and lower germination rates than other districts. The Wollega variety of sesame had higher kernel infection, total aflatoxins concentration, and lower germination rates than other varieties. Grain age had a statistically significant (p<0.05) effect on both kernel infection and germination. The storage methods used for sesame in major-growing districts of Ethiopia favor mycotoxin-producing fungi. As the levels of mycotoxins in sesame are of public health significance, stakeholders should come together to identify secure and suitable storage technologies to maintain the quantity and quality of sesame at the level of smallholder farmers. This study suggests the need for suitable storage technologies to maintain the quality of sesame and reduce the risk of mycotoxin contamination.Keywords: districts, seed germination, kernel infection, moisture content, relative humidity, temperature
Procedia PDF Downloads 1371444 Syllogistic Reasoning with 108 Inference Rules While Case Quantities Change
Authors: Mikhail Zarechnev, Bora I. Kumova
Abstract:
A syllogism is a deductive inference scheme used to derive a conclusion from a set of premises. In a categorical syllogisms, there are only two premises and every premise and conclusion is given in form of a quantified relationship between two objects. The different order of objects in premises give classification known as figures. We have shown that the ordered combinations of 3 generalized quantifiers with certain figure provide in total of 108 syllogistic moods which can be considered as different inference rules. The classical syllogistic system allows to model human thought and reasoning with syllogistic structures always attracted the attention of cognitive scientists. Since automated reasoning is considered as part of learning subsystem of AI agents, syllogistic system can be applied for this approach. Another application of syllogistic system is related to inference mechanisms on the Semantic Web applications. In this paper we proposed the mathematical model and algorithm for syllogistic reasoning. Also the model of iterative syllogistic reasoning in case of continuous flows of incoming data based on case–based reasoning and possible applications of proposed system were discussed.Keywords: categorical syllogism, case-based reasoning, cognitive architecture, inference on the semantic web, syllogistic reasoning
Procedia PDF Downloads 4111443 Low Density Parity Check Codes
Authors: Kassoul Ilyes
Abstract:
The field of error correcting codes has been revolutionized by the introduction of iteratively decoded codes. Among these, LDPC codes are now a preferred solution thanks to their remarkable performance and low complexity. The binary version of LDPC codes showed even better performance, although it’s decoding introduced greater complexity. This thesis studies the performance of binary LDPC codes using simplified weighted decisions. Information is transported between a transmitter and a receiver by digital transmission systems, either by propagating over a radio channel or also by using a transmission medium such as the transmission line. The purpose of the transmission system is then to carry the information from the transmitter to the receiver as reliably as possible. These codes have not generated enough interest within the coding theory community. This forgetfulness will last until the introduction of Turbo-codes and the iterative principle. Then it was proposed to adopt Pearl's Belief Propagation (BP) algorithm for decoding these codes. Subsequently, Luby introduced irregular LDPC codes characterized by a parity check matrix. And finally, we study simplifications on binary LDPC codes. Thus, we propose a method to make the exact calculation of the APP simpler. This method leads to simplifying the implementation of the system.Keywords: LDPC, parity check matrix, 5G, BER, SNR
Procedia PDF Downloads 1561442 Initial Dip: An Early Indicator of Neural Activity in Functional Near Infrared Spectroscopy Waveform
Authors: Mannan Malik Muhammad Naeem, Jeong Myung Yung
Abstract:
Functional near infrared spectroscopy (fNIRS) has a favorable position in non-invasive brain imaging techniques. The concentration change of oxygenated hemoglobin and de-oxygenated hemoglobin during particular cognitive activity is the basis for this neuro-imaging modality. Two wavelengths of near-infrared light can be used with modified Beer-Lambert law to explain the indirect status of neuronal activity inside brain. The temporal resolution of fNIRS is very good for real-time brain computer-interface applications. The portability, low cost and an acceptable temporal resolution of fNIRS put it on a better position in neuro-imaging modalities. In this study, an optimization model for impulse response function has been used to estimate/predict initial dip using fNIRS data. In addition, the activity strength parameter related to motor based cognitive task has been analyzed. We found an initial dip that remains around 200-300 millisecond and better localize neural activity.Keywords: fNIRS, brain-computer interface, optimization algorithm, adaptive signal processing
Procedia PDF Downloads 2261441 A Scalable Model of Fair Socioeconomic Relations Based on Blockchain and Machine Learning Algorithms-1: On Hyperinteraction and Intuition
Authors: Merey M. Sarsengeldin, Alexandr S. Kolokhmatov, Galiya Seidaliyeva, Alexandr Ozerov, Sanim T. Imatayeva
Abstract:
This series of interdisciplinary studies is an attempt to investigate and develop a scalable model of fair socioeconomic relations on the base of blockchain using positive psychology techniques and Machine Learning algorithms for data analytics. In this particular study, we use hyperinteraction approach and intuition to investigate their influence on 'wisdom of crowds' via created mobile application which was created for the purpose of this research. Along with the public blockchain and private Decentralized Autonomous Organization (DAO) which were elaborated by us on the base of Ethereum blockchain, a model of fair financial relations of members of DAO was developed. We developed a smart contract, so-called, Fair Price Protocol and use it for implementation of model. The data obtained from mobile application was analyzed by ML algorithms. A model was tested on football matches.Keywords: blockchain, Naïve Bayes algorithm, hyperinteraction, intuition, wisdom of crowd, decentralized autonomous organization
Procedia PDF Downloads 1711440 Design Thinking Activities: A Tool in Overcoming Student Reticence
Authors: Marinel Dayawon
Abstract:
Student participation in classroom activities is vital in the teaching- learning the process as it develops self-confidence, social relationships and good academic performance of students. It is the teacher’s empathetic manner and creativity to create solutions that encourage teamwork and mutual support while dropping the academic competition within the class that hinder every shy student to walk with courage and talk with conviction because they consider their ideas, weak, as compared to the bright students. This study aimed to explore the different design thinking strategies that will change the mindset of shy students in classroom activities, maximizing their participation in all given tasks while sharing their views through ideation and providing them a wider world through compromise agreement within the members of the group, sensitivity to one’s idea, thus, arriving at a collective decision in the development of a prototype that indicates improvement in their classroom involvement. The study used the qualitative type of research. Triangulation is done through participant observation, focus group discussion and interview, documented through photos and videos. The respondents were the second- year Bachelor of Secondary Education students of the Institute of Teacher Education at Isabela State University- Cauayan City Campus. The result of the study revealed that reticent students when involved in game activities through a slap and tap method, writing their clustered ideas, using sticky notes is excited in sharing ideas as it doesn’t use oral communication. It is also observed after three weeks of using the design thinking strategies; shy students volunteer as secretary, rapporteur or group leader in the team- building activities as it represents the ideas of the heterogeneous group, removing the individual identity of the ideas. Superior students learned to listen to the ideas of the reticent students and involved them in the prototyping process of designing a remediation program for high school students showing reticence in the classroom, making their experience as a benchmark. The strategies made a 360- degrees transformation of the shy students, producing their journal log, in their journey to being open. Thus, faculty members are now adopting the design thinking approach.Keywords: design thinking activities, qualitative, reticent students, Isabela, Philippines
Procedia PDF Downloads 2251439 Little Girls and Big Stories: A Thematic Analysis of Gender Representations in Selected Asian Room to Read Storybooks
Authors: Cheeno Marlo Sayuno
Abstract:
Room to Read is an international nonprofit organization aimed at empowering young readers through literature and literacy education. In particular, the organization is focused on girls’ education in schools and bettering their social status through crafting stories and making sure that these stories are accessible to them. In 2019, Room to Read visited the Philippines and partnered with Philippine children’s literature publishers Adarna House, Lampara Books, Anvil Publishing, and OMF-Hiyas with the goal of producing contextualized stories that Filipino children can read. The result is a set of 20 storybooks developed by Filipino writers and illustrators, the author of this paper included. The project led to narratives of experiences in storybook production from conceptualization to publication, towards translations and reimagining in online repository, storytelling, and audiobook formats. During the production process, we were particularly reminded of gender representations, child’s rights, and telling stories that can empower the children in vulnerable communities, who are the beneficiaries of the project. The storybooks, along with many others produced in Asia and the world, are available online through the literacycloud.org website of Room to Read. In this study, the goal is to survey the stories produced in Asia and look at how gender is represented in the storybooks. By analyzing both the texts and the illustrations of the storybooks produced across Asian countries, themes of portrayals of young boys and girls, their characteristics and narratives, and how they are empowered in the stories are identified, with the goal of mapping how Room to Read is able to address the problem of access to literacy among young girls and ensuring them that they can do anything, the way they are portrayed in the stories. The paper hopes to determine how gender is represented in Asian storybooks produced by the international nonprofit organization Room to Read. Thematic textual analysis was used as methodology, where the storybooks are analyzed qualitatively to identify arising themes of gender representation. This study will shed light on the importance of responsible portrayal of gender in storybooks and how it can impact and empower children. The results of the study can also aid writers and illustrators in developing gender-sensitive storybooks.Keywords: room to read, asian storybooks, young girls, thematic analysis, child empowerment, literacy, education
Procedia PDF Downloads 791438 Clustering Performance Analysis using New Correlation-Based Cluster Validity Indices
Authors: Nathakhun Wiroonsri
Abstract:
There are various cluster validity measures used for evaluating clustering results. One of the main objectives of using these measures is to seek the optimal unknown number of clusters. Some measures work well for clusters with different densities, sizes and shapes. Yet, one of the weaknesses that those validity measures share is that they sometimes provide only one clear optimal number of clusters. That number is actually unknown and there might be more than one potential sub-optimal option that a user may wish to choose based on different applications. We develop two new cluster validity indices based on a correlation between an actual distance between a pair of data points and a centroid distance of clusters that the two points are located in. Our proposed indices constantly yield several peaks at different numbers of clusters which overcome the weakness previously stated. Furthermore, the introduced correlation can also be used for evaluating the quality of a selected clustering result. Several experiments in different scenarios, including the well-known iris data set and a real-world marketing application, have been conducted to compare the proposed validity indices with several well-known ones.Keywords: clustering algorithm, cluster validity measure, correlation, data partitions, iris data set, marketing, pattern recognition
Procedia PDF Downloads 1031437 Despiking of Turbulent Flow Data in Gravel Bed Stream
Authors: Ratul Das
Abstract:
The present experimental study insights the decontamination of instantaneous velocity fluctuations captured by Acoustic Doppler Velocimeter (ADV) in gravel-bed streams to ascertain near-bed turbulence for low Reynolds number. The interference between incidental and reflected pulses produce spikes in the ADV data especially in the near-bed flow zone and therefore filtering the data are very essential. Nortek’s Vectrino four-receiver ADV probe was used to capture the instantaneous three-dimensional velocity fluctuations over a non-cohesive bed. A spike removal algorithm based on the acceleration threshold method was applied to note the bed roughness and its influence on velocity fluctuations and velocity power spectra in the carrier fluid. The velocity power spectra of despiked signals with a best combination of velocity threshold (VT) and acceleration threshold (AT) are proposed which ascertained velocity power spectra a satisfactory fit with the Kolmogorov “–5/3 scaling-law” in the inertial sub-range. Also, velocity distributions below the roughness crest level fairly follows a third-degree polynomial series.Keywords: acoustic doppler velocimeter, gravel-bed, spike removal, reynolds shear stress, near-bed turbulence, velocity power spectra
Procedia PDF Downloads 3001436 Realistic Testing Procedure of Power Swing Blocking Function in Distance Relay
Authors: Farzad Razavi, Behrooz Taheri, Mohammad Parpaei, Mehdi Mohammadi Ghalesefidi, Siamak Zarei
Abstract:
As one of the major problems in protecting large-dimension power systems, power swing and its effect on distance have caused a lot of damages to energy transfer systems in many parts of the world. Therefore, power swing has gained attentions of many researchers, which has led to invention of different methods for power swing detection. Power swing detection algorithm is highly important in distance relay, but protection relays should have general requirements such as correct fault detection, response rate, and minimization of disturbances in a power system. To ensure meeting the requirements, protection relays need different tests during development, setup, maintenance, configuration, and troubleshooting steps. This paper covers power swing scheme of the modern numerical relay protection, 7sa522 to address the effect of the different fault types on the function of the power swing blocking. In this study, it was shown that the different fault types during power swing cause different time for unblocking distance relay.Keywords: power swing, distance relay, power system protection, relay test, transient in power system
Procedia PDF Downloads 3861435 Voice and Head Controlled Intelligent Wheelchair
Authors: Dechrit Maneetham
Abstract:
The aim of this paper was to design a void and head controlled electric power wheelchair (EPW). A novel activate the control system for quadriplegics with voice, head and neck mobility. Head movement has been used as a control interface for people with motor impairments in a range of applications. Acquiring measurements from the module is simplified through a synchronous a motor. Axis measures the two directions namely x and y. At the same time, patients can control the motorized wheelchair using voice signals (forward, backward, turn left, turn right, and stop) given by it self. The model of a dc motor is considered as a speed control by selection of a PID parameters using genetic algorithm. An experimental set-up constructed, which consists of micro controller as controller, a DC motor driven EPW and feedback elements. This paper is tuning methods of parameter for a pulse width modulation (PWM) control system. A speed controller has been designed successfully for closed loop of the dc motor so that the motor runs very closed to the reference speed and angle. Intelligent wheelchair can be used to ensure the person’s voice and head are attending the direction of travel asserted by a conventional, direction and speed control.Keywords: wheelchair, quadriplegia, rehabilitation , medical devices, speed control
Procedia PDF Downloads 5401434 Non-Local Simultaneous Sparse Unmixing for Hyperspectral Data
Authors: Fanqiang Kong, Chending Bian
Abstract:
Sparse unmixing is a promising approach in a semisupervised fashion by assuming that the observed pixels of a hyperspectral image can be expressed in the form of linear combination of only a few pure spectral signatures (end members) in an available spectral library. However, the sparse unmixing problem still remains a great challenge at finding the optimal subset of endmembers for the observed data from a large standard spectral library, without considering the spatial information. Under such circumstances, a sparse unmixing algorithm termed as non-local simultaneous sparse unmixing (NLSSU) is presented. In NLSSU, the non-local simultaneous sparse representation method for endmember selection of sparse unmixing, is used to finding the optimal subset of endmembers for the similar image patch set in the hyperspectral image. And then, the non-local means method, as a regularizer for abundance estimation of sparse unmixing, is used to exploit the abundance image non-local self-similarity. Experimental results on both simulated and real data demonstrate that NLSSU outperforms the other algorithms, with a better spectral unmixing accuracy.Keywords: hyperspectral unmixing, simultaneous sparse representation, sparse regression, non-local means
Procedia PDF Downloads 2481433 Optimization of Multiplier Extraction Digital Filter On FPGA
Authors: Shiksha Jain, Ramesh Mishra
Abstract:
One of the most widely used complex signals processing operation is filtering. The most important FIR digital filter are widely used in DSP for filtering to alter the spectrum according to some given specifications. Power consumption and Area complexity in the algorithm of Finite Impulse Response (FIR) filter is mainly caused by multipliers. So we present a multiplier less technique (DA technique). In this technique, precomputed value of inner product is stored in LUT. Which are further added and shifted with number of iterations equal to the precision of input sample. But the exponential growth of LUT with the order of FIR filter, in this basic structure, makes it prohibitive for many applications. The significant area and power reduction over traditional Distributed Arithmetic (DA) structure is presented in this paper, by the use of slicing of LUT to the desired length. An architecture of 16 tap FIR filter is presented, with different length of slice of LUT. The result of FIR Filter implementation on Xilinx ISE synthesis tool (XST) vertex-4 FPGA Tool by using proposed method shows the increase of the maximum frequency, the decrease of the resources as usage saving in area with more number of slices and the reduction dynamic power.Keywords: multiplier less technique, linear phase symmetric FIR filter, FPGA tool, look up table
Procedia PDF Downloads 3921432 Detection and Molecular Identification of Bacteria Forming Polyhydroxyalkanoate and Polyhydroxybutyrate Isolated from Soil in Saudi Arabia
Authors: Ali Bahkali, Rayan Yousef Booq, Mohammad Khiyami
Abstract:
Soil samples were collected from five different regions in the Kingdom of Saudi Arabia. Microbiological methods included dilution methods and pour plates to isolate and purify bacteria soil. The ability of isolates to develop biopolymer was investigated on petri dishes containing elements and substance concentrations stimulating developing biopolymer. Fluorescent stains, Nile red and Nile blue were used to stain the bacterial cells developing biopolymers. In addition, Sudan black was used to detect biopolymers in bacterial cells. The isolates which developed biopolymers were identified based on their gene sequence of 1 6sRNA and their ability to grow and synthesize PHAs on mineral medium supplemented with 1% dates molasses as the only carbon source under nitrogen limitation. During the study 293 bacterial isolates were isolated and detected. Through the initial survey on the petri dishes, 84 isolates showed the ability to develop biopolymers. These bacterial colonies developed a pink color due to accumulation of the biopolymers in the cells. Twenty-three isolates were able to grow on dates molasses, three strains of which showed the ability to accumulate biopolymers. These strains included Bacillus sp., Ralstonia sp. and Microbacterium sp. They were detected by Nile blue A stain with fluorescence microscopy (OLYMPUS IX 51). Among the isolated strains Ralstonia sp. was selected after its ability to grow on molasses dates in the presence of a limited nitrogen source was detected. The optimum conditions for formation of biopolymers by isolated strains were investigated. Conditions studied included, best incubation duration (2 days), temperature (30°C) and pH (7-8). The maximum PHB production was raised by 1% (v1v) when using concentrations of dates molasses 1, 2, 3, 4 and 5% in MSM. The best inoculated with 1% old inoculum (1= OD). The ideal extraction method of PHA and PHB proved to be 0.4% sodium hypochlorite solution, producing a quantity of polymer 98.79% of the cell's dry weight. The maximum PHB production was 1.79 g/L recorded by Ralstonia sp. after 48 h, while it was 1.40 g/L produced by R.eutropha ATCC 17697 after 48 h.Keywords: bacteria forming polyhydroxyalkanoate, detection, molecular, Saudi Arabia
Procedia PDF Downloads 3471431 Measuring Elemental Sulfur in Late Manually-Treated Grape Juice in Relation to Polyfunctional Mercaptan Formation in Sauvignon Blanc Wines
Authors: Bahareh Sarmadi, Paul A. Kilmartin, Leandro D. Araújo, Brandt P. Bastow
Abstract:
Aim: Sauvignon blanc is the most substantial variety cultivated in almost 62% of all producing vineyards of New Zealand. The popularity of New Zealand Sauvignon blanc is due to its unique taste. It is the most famous wine characterized by its aroma profile derived from mercaptans. 3-mercaptohexan-1-ol (3MH) and 3-mercaptohexyl acetate (3MHA) are two of the most important volatile mercaptans found in Sauvignon blanc wines. “Viticultural” and “Enological” factors such as machine-harvesting, the most common harvesting practice used in New Zealand, can be among the reasons for this distinct flavor. Elemental sulfur is commonly sprayed in the fields to protect berries against powdery mildew. Although it is not the only source of sulfur, this practice creates a source of elemental sulfur that can be transferred into the must and eventually into wines. Despite the clear effects of residual elemental sulfur present in the must on the quality and aroma of the final wines, its measurement before harvest or fermentation is not a regular practice in the wineries. This can be due to the lack of accessible and applicable methods for the equipment at most commercial wineries. This study aims to establish a relationship between the number and frequency of elemental sulfur applications and the concentration of polyfunctional mercaptans in the final wines. Methods: An apparatus was designed to reduce elemental sulfur to sulfide, then an ion-selective electrode to measure sulfide concentration. During harvest 2022, we explored a wider range of residual elemental sulfur levels than what typically applies in the vineyards. This has been done through later manual elemental sulfur applications in the vineyard. Additional sulfur applications were made 20, 10 and 5 days prior to harvesting the treated grapes, covering long and short pre-harvest intervals (PHI). The grapes were processed into juice and fermented into wine; then, they were analyzed to find the correlation between polyfunctional mercaptans concentrations in the wines and residual elemental sulfur in the juice samples. Results: The research showed that higher 3MH/3MHA was formed when elemental sulfur was applied more frequent in the vineyards and supported the proposed pathway in which elemental sulfur is a source of 3MH formation in wines.Keywords: sauvignon blanc, elemental sulfur, polyfunctional mercaptans, varietal thiols
Procedia PDF Downloads 1051430 Protocol for Dynamic Load Distributed Low Latency Web-Based Augmented Reality and Virtual Reality
Authors: Rohit T. P., Sahil Athrij, Sasi Gopalan
Abstract:
Currently, the content entertainment industry is dominated by mobile devices. As the trends slowly shift towards Augmented/Virtual Reality applications the computational demands on these devices are increasing exponentially and we are already reaching the limits of hardware optimizations. This paper proposes a software solution to this problem. By leveraging the capabilities of cloud computing we can offload the work from mobile devices to dedicated rendering servers that are way more powerful. But this introduces the problem of latency. This paper introduces a protocol that can achieve high-performance low latency Augmented/Virtual Reality experience. There are two parts to the protocol, 1) In-flight compression The main cause of latency in the system is the time required to transmit the camera frame from client to server. The round trip time is directly proportional to the amount of data transmitted. This can therefore be reduced by compressing the frames before sending. Using some standard compression algorithms like JPEG can result in minor size reduction only. Since the images to be compressed are consecutive camera frames there won't be a lot of changes between two consecutive images. So inter-frame compression is preferred. Inter-frame compression can be implemented efficiently using WebGL but the implementation of WebGL limits the precision of floating point numbers to 16bit in most devices. This can introduce noise to the image due to rounding errors, which will add up eventually. This can be solved using an improved interframe compression algorithm. The algorithm detects changes between frames and reuses unchanged pixels from the previous frame. This eliminates the need for floating point subtraction thereby cutting down on noise. The change detection is also improved drastically by taking the weighted average difference of pixels instead of the absolute difference. The kernel weights for this comparison can be fine-tuned to match the type of image to be compressed. 2) Dynamic Load distribution Conventional cloud computing architectures work by offloading as much work as possible to the servers, but this approach can cause a hit on bandwidth and server costs. The most optimal solution is obtained when the device utilizes 100% of its resources and the rest is done by the server. The protocol balances the load between the server and the client by doing a fraction of the computing on the device depending on the power of the device and network conditions. The protocol will be responsible for dynamically partitioning the tasks. Special flags will be used to communicate the workload fraction between the client and the server and will be updated in a constant interval of time ( or frames ). The whole of the protocol is designed so that it can be client agnostic. Flags are available to the client for resetting the frame, indicating latency, switching mode, etc. The server can react to client-side changes on the fly and adapt accordingly by switching to different pipelines. The server is designed to effectively spread the load and thereby scale horizontally. This is achieved by isolating client connections into different processes.Keywords: 2D kernelling, augmented reality, cloud computing, dynamic load distribution, immersive experience, mobile computing, motion tracking, protocols, real-time systems, web-based augmented reality application
Procedia PDF Downloads 771429 Application of Additive Manufacturing for Production of Optimum Topologies
Authors: Mahdi Mottahedi, Peter Zahn, Armin Lechler, Alexander Verl
Abstract:
Optimal topology of components leads to the maximum stiffness with the minimum material use. For the generation of these topologies, normally algorithms are employed, which tackle manufacturing limitations, at the cost of the optimal result. The global optimum result with penalty factor one, however, cannot be fabricated with conventional methods. In this article, an additive manufacturing method is introduced, in order to enable the production of global topology optimization results. For a benchmark, topology optimization with higher and lower penalty factors are performed. Different algorithms are employed in order to interpret the results of topology optimization with lower factors in many microstructure layers. These layers are then joined to form the final geometry. The algorithms’ benefits are then compared experimentally and numerically for the best interpretation. The findings demonstrate that by implementation of the selected algorithm, the stiffness of the components produced with this method is higher than what could have been produced by conventional techniques.Keywords: topology optimization, additive manufacturing, 3D-printer, laminated object manufacturing
Procedia PDF Downloads 3401428 Application of Artificial Neural Network for Prediction of Load-Haul-Dump Machine Performance Characteristics
Authors: J. Balaraju, M. Govinda Raj, C. S. N. Murthy
Abstract:
Every industry is constantly looking for enhancement of its day to day production and productivity. This can be possible only by maintaining the men and machinery at its adequate level. Prediction of performance characteristics plays an important role in performance evaluation of the equipment. Analytical and statistical approaches will take a bit more time to solve complex problems such as performance estimations as compared with software-based approaches. Keeping this in view the present study deals with an Artificial Neural Network (ANN) modelling of a Load-Haul-Dump (LHD) machine to predict the performance characteristics such as reliability, availability and preventive maintenance (PM). A feed-forward-back-propagation ANN technique has been used to model the Levenberg-Marquardt (LM) training algorithm. The performance characteristics were computed using Isograph Reliability Workbench 13.0 software. These computed values were validated using predicted output responses of ANN models. Further, recommendations are given to the industry based on the performed analysis for improvement of equipment performance.Keywords: load-haul-dump, LHD, artificial neural network, ANN, performance, reliability, availability, preventive maintenance
Procedia PDF Downloads 1511427 New Estimation in Autoregressive Models with Exponential White Noise by Using Reversible Jump MCMC Algorithm
Authors: Suparman Suparman
Abstract:
A white noise in autoregressive (AR) model is often assumed to be normally distributed. In application, the white noise usually do not follows a normal distribution. This paper aims to estimate a parameter of AR model that has a exponential white noise. A Bayesian method is adopted. A prior distribution of the parameter of AR model is selected and then this prior distribution is combined with a likelihood function of data to get a posterior distribution. Based on this posterior distribution, a Bayesian estimator for the parameter of AR model is estimated. Because the order of AR model is considered a parameter, this Bayesian estimator cannot be explicitly calculated. To resolve this problem, a method of reversible jump Markov Chain Monte Carlo (MCMC) is adopted. A result is a estimation of the parameter AR model can be simultaneously calculated.Keywords: autoregressive (AR) model, exponential white Noise, bayesian, reversible jump Markov Chain Monte Carlo (MCMC)
Procedia PDF Downloads 3561426 Organic Geochemical Evaluation of the Ecca Group Shale: Implications for Hydrocarbon Potential
Authors: Temitope L. Baiyegunhi, Kuiwu Liu, Oswald Gwavava, Christopher Baiyegunhi
Abstract:
Shale gas has recently been the exploration focus for future energy resource in South Africa. Specifically, the black shales of the lower Ecca Group in the study area are considered to be one of the most prospective targets for shale gas exploration. Evaluation of this potential resource has been restricted due to the lack of exploration and scarcity of existing drill core data. Thus, only limited previous geochemical data exist for these formations. In this study, outcrop and core samples of the Ecca Group were analysed to assess their total organic carbon (TOC), organic matter type, thermal maturity and hydrocarbon generation potential (SP). The results show that these rocks have TOC ranging from 0.11 to 7.35 wt.%. The SP values vary from 0.09 to 0.53 mg HC/g, suggesting poor hydrocarbon generative potential. The plot of S1 versus TOC shows that the source rocks were characterized by autochthonous hydrocarbons. S2/S3 values range between 0.40 and 7.5, indicating Type- II/III, III, and IV kerogen. With the exception of one sample from the collingham formation which has HI value of 53 mg HC/g TOC, all other samples have HI values of less than 50 mg HC/g TOC, thus suggesting Type-IV kerogen, which is mostly derived from reworked organic matter (mainly dead carbon) with little or no potential for hydrocarbon generation. Tmax values range from 318 to 601℃, indicating immature to over-maturity of hydrocarbon. The vitrinite reflectance values range from 2.22 to 3.93%, indicating over-maturity of the kerogen. Binary plots of HI against OI and HI versus Tmax show that the shales are of Type II and mixed Type II-III kerogen, which are capable of generating both natural gas and minor oil at suitable burial depth. Based on the geochemical data, it can be inferred that the source rocks are immature to over-matured variable from localities and have potential of producing wet to dry gas at present-stage. Generally, the Whitehill formation of the Ecca Group is comparable to the Marcellus and Barnett Shales. This further supports the assumption that the Whitehill Formation has a high probability of being a profitable shale gas play, but only when explored in dolerite-free area and away from the Cape Fold Belt.Keywords: source rock, organic matter type, thermal maturity, hydrocarbon generation potential, Ecca Group
Procedia PDF Downloads 1451425 Income and Factor Analysis of Small Scale Broiler Production in Imo State, Nigeria
Authors: Ubon Asuquo Essien, Okwudili Bismark Ibeagwa, Daberechi Peace Ubabuko
Abstract:
The Broiler Poultry subsector is dominated by small scale production with low aggregate output. The high cost of inputs currently experienced in Nigeria tends to aggravate the situation; hence many broiler farmers struggle to break-even. This study was designed to examine income and input factors in small scale deep liter broiler production in Imo state, Nigeria. Specifically, the study examined; socio-economic characteristics of small scale deep liter broiler producing Poultry farmers; estimate cost and returns of broiler production in the area; analyze input factors in broiler production in the area and examined marketability, age and profitability of the enterprise. A multi-stage sampling technique was adopted in selecting 60 small scale broiler farmers who use deep liter system from 6 communities through the use of structured questionnaire. The socioeconomic characteristics of the broiler farmers and the profitability/ marketability age of the birds were described using descriptive statistical tools such as frequencies, means and percentages. Gross margin analysis was used to analyze the cost and returns to broiler production, while Cobb Douglas production function was employed to analyze input factors in broiler production. The result of the study revealed that the cost of feed (P<0.1), deep liter material (P<0.05) and medication (P<0.05) had a significant positive relationship with the gross return of broiler farmers in the study area, while cost of labour, fuel and day old chicks were not significant. Furthermore, Gross profit margin of the farmers who market their broiler at the 8th week of rearing was 80.7%; and 78.7% and 60.8% for farmers who market at the 10th week and 12th week of rearing, respectively. The business is, therefore, profitable but at varying degree. Government and Development partners should make deliberate efforts to curb the current rise in the prices of poultry feeds, drugs and timber materials used as bedding so as to widen the profit margin and encourage more farmers to go into the business. The farmers equally need more technical assistance from extension agents with regards to timely and profitable marketing.Keywords: broilers, factor analysis, income, small scale
Procedia PDF Downloads 821424 The Application of Artificial Neural Networks for the Performance Prediction of Evacuated Tube Solar Air Collector with Phase Change Material
Authors: Sukhbir Singh
Abstract:
This paper describes the modeling of novel solar air collector (NSAC) system by using artificial neural network (ANN) model. The objective of the study is to demonstrate the application of the ANN model to predict the performance of the NSAC with acetamide as a phase change material (PCM) storage. Input data set consist of time, solar intensity and ambient temperature wherever as outlet air temperature of NSAC was considered as output. Experiments were conducted between 9.00 and 24.00 h in June and July 2014 underneath the prevailing atmospheric condition of Kurukshetra (city of the India). After that, experimental results were utilized to train the back propagation neural network (BPNN) to predict the outlet air temperature of NSAC. The results of proposed algorithm show that the BPNN is effective tool for the prediction of responses. The BPNN predicted results are 99% in agreement with the experimental results.Keywords: Evacuated tube solar air collector, Artificial neural network, Phase change material, solar air collector
Procedia PDF Downloads 1221423 Solid Waste Generation, Composition and Potentiality of Waste to Resource Recovery in Narayanganj City Corporation
Authors: Md. Jisan Ahmed, M. A. Taher
Abstract:
One of the cities in Bangladesh that is developing the fastest is Narayanganj City Corporation. In 2011, the municipality of Narayanganj was transformed into a city corporation, with 27 wards combining Kadamrasul Municipality, Siddhirganj Municipality, and Narayanganj Town. It is also one of Bangladesh's most important industrial centers in Bangladesh. Narayanganj City Corporation (NCC), which has had high development growth, is also generating more solid waste on a high per-capita basis. Because of the increasing rate of population expansion, business activity, industrial development, and fast urbanization, NCC is today creating more waste than ever before. The enormous amount of solid garbage produced in NCC is currently causing air pollution, soil contamination, water pollution, drainage system blockages, and an unpleasant urban environment. The study aimed to find out the amount of solid waste produced per day in NCC by exploring the waste composition and potentiality of resource recovery from the produced solid waste. This study considered household surveys, polythene bag surveys, questionnaire surveys in commercial and industrial sectors, KIIs, FGDs, and lab tests to identify the total amount of waste generated in NCC with waste composition and potentiality for energy recovery from the generated waste. This study has explored that NCC is producing about 922 tons of solid waste per day from households, commercial activities, and industrial sectors where the existing waste collection rate by NCC authority is only about 50% of total generated waste. This study has also explored that about 75% of daily-produced solid waste is perishable with comparatively high moisture content whereas 18 % and 7% are non-perishable and hazardous. It is also explored that there is no resource recovery plant for solid waste management in NCC. On the other hand, this study has explored that the calorific value of the produced solid waste favors resource recovery like waste to electricity. The generated solid waste composition is also in favor of waste-to-biogas, and waste-to-compost fertilizer production. This study has advocated that initiatives need to develop a solid waste management plant in NCC for resource recovery from solid waste. This research may provide a quick overview of the rate of solid waste generation, its composition, and the potential for resource recovery from solid waste in Bangladesh's metropolitan regions. It can also provide information and knowledge to other trash departments in different cities and municipalities in Bangladesh.Keywords: solid waste, waste composition, waste management, resource recovery from solid waste
Procedia PDF Downloads 241422 A Neural Network Modelling Approach for Predicting Permeability from Well Logs Data
Authors: Chico Horacio Jose Sambo
Abstract:
Recently neural network has gained popularity when come to solve complex nonlinear problems. Permeability is one of fundamental reservoir characteristics system that are anisotropic distributed and non-linear manner. For this reason, permeability prediction from well log data is well suited by using neural networks and other computer-based techniques. The main goal of this paper is to predict reservoir permeability from well logs data by using neural network approach. A multi-layered perceptron trained by back propagation algorithm was used to build the predictive model. The performance of the model on net results was measured by correlation coefficient. The correlation coefficient from testing, training, validation and all data sets was evaluated. The results show that neural network was capable of reproducing permeability with accuracy in all cases, so that the calculated correlation coefficients for training, testing and validation permeability were 0.96273, 0.89991 and 0.87858, respectively. The generalization of the results to other field can be made after examining new data, and a regional study might be possible to study reservoir properties with cheap and very fast constructed models.Keywords: neural network, permeability, multilayer perceptron, well log
Procedia PDF Downloads 405