Search results for: estimation after selection
1484 Analysis of Brain Activities due to Differences in Running Shoe Properties
Authors: Kei Okubo, Yosuke Kurihara, Takashi Kaburagi, Kajiro Watanabe
Abstract:
Many of the ever-growing elderly population require exercise, such as running, for health management. One important element of a runner’s training is the choice of shoes for exercise; shoes are important because they provide the interface between the feet and road. When we purchase shoes, we may instinctively choose a pair after trying on many different pairs of shoes. Selecting the shoes instinctively may work, but it does not guarantee a suitable fit for running activities. Therefore, if we could select suitable shoes for each runner from the viewpoint of brain activities, it would be helpful for validating shoe selection. In this paper, we describe how brain activities show different characteristics during particular task, corresponding to different properties of shoes. Using five subjects, we performed a verification experiment, applying weight, softness, and flexibility as shoe properties. In order to affect the shoe property’s differences to the brain, subjects run for ten min. Before and after running, subjects conducted a paced auditory serial addition task (PASAT) as the particular task; and the subjects’ brain activities during the PASAT are evaluated based on oxyhemoglobin and deoxyhemoglobin relative concentration changes, measured by near-infrared spectroscopy (NIRS). When the brain works actively, oxihemoglobin and deoxyhemoglobin concentration drastically changes; therefore, we calculate the maximum values of concentration changes. In order to normalize relative concentration changes after running, the maximum value are divided by before running maximum value as evaluation parameters. The classification of the groups of shoes is expressed on a self-organizing map (SOM). As a result, deoxyhemoglobin can make clusters for two of the three types of shoes.Keywords: brain activities, NIRS, PASAT, running shoes
Procedia PDF Downloads 3731483 Modeling and Numerical Simulation of Heat Transfer and Internal Loads at Insulating Glass Units
Authors: Nina Penkova, Kalin Krumov, Liliana Zashcova, Ivan Kassabov
Abstract:
The insulating glass units (IGU) are widely used in the advanced and renovated buildings in order to reduce the energy for heating and cooling. Rules for the choice of IGU to ensure energy efficiency and thermal comfort in the indoor space are well known. The existing of internal loads - gage or vacuum pressure in the hermetized gas space, requires additional attention at the design of the facades. The internal loads appear at variations of the altitude, meteorological pressure and gas temperature according to the same at the process of sealing. The gas temperature depends on the presence of coatings, coating position in the transparent multi-layer system, IGU geometry and space orientation, its fixing on the facades and varies with the climate conditions. An algorithm for modeling and numerical simulation of thermal fields and internal pressure in the gas cavity at insulating glass units as function of the meteorological conditions is developed. It includes models of the radiation heat transfer in solar and infrared wave length, indoor and outdoor convection heat transfer and free convection in the hermetized gas space, assuming the gas as compressible. The algorithm allows prediction of temperature and pressure stratification in the gas domain of the IGU at different fixing system. The models are validated by comparison of the numerical results with experimental data obtained by Hot-box testing. Numerical calculations and estimation of 3D temperature, fluid flow fields, thermal performances and internal loads at IGU in window system are implemented.Keywords: insulating glass units, thermal loads, internal pressure, CFD analysis
Procedia PDF Downloads 2731482 Desk Graffiti as Art, Archive or Collective Knowledge Sharing: A Case Study of Schools in Addis Ababa, Ethiopia
Authors: Behailu Bezabih Ayele
Abstract:
Illustrative expressions in art education and in overall learning are being given increasing attention in the transmission of knowledge. The objective of this paper, therefore, is to present an analysis of graffiti on school desks-a way of smuggling knowledge on the edge of classroom education and learning. The methodological approach focuses on the systematic collection and selection of desk graffiti. Four schools are chosen to reflect socioeconomic status and gender composition. The analysis focused on the categorization of graffiti by genre. This was followed by an analysis of the style, intensity as well as content of the messages in terms of overall social impacts. The paper grounds the analysis by reviewing the literature on modern education and art education in the Ethiopian context, as well as the place of desk graffiti. The findings generally show that the school desks and the school environment, by and large, have managed to serve as vessels through which formal and informal knowledge is acquired, transmitted, engrained into the students and transformed into messages by the students. The desks have also apparently served as a springboard to maximize the interfaces between several ideas and disciplines and communications. However, the very fact that the desks serve as massive channels of expression and knowledge transmission also points to a lack of breadth availability of channels of expression, perhaps confounding the ability of classrooms as means of outlet of expression and documentation for the students. This points to the need for efforts in education policy and funding of artistic endeavors for young students.Keywords: artistic expression, desk graffiti, education, school children, Ethiopia
Procedia PDF Downloads 681481 An Assessment of the Extent and Impact of Motor Insurance Fraud Claims in Nigeria
Authors: Olatokunbo Shoyemi, Mario Brito, Ian Dawson
Abstract:
In recent times, the Nigerian motor insurers have experienced high volume of motor insurance claim pay-outs and insignificant contribution to the net premium income of the Nigerian insurance market, which has been a major concern for the shareholders/stakeholders. It has been argued that there are many factors that have brought about these concerns. However, anecdotal evidence (ongoing debates among industry practitioners) suggests prevalence of fraud due to poor practices in motor insurance business in Nigeria. This study is therefore aimed to carry out an assessment of fraud in motor insurance claims as perceived by experts in the Nigerian insurance market. This study adopted a descriptive research design, and the analysis was built on a survey among insurance experts in Nigeria using a designed questionnaire. A purposive and snowball sampling were used to select our sample (N = 120) - representing a selection of all professionally qualified insurance experts in Nigeria insurance industry. The study found that Nigerian insurance experts (i) largely agree that there is a problematic level of fraud in the Nigerian motor insurance industry; (ii) perceive soft fraud to be about 3 times more common than hard fraud in the Nigerian motor insurance industry, and (iii) strongly agree there are problematic impacts from fraud on the solvency of the Nigerian motor insurers. This paper has provided an empirical understanding of the existence, extent, and impact of fraud risks within the Nigerian insurance market based on expert knowledge and insights rather than, as has often been the case, a reliance on individual anecdotes.Keywords: claims, net premium income, motor insurance, soft fraud, hard fraud
Procedia PDF Downloads 1081480 Technical, Environmental and Financial Assessment for Optimal Sizing of Run-of-River Small Hydropower Project: Case Study in Colombia
Authors: David Calderon Villegas, Thomas Kaltizky
Abstract:
Run-of-river (RoR) hydropower projects represent a viable, clean, and cost-effective alternative to dam-based plants and provide decentralized power production. However, RoR schemes cost-effectiveness depends on the proper selection of site and design flow, which is a challenging task because it requires multivariate analysis. In this respect, this study presents the development of an investment decision support tool for assessing the optimal size of an RoR scheme considering the technical, environmental, and cost constraints. The net present value (NPV) from a project perspective is used as an objective function for supporting the investment decision. The tool has been tested by applying it to an actual RoR project recently proposed in Colombia. The obtained results show that the optimum point in financial terms does not match the flow that maximizes energy generation from exploiting the river's available flow. For the case study, the flow that maximizes energy corresponds to a value of 5.1 m3/s. In comparison, an amount of 2.1 m3/s maximizes the investors NPV. Finally, a sensitivity analysis is performed to determine the NPV as a function of the debt rate changes and the electricity prices and the CapEx. Even for the worst-case scenario, the optimal size represents a positive business case with an NPV of 2.2 USD million and an IRR 1.5 times higher than the discount rate.Keywords: small hydropower, renewable energy, RoR schemes, optimal sizing, objective function
Procedia PDF Downloads 1321479 A Stochastic Diffusion Process Based on the Two-Parameters Weibull Density Function
Authors: Meriem Bahij, Ahmed Nafidi, Boujemâa Achchab, Sílvio M. A. Gama, José A. O. Matos
Abstract:
Stochastic modeling concerns the use of probability to model real-world situations in which uncertainty is present. Therefore, the purpose of stochastic modeling is to estimate the probability of outcomes within a forecast, i.e. to be able to predict what conditions or decisions might happen under different situations. In the present study, we present a model of a stochastic diffusion process based on the bi-Weibull distribution function (its trend is proportional to the bi-Weibull probability density function). In general, the Weibull distribution has the ability to assume the characteristics of many different types of distributions. This has made it very popular among engineers and quality practitioners, who have considered it the most commonly used distribution for studying problems such as modeling reliability data, accelerated life testing, and maintainability modeling and analysis. In this work, we start by obtaining the probabilistic characteristics of this model, as the explicit expression of the process, its trends, and its distribution by transforming the diffusion process in a Wiener process as shown in the Ricciaardi theorem. Then, we develop the statistical inference of this model using the maximum likelihood methodology. Finally, we analyse with simulated data the computational problems associated with the parameters, an issue of great importance in its application to real data with the use of the convergence analysis methods. Overall, the use of a stochastic model reflects only a pragmatic decision on the part of the modeler. According to the data that is available and the universe of models known to the modeler, this model represents the best currently available description of the phenomenon under consideration.Keywords: diffusion process, discrete sampling, likelihood estimation method, simulation, stochastic diffusion process, trends functions, bi-parameters weibull density function
Procedia PDF Downloads 3081478 Comparative Analysis of the Third Generation of Research Data for Evaluation of Solar Energy Potential
Authors: Claudineia Brazil, Elison Eduardo Jardim Bierhals, Luciane Teresa Salvi, Rafael Haag
Abstract:
Renewable energy sources are dependent on climatic variability, so for adequate energy planning, observations of the meteorological variables are required, preferably representing long-period series. Despite the scientific and technological advances that meteorological measurement systems have undergone in the last decades, there is still a considerable lack of meteorological observations that form series of long periods. The reanalysis is a system of assimilation of data prepared using general atmospheric circulation models, based on the combination of data collected at surface stations, ocean buoys, satellites and radiosondes, allowing the production of long period data, for a wide gamma. The third generation of reanalysis data emerged in 2010, among them is the Climate Forecast System Reanalysis (CFSR) developed by the National Centers for Environmental Prediction (NCEP), these data have a spatial resolution of 0.50 x 0.50. In order to overcome these difficulties, it aims to evaluate the performance of solar radiation estimation through alternative data bases, such as data from Reanalysis and from meteorological satellites that satisfactorily meet the absence of observations of solar radiation at global and/or regional level. The results of the analysis of the solar radiation data indicated that the reanalysis data of the CFSR model presented a good performance in relation to the observed data, with determination coefficient around 0.90. Therefore, it is concluded that these data have the potential to be used as an alternative source in locations with no seasons or long series of solar radiation, important for the evaluation of solar energy potential.Keywords: climate, reanalysis, renewable energy, solar radiation
Procedia PDF Downloads 2091477 An Approach to Correlate the Statistical-Based Lorenz Method, as a Way of Measuring Heterogeneity, with Kozeny-Carman Equation
Authors: H. Khanfari, M. Johari Fard
Abstract:
Dealing with carbonate reservoirs can be mind-boggling for the reservoir engineers due to various digenetic processes that cause a variety of properties through the reservoir. A good estimation of the reservoir heterogeneity which is defined as the quality of variation in rock properties with location in a reservoir or formation, can better help modeling the reservoir and thus can offer better understanding of the behavior of that reservoir. Most of reservoirs are heterogeneous formations whose mineralogy, organic content, natural fractures, and other properties vary from place to place. Over years, reservoir engineers have tried to establish methods to describe the heterogeneity, because heterogeneity is important in modeling the reservoir flow and in well testing. Geological methods are used to describe the variations in the rock properties because of the similarities of environments in which different beds have deposited in. To illustrate the heterogeneity of a reservoir vertically, two methods are generally used in petroleum work: Dykstra-Parsons permeability variations (V) and Lorenz coefficient (L) that are reviewed briefly in this paper. The concept of Lorenz is based on statistics and has been used in petroleum from that point of view. In this paper, we correlated the statistical-based Lorenz method to a petroleum concept, i.e. Kozeny-Carman equation and derived the straight line plot of Lorenz graph for a homogeneous system. Finally, we applied the two methods on a heterogeneous field in South Iran and discussed each, separately, with numbers and figures. As expected, these methods show great departure from homogeneity. Therefore, for future investment, the reservoir needs to be treated carefully.Keywords: carbonate reservoirs, heterogeneity, homogeneous system, Dykstra-Parsons permeability variations (V), Lorenz coefficient (L)
Procedia PDF Downloads 2201476 Good Governance in Perspective: An Example of Transition from Corruption towards Integrity within a Developing Country (Pakistan)
Authors: Saifullah Khalid
Abstract:
Governance and good governance are among the main topics in international discussions about the success factors for social and economic development. The image of developing countries as for example Pakistan in this respect is bad (in TI Corruption Index nr. among countries). Additionally, the police are among the sectors and organizations which are seen as most corrupt in many countries. However, in case of Pakistan there seem to be exceptions to the rule, and improvement can be brought in specific police departments. This paper represents the findings of Islamabad traffic police (ITP). In Pakistan, the police, in general, have been stigmatized for being the most corrupt department in the country. However, the few recent examples of Motorway police and its replicated model of Islamabad traffic police changed the perception about police and policing. These police forces have shown that Policing in Pakistan can be changed for better. In this paper, the research question that is addressed is: How corrupt are (traffic) police forces in Pakistan and what factors influence corruption within that police force? And What lessons can be learned from that to improve police integrity? Both qualitative and quantitative tools are utilized for data collection. The overall picture of the factors is not so easy to interpret and summarise. Nevertheless paying a better salary does not seem to limit integrity violations, neither does recruitment and selection and leadership, while supervision and control, training and stimulating the positive and limiting the negative elements of culture appear to be important in curbing (sometimes specific) integrity violations in the context of Pakistani police forces. The study also leads to a number of suggestions for curbing corruption and other integrity violations in the Pakistan police.Keywords: corruption control, governance, integrity violations, Islamabad traffic police, Pakistan
Procedia PDF Downloads 2161475 Design and Optimization of Sustainable Buildings by Combined Cooling, Heating and Power System (CCHP) Based on Exergy Analysis
Authors: Saeed Karimi, Ali Behbahaninia
Abstract:
In this study, the design and optimization of combined cooling, heating, and power system (CCHP) for a sustainable building are dealt with. Sustainable buildings are environmentally responsible and help us to save energy also reducing waste, pollution and environmental degradation. CCHP systems are widely used to save energy sources. In these systems, electricity, cooling, and heating are generating using just one primary energy source. The selection of the size of components based on the maximum demand of users will lead to an increase in the total cost of energy and equipment for the building complex. For this purpose, a system was designed in which the prime mover (gas turbine), heat recovery boiler, and absorption chiller are lower than the needed maximum. The difference in months with peak consumption is supplied with the help of electrical absorption chiller and auxiliary boiler (and the national electricity network). In this study, the optimum capacities of each of the equipment are determined based on Thermo economic method, in a way that the annual capital cost and energy consumption will be the lowest. The design was done for a gas turbine prime mover, and finally, the optimum designs were investigated using exergy analysis and were compared with a traditional energy supply system.Keywords: sustainable building, CCHP, energy optimization, gas turbine, exergy, thermo-economic
Procedia PDF Downloads 931474 Analysing Representations of ‘Leftover’ Women in Chinese Media: Taking the Film ‘The Last Woman Standing’ and ‘I Do’ as Examples
Authors: Ting Li Liu
Abstract:
‘Leftover woman’ or ‘3S’ woman is the term used to describe a well-educated, high income, independent woman who is single and never married around 30 years in Chinese society. With the naming of this demographic of ‘leftover women’, their family, dating culture, mate selection and marriage attract public concern. Massive media representations of ‘leftover women’ occur daily; the research aims to present several media representations of women’s anxiety about their singlehood and related marital issues around thirty. The research triangulates two areas of media representation of ‘leftover women’: films and audience reviews on ‘Douban Movie’ website. Drawing on traditional media studies, Fairclough’s critical discourse analysis combined with multimodal techniques is applied to the research to analyze the representations of ‘leftover women’ and their implications for marital culture in China, in conjunction with a feminist perspective. The conference paper will discuss two case studies: the film ‘The last woman standing’ and ‘I Do’. Paying attention to different aspects of ‘leftover women’, the research aims to re-examine the representations of ‘leftover women’ in selected scenes, such as their age anxiety, family, marriage, dating process, careers, etc. The paper also includes public beliefs about ‘leftover women’ from online audience reviews. In conclusion, the emergence of ‘leftover women’ is a reflection of Chinese tradition’s impact on people’s lives and new changes in Chinese families and their attitude to marriage.Keywords: leftover women, marriage, family, media culture, China
Procedia PDF Downloads 2541473 Applying Theory of Inventive Problem Solving to Develop Innovative Solutions: A Case Study
Authors: Y. H. Wang, C. C. Hsieh
Abstract:
Good service design can increase organization revenue and consumer satisfaction while reducing labor and time costs. The problems facing consumers in the original serve model for eyewear and optical industry includes the following issues: 1. Insufficient information on eyewear products 2. Passively dependent on recommendations, insufficient selection 3. Incomplete records on progression of vision conditions 4. Lack of complete customer records. This study investigates the case of Kobayashi Optical, applying the Theory of Inventive Problem Solving (TRIZ) to develop innovative solutions for eyewear and optical industry. Analysis results raise the following conclusions and management implications: In order to provide customers with improved professional information and recommendations, Kobayashi Optical is suggested to establish customer purchasing records. Overall service efficiency can be enhanced by applying data mining techniques to analyze past consumer preferences and purchase histories. Furthermore, Kobayashi Optical should continue to develop a 3D virtual trial service which can allow customers for easy browsing of different frame styles and colors. This 3D virtual trial service will save customer waiting times in during peak service times at stores.Keywords: theory of inventive problem solving (TRIZ), service design, augmented reality (AR), eyewear and optical industry
Procedia PDF Downloads 2791472 A Comparative Soft Computing Approach to Supplier Performance Prediction Using GEP and ANN Models: An Automotive Case Study
Authors: Seyed Esmail Seyedi Bariran, Khairul Salleh Mohamed Sahari
Abstract:
In multi-echelon supply chain networks, optimal supplier selection significantly depends on the accuracy of suppliers’ performance prediction. Different methods of multi criteria decision making such as ANN, GA, Fuzzy, AHP, etc have been previously used to predict the supplier performance but the “black-box” characteristic of these methods is yet a major concern to be resolved. Therefore, the primary objective in this paper is to implement an artificial intelligence-based gene expression programming (GEP) model to compare the prediction accuracy with that of ANN. A full factorial design with %95 confidence interval is initially applied to determine the appropriate set of criteria for supplier performance evaluation. A test-train approach is then utilized for the ANN and GEP exclusively. The training results are used to find the optimal network architecture and the testing data will determine the prediction accuracy of each method based on measures of root mean square error (RMSE) and correlation coefficient (R2). The results of a case study conducted in Supplying Automotive Parts Co. (SAPCO) with more than 100 local and foreign supply chain members revealed that, in comparison with ANN, gene expression programming has a significant preference in predicting supplier performance by referring to the respective RMSE and R-squared values. Moreover, using GEP, a mathematical function was also derived to solve the issue of ANN black-box structure in modeling the performance prediction.Keywords: Supplier Performance Prediction, ANN, GEP, Automotive, SAPCO
Procedia PDF Downloads 4191471 Bias Prevention in Automated Diagnosis of Melanoma: Augmentation of a Convolutional Neural Network Classifier
Authors: Kemka Ihemelandu, Chukwuemeka Ihemelandu
Abstract:
Melanoma remains a public health crisis, with incidence rates increasing rapidly in the past decades. Improving diagnostic accuracy to decrease misdiagnosis using Artificial intelligence (AI) continues to be documented. Unfortunately, unintended racially biased outcomes, a product of lack of diversity in the dataset used, with a noted class imbalance favoring lighter vs. darker skin tone, have increasingly been recognized as a problem.Resulting in noted limitations of the accuracy of the Convolutional neural network (CNN)models. CNN models are prone to biased output due to biases in the dataset used to train them. Our aim in this study was the optimization of convolutional neural network algorithms to mitigate bias in the automated diagnosis of melanoma. We hypothesized that our proposed training algorithms based on a data augmentation method to optimize the diagnostic accuracy of a CNN classifier by generating new training samples from the original ones will reduce bias in the automated diagnosis of melanoma. We applied geometric transformation, including; rotations, translations, scale change, flipping, and shearing. Resulting in a CNN model that provided a modifiedinput data making for a model that could learn subtle racial features. Optimal selection of the momentum and batch hyperparameter increased our model accuracy. We show that our augmented model reduces bias while maintaining accuracy in the automated diagnosis of melanoma.Keywords: bias, augmentation, melanoma, convolutional neural network
Procedia PDF Downloads 2111470 Bacteriological and Mineral Analyses of Leachate Samples from Erifun Dumpsite, Ado-Ekiti, Ekiti State, Nigeria
Authors: Adebowale T. Odeyemi, Oluwafemi A. Ajenifuja
Abstract:
The leachate samples collected from Erifun dumpsite along Federal Polythenic road, Ado-Ekiti, Ekiti State, were subjected to bacteriological and mineral analyses. The bacteriological estimation and isolation were done using serial dilution and pour plating techniques. Antibiotic susceptibility test was done using agar disc diffusion technique. Atomic Absorption Spectophotometry method was used to analyze the heavy metal contents in the leachate samples. The bacterial and coliform counts ranged from 4.2 × 105 CFU/ml to 2.97 × 106 CFU/ml and 5.0 × 104 CFU/ml to 2.45 x 106 CFU/ml, respectively. The isolated bacteria and percentage of occurrence include Bacillus cereus (22%), Enterobacter aerogenes (18%), Staphylococcus aureus (16%), Proteus vulgaris (14%), Escherichia coli (14%), Bacillus licheniformis (12%) and Klebsiella aerogenes (4%). The mineral value ranged as follow; iron (21.30mg/L - 25.60mg/L), zinc (1.80mg/L - 5.60mg/L), copper (1.00mg/L - 2.60mg/L), chromium (0.50mg/L - 1.30mg/L), candium (0.20mg/L - 1.30mg/L), nickel (0.20mg/L - 0.80mg/L), lead (0.05mg/L-0.30mg/L), cobalt (0.03mg/L - 0.30mg/L) and in all samples manganese was not detected. The entire organisms isolated exhibited a high level of resistance to most of the antibiotics used. There is an urgent need for awareness to be created about the present situation of the leachate in Erifun, on the need for treatment of the nearby stream and other water sources before they can be used for drinking and other domestic use. In conclusion, a good method of waste disposal is required in those communities to prevent leachate formation, percolation, and runoff into water bodies during the raining season.Keywords: antibiotic susceptibility, dumpsite, bacteriological analysis, heavy metal
Procedia PDF Downloads 1421469 Microscopic Analysis of Bulk, High-Tc Superconductors by Transmission Kikuchi Diffraction
Authors: Anjela Koblischka-Veneva, Michael R. Koblischka
Abstract:
In this contribution, the Transmission-Kikuchi Diffraction (TKD, or sometimes called t-EBSD) is applied to bulk, melt-grown YBa₂Cu₃O₇ (YBCO) superconductors prepared by the MTMG (melt-textured melt-grown) technique and the infiltration growth (IG) technique. TEM slices required for the analysis were prepared by means of Focused Ion-Beam (FIB) milling using mechanically polished sample surfaces, which enable a proper selection of the interesting regions for investigations. The required optical transparency was reached by an additional polishing step of the resulting surfaces using FIB-Ga-ion and Ar-ion milling. The improved spatial resolution of TKD enabled the investigation of the tiny YBa₂Cu₃O₅ (Y-211) particles having a diameter of about 50-100 nm embedded within the YBCO matrix and of other added secondary phase particles. With the TKD technique, the microstructural properties of the YBCO matrix are studied in detail. It is observed that the matrix shows the effects of stress/strain, depending on the size and distribution of the embedded particles, which are important for providing additional flux pinning centers in such superconducting bulk samples. Using the Kernel Average Misorientation (KAM) maps, the strain induced in the superconducting matrix around the particles, which increases the flux pinning effectivity, can be clearly revealed. This type of analysis of the EBSD/TKD data is, therefore, also important for other material systems, where nanoparticles are embedded in a matrix.Keywords: transmission Kikuchi diffraction, EBSD, TKD, embedded particles, superconductors YBa₂Cu₃O₇
Procedia PDF Downloads 1351468 Saudi Arabia Border Security Informatics: Challenges of a Harsh Environment
Authors: Syed Ahsan, Saleh Alshomrani, Ishtiaq Rasool, Ali Hassan
Abstract:
In this oral presentation, we will provide an overview of the technical and semantic architecture of a desert border security and critical infrastructure protection security system. Modern border security systems are designed to reduce the dependability and intrusion of human operators. To achieve this, different types of sensors are use along with video surveillance technologies. Application of these technologies in a harsh desert environment of Saudi Arabia poses unique challenges. Environmental and geographical factors including high temperatures, desert storms, temperature variations and remoteness adversely affect the reliability of surveillance systems. To successfully implement a reliable, effective system in a harsh desert environment, the following must be achieved: i) Selection of technology including sensors, video cameras, and communication infrastructure that suit desert environments. ii) Reduced power consumption and efficient usage of equipment to increase the battery life of the equipment. iii) A reliable and robust communication network with efficient usage of bandwidth. Also, to reduce the expert bottleneck, an ontology-based intelligent information systems needs to be developed. Domain knowledge unique and peculiar to Saudi Arabia needs to be formalized to develop an expert system that can detect abnormal activities and any intrusion.Keywords: border security, sensors, abnormal activity detection, ontologies
Procedia PDF Downloads 4811467 Optimized Brain Computer Interface System for Unspoken Speech Recognition: Role of Wernicke Area
Authors: Nassib Abdallah, Pierre Chauvet, Abd El Salam Hajjar, Bassam Daya
Abstract:
In this paper, we propose an optimized brain computer interface (BCI) system for unspoken speech recognition, based on the fact that the constructions of unspoken words rely strongly on the Wernicke area, situated in the temporal lobe. Our BCI system has four modules: (i) the EEG Acquisition module based on a non-invasive headset with 14 electrodes; (ii) the Preprocessing module to remove noise and artifacts, using the Common Average Reference method; (iii) the Features Extraction module, using Wavelet Packet Transform (WPT); (iv) the Classification module based on a one-hidden layer artificial neural network. The present study consists of comparing the recognition accuracy of 5 Arabic words, when using all the headset electrodes or only the 4 electrodes situated near the Wernicke area, as well as the selection effect of the subbands produced by the WPT module. After applying the articial neural network on the produced database, we obtain, on the test dataset, an accuracy of 83.4% with all the electrodes and all the subbands of 8 levels of the WPT decomposition. However, by using only the 4 electrodes near Wernicke Area and the 6 middle subbands of the WPT, we obtain a high reduction of the dataset size, equal to approximately 19% of the total dataset, with 67.5% of accuracy rate. This reduction appears particularly important to improve the design of a low cost and simple to use BCI, trained for several words.Keywords: brain-computer interface, speech recognition, artificial neural network, electroencephalography, EEG, wernicke area
Procedia PDF Downloads 2721466 Analysis of Tourism Development Level and Research on Improvement Strategies - Take Chongqing as an Example
Abstract:
As a member of the tertiary industry, tourism is an important driving factor for urban economic development. As a well-known tourist city in China, according to statistics, the added value of tourism and related industries in 2022 will reach 106.326 billion yuan, a year-on-year increase of 1.2%, accounting for 3.7% of the city's GDP. However, the overall tourism development level of Chongqing is seriously unbalanced, and the tourism strength of the main urban area is much higher than that of the southeast Chongqing, northeast Chongqing and the surrounding city tourism area, and the overall tourism strength of the other three regions is relatively balanced. Based on the estimation of tourism development level and the geographic detector method, this paper finds that the important factors affecting the tourism development level of non-main urban areas in Chongqing are A-level tourist attractions. Through GIS geospatial analysis technology and SPSS data correlation research method, the spatial distribution characteristics and influencing factors of A-level tourist attractions in Chongqing were quantitatively analyzed by using data such as geospatial data cloud, relevant documents of Chongqing Municipal Commission of Culture and Tourism Development, planning cloud, and relevant statistical yearbooks. The results show that: (1) The spatial distribution of tourist attractions in non-main urban areas of Chongqing is agglomeration and uneven. (2) The spatial distribution of A-level tourist attractions in non-main urban areas of Chongqing is affected by ecological factors, and the degree of influence is in the order of water factors> topographic factors > green space factors.Keywords: tourist attractions, geographic detectors, quantitative research, ecological factors, GIS technology, SPSS analysis
Procedia PDF Downloads 131465 Estimation of Carbon Uptake of Seoul City Street Trees in Seoul and Plans for Increase Carbon Uptake by Improving Species
Authors: Min Woo Park, Jin Do Chung, Kyu Yeol Kim, Byoung Uk Im, Jang Woo Kim, Hae Yeul Ryu
Abstract:
Nine representative species of trees among all the street trees were selected to estimate the absorption amount of carbon dioxide emitted from street trees in Seoul calculating the biomass, amount of carbon saved, and annual absorption amount of carbon dioxide in each of the species. Planting distance of street trees in Seoul was 1,851,180 m, the number of planting lines was 1,287, the number of planted trees was 284,498 and 46 species of trees were planted as of 2013. According to the result of plugging the quantity of species of street trees in Seoul on the absorption amount of each of the species, 120,097 ton of biomass, 60,049.8 ton of amount of carbon saved, and 11,294 t CO2/year of annual absorption amount of carbon dioxide were calculated. Street ratio mentioned on the road statistics in Seoul in 2022 is 23.13%. If the street trees are assumed to be increased in the same rate, the number of street trees in Seoul was calculated to be 294,823. The planting distance was estimated to be 1,918,360 m, and the annual absorption amount of carbon dioxide was measured to be 11,704 t CO2/year. Plans for improving the annual absorption amount of carbon dioxide from street trees were established based on the expected amount of absorption. First of all, it is to improve the annual absorption amount of carbon dioxide by increasing the number of planted street trees after adjusting the planting distance of street trees. If adjusting the current planting distance to 6 m, it was turned out that 12,692.7 t CO2/year was absorbed on an annual basis. Secondly, it is to change the species of trees to tulip trees that represent high absorption rate. If increasing the proportion of tulip trees to 30% up to 2022, the annual absorption rate of carbon dioxide was calculated to be 17804.4 t CO2/year.Keywords: absorption of carbon dioxide, source of absorbing carbon dioxide, trees in city, improving species
Procedia PDF Downloads 3621464 Analysis of Earthquake Potential and Shock Level Scenarios in South Sulawesi
Authors: Takhul Bakhtiar
Abstract:
In South Sulawesi Province, there is an active Walanae Fault causing this area to frequently experience earthquakes. This study aims to determine the level of seismicity of the earthquake in order to obtain the potential for earthquakes in the future. The estimation of the potential for earthquakes is then made a scenario model determine the estimated level of shocks as an effort to mitigate earthquake disasters in the region. The method used in this study is the Gutenberg Richter Method through the statistical likelihood approach. This study used earthquake data in the South Sulawesi region in 1972 - 2022. The research location is located at the coordinates of 3.5° – 5.5° South Latitude and 119.5° – 120.5° East Longitude and divided into two segments, namely the northern segment at the coordinates of 3.5° – 4.5° South Latitude and 119,5° – 120,5° East Longitude then the southern segment with coordinates of 4.5° – 5.5° South Latitude and 119,5° – 120.5° East Longitude. This study uses earthquake parameters with a magnitude > 1 and a depth < 50 km. The results of the analysis show that the potential for earthquakes in the next ten years with a magnitude of M = 7 in the northern segment is estimated at 98.81% with an estimated shock level of VI-VII MMI around the cities of Pare-Pare, Barru, Pinrang and Soppeng then IV - V MMI in the cities of Bulukumba, Selayar, Makassar and Gowa. In the southern segment, the potential for earthquakes in the next ten years with a magnitude of M = 7 is estimated at 32.89% with an estimated VI-VII MMI shock level in the cities of Bulukumba, Selayar, Makassar and Gowa, then III-IV MMI around the cities of Pare-Pare, Barru, Pinrang and Soppeng.Keywords: Gutenberg Richter, likelihood method, seismicity, shakemap and MMI scale
Procedia PDF Downloads 1201463 Aerodynamic Modeling Using Flight Data at High Angle of Attack
Authors: Rakesh Kumar, A. K. Ghosh
Abstract:
The paper presents the modeling of linear and nonlinear longitudinal aerodynamics using real flight data of Hansa-3 aircraft gathered at low and high angles of attack. The Neural-Gauss-Newton (NGN) method has been applied to model the linear and nonlinear longitudinal dynamics and estimate parameters from flight data. Unsteady aerodynamics due to flow separation at high angles of attack near stall has been included in the aerodynamic model using Kirchhoff’s quasi-steady stall model. NGN method is an algorithm that utilizes Feed Forward Neural Network (FFNN) and Gauss-Newton optimization to estimate the parameters and it does not require any a priori postulation of mathematical model or solving of equations of motion. NGN method was validated on real flight data generated at moderate angles of attack before application to the data at high angles of attack. The estimates obtained from compatible flight data using NGN method were validated by comparing with wind tunnel values and the maximum likelihood estimates. Validation was also carried out by comparing the response of measured motion variables with the response generated by using estimates a different control input. Next, NGN method was applied to real flight data generated by executing a well-designed quasi-steady stall maneuver. The results obtained in terms of stall characteristics and aerodynamic parameters were encouraging and reasonably accurate to establish NGN as a method for modeling nonlinear aerodynamics from real flight data at high angles of attack.Keywords: parameter estimation, NGN method, linear and nonlinear, aerodynamic modeling
Procedia PDF Downloads 4451462 Impact of Vehicle Travel Characteristics on Level of Service: A Comparative Analysis of Rural and Urban Freeways
Authors: Anwaar Ahmed, Muhammad Bilal Khurshid, Samuel Labi
Abstract:
The effect of trucks on the level of service is determined by considering passenger car equivalents (PCE) of trucks. The current version of Highway Capacity Manual (HCM) uses a single PCE value for all tucks combined. However, the composition of truck traffic varies from location to location; therefore a single PCE-value for all trucks may not correctly represent the impact of truck traffic at specific locations. Consequently, present study developed separate PCE values for single-unit and combination trucks to replace the single value provided in the HCM on different freeways. Site specific PCE values, were developed using concept of spatial lagging headways (the distance from the rear bumper of a leading vehicle to the rear bumper of the following vehicle) measured from field traffic data. The study used data from four locations on a single urban freeway and three different rural freeways in Indiana. Three-stage-least-squares (3SLS) regression techniques were used to generate models that predicted lagging headways for passenger cars, single unit trucks (SUT), and combination trucks (CT). The estimated PCE values for single-unit and combination truck for basic urban freeways (level terrain) were: 1.35 and 1.60, respectively. For rural freeways the estimated PCE values for single-unit and combination truck were: 1.30 and 1.45, respectively. As expected, traffic variables such as vehicle flow rates and speed have significant impacts on vehicle headways. Study results revealed that the use of separate PCE values for different truck classes can have significant influence on the LOS estimation.Keywords: level of service, capacity analysis, lagging headway, trucks
Procedia PDF Downloads 3551461 A Static Android Malware Detection Based on Actual Used Permissions Combination and API Calls
Authors: Xiaoqing Wang, Junfeng Wang, Xiaolan Zhu
Abstract:
Android operating system has been recognized by most application developers because of its good open-source and compatibility, which enriches the categories of applications greatly. However, it has become the target of malware attackers due to the lack of strict security supervision mechanisms, which leads to the rapid growth of malware, thus bringing serious safety hazards to users. Therefore, it is critical to detect Android malware effectively. Generally, the permissions declared in the AndroidManifest.xml can reflect the function and behavior of the application to a large extent. Since current Android system has not any restrictions to the number of permissions that an application can request, developers tend to apply more than actually needed permissions in order to ensure the successful running of the application, which results in the abuse of permissions. However, some traditional detection methods only consider the requested permissions and ignore whether it is actually used, which leads to incorrect identification of some malwares. Therefore, a machine learning detection method based on the actually used permissions combination and API calls was put forward in this paper. Meanwhile, several experiments are conducted to evaluate our methodology. The result shows that it can detect unknown malware effectively with higher true positive rate and accuracy while maintaining a low false positive rate. Consequently, the AdaboostM1 (J48) classification algorithm based on information gain feature selection algorithm has the best detection result, which can achieve an accuracy of 99.8%, a true positive rate of 99.6% and a lowest false positive rate of 0.Keywords: android, API Calls, machine learning, permissions combination
Procedia PDF Downloads 3291460 Artificial intelligence and Law
Authors: Mehrnoosh Abouzari, Shahrokh Shahraei
Abstract:
With the development of artificial intelligence in the present age, intelligent machines and systems have proven their actual and potential capabilities and are mindful of increasing their presence in various fields of human life in the fields of industry, financial transactions, marketing, manufacturing, service affairs, politics, economics and various branches of the humanities .Therefore, despite the conservatism and prudence of law enforcement, the traces of artificial intelligence can be seen in various areas of law. Including judicial robotics capability estimation, intelligent judicial decision making system, intelligent defender and attorney strategy adjustment, dissemination and regulation of different and scattered laws in each case to achieve judicial coherence and reduce opinion, reduce prolonged hearing and discontent compared to the current legal system with designing rule-based systems, case-based, knowledge-based systems, etc. are efforts to apply AI in law. In this article, we will identify the ways in which AI is applied in its laws and regulations, identify the dominant concerns in this area and outline the relationship between these two areas in order to answer the question of how artificial intelligence can be used in different areas of law and what the implications of this application will be. The authors believe that the use of artificial intelligence in the three areas of legislative, judiciary and executive power can be very effective in governments' decisions and smart governance, and helping to reach smart communities across human and geographical boundaries that humanity's long-held dream of achieving is a global village free of violence and personalization and human error. Therefore, in this article, we are going to analyze the dimensions of how to use artificial intelligence in the three legislative, judicial and executive branches of government in order to realize its application.Keywords: artificial intelligence, law, intelligent system, judge
Procedia PDF Downloads 1191459 Consumer Behavior and Knowledge on Organic Products in Thailand
Authors: Warunpun Kongsom, Chaiwat Kongsom
Abstract:
The objective of this study was to investigate the awareness, knowledge and consumer behavior towards organic products in Thailand. For this study, a purposive sampling technique was used to identify a sample group of 2,575 consumers over the age of 20 years who intended or made purchases from 1) green shops; 2) supermarkets with branches; and, 3) green markets. A questionnaire was used for data collection across the country. Descriptive statistics were used for data analysis. The results showed that more than 92% of consumers were aware of organic agriculture, but had less knowledge about it. More than 60% of consumers knew that organic agriculture production and processing did not allow the use of chemicals. And about 40% of consumers were confused between the food safety logo and the certified organic logo, and whether GMO was allowed in organic agriculture practice or not. In addition, most consumers perceived that organic agricultural products, good agricultural practice (GAP) products, agricultural chemicals free products, and hydroponic vegetable products had the same standard. In the view of organic consumers, the organic Thailand label was the most seen and reliable among various organic labels. Less than 3% of consumers thought that the International Federation of Organic Agriculture Movements (IFOAM) Global Organic Mark (GOM) was the most seen and reliable. For the behaviors of organic consumers, they purchased organic products mainly at the supermarket and green shop (55.4%), one to two times per month, and with a total expenditure of about 200 to 400 baht each time. The main reason for buying organic products was safety and free from agricultural chemicals. The considered factors in organic product selection were price (29.5%), convenience (22.4%), and a reliable certification system (21.3%). The demands for organic products were mainly rice, vegetables and fruits. Processed organic products were relatively small in quantity.Keywords: consumer behavior, consumer knowledge, organic products, Thailand
Procedia PDF Downloads 2961458 Detection and Classification of Mammogram Images Using Principle Component Analysis and Lazy Classifiers
Authors: Rajkumar Kolangarakandy
Abstract:
Feature extraction and selection is the primary part of any mammogram classification algorithms. The choice of feature, attribute or measurements have an important influence in any classification system. Discrete Wavelet Transformation (DWT) coefficients are one of the prominent features for representing images in frequency domain. The features obtained after the decomposition of the mammogram images using wavelet transformations have higher dimension. Even though the features are higher in dimension, they were highly correlated and redundant in nature. The dimensionality reduction techniques play an important role in selecting the optimum number of features from the higher dimension data, which are highly correlated. PCA is a mathematical tool that reduces the dimensionality of the data while retaining most of the variation in the dataset. In this paper, a multilevel classification of mammogram images using reduced discrete wavelet transformation coefficients and lazy classifiers is proposed. The classification is accomplished in two different levels. In the first level, mammogram ROIs extracted from the dataset is classified as normal and abnormal types. In the second level, all the abnormal mammogram ROIs is classified into benign and malignant too. A further classification is also accomplished based on the variation in structure and intensity distribution of the images in the dataset. The Lazy classifiers called Kstar, IBL and LWL are used for classification. The classification results obtained with the reduced feature set is highly promising and the result is also compared with the performance obtained without dimension reduction.Keywords: PCA, wavelet transformation, lazy classifiers, Kstar, IBL, LWL
Procedia PDF Downloads 3351457 System-level Factors, Presidential Coattails and Mass Preferences: Dynamics of Party Nationalization in Contemporary Brazil (1990-2014)
Authors: Kazuma Mizukoshi
Abstract:
Are electoral politics in contemporary Brazil still local in organization and focus? The importance of this question lies in its paradoxical trajectories. First, often coupled with institutional and sociological ‘barriers’ (e.g. the selection and election of candidates relatively loyal to the local party leadership, the predominance of territorialized electoral campaigns, and the resilience of political clientelism), the regionalization of electoral politics has been a viable and practical solution especially for pragmatic politicians in some Latin American countries. On the other hand, some leftist parties that once served as minor opposition forces at the time of foundational or initial elections have certainly expanded vote shares. Some were eventually capable of holding most (if not a majority) legislative seats since the 1990s. Though not yet rigorously demonstrated, theoretically implicit in the rise of leftist parties in legislative elections is the gradual (if not complete) nationalization of electoral support—meaning the growing equality of a party’s vote share across electoral districts and its change over time. This study will develop four hypotheses to explain the dynamics of party nationalization in contemporary Brazil: district magnitude, ethnic and class fractionalization of each district, voting intentions in federal and state executive elections, and finally the left-right stances of electorates. The study will demonstrate these hypotheses by closely working with the Brazilian Electoral Study (2002-2014).Keywords: party nationalization, presidential coattails, Left, Brazil
Procedia PDF Downloads 1381456 Development and Validation of Selective Methods for Estimation of Valaciclovir in Pharmaceutical Dosage Form
Authors: Eman M. Morgan, Hayam M. Lotfy, Yasmin M. Fayez, Mohamed Abdelkawy, Engy Shokry
Abstract:
Two simple, selective, economic, safe, accurate, precise and environmentally friendly methods were developed and validated for the quantitative determination of valaciclovir (VAL) in the presence of its related substances R1 (acyclovir), R2 (guanine) in bulk powder and in the commercial pharmaceutical product containing the drug. Method A is a colorimetric method where VAL selectively reacts with ferric hydroxamate and the developed color was measured at 490 nm over a concentration range of 0.4-2 mg/mL with percentage recovery 100.05 ± 0.58 and correlation coefficient 0.9999. Method B is a reversed phase ultra performance liquid chromatographic technique (UPLC) which is considered superior in technology to the high-performance liquid chromatography with respect to speed, resolution, solvent consumption, time, and cost of analysis. Efficient separation was achieved on Agilent Zorbax CN column using ammonium acetate (0.1%) and acetonitrile as a mobile phase in a linear gradient program. Elution time for the separation was less than 5 min and ultraviolet detection was carried out at 256 nm over a concentration range of 2-50 μg/mL with mean percentage recovery 100.11±0.55 and correlation coefficient 0.9999. The proposed methods were fully validated as per International Conference on Harmonization specifications and effectively applied for the analysis of valaciclovir in pure form and tablets dosage form. Statistical comparison of the results obtained by the proposed and official or reported methods revealed no significant difference in the performance of these methods regarding the accuracy and precision respectively.Keywords: hydroxamic acid, related substances, UPLC, valaciclovir
Procedia PDF Downloads 2471455 Automatic Multi-Label Image Annotation System Guided by Firefly Algorithm and Bayesian Method
Authors: Saad M. Darwish, Mohamed A. El-Iskandarani, Guitar M. Shawkat
Abstract:
Nowadays, the amount of available multimedia data is continuously on the rise. The need to find a required image for an ordinary user is a challenging task. Content based image retrieval (CBIR) computes relevance based on the visual similarity of low-level image features such as color, textures, etc. However, there is a gap between low-level visual features and semantic meanings required by applications. The typical method of bridging the semantic gap is through the automatic image annotation (AIA) that extracts semantic features using machine learning techniques. In this paper, a multi-label image annotation system guided by Firefly and Bayesian method is proposed. Firstly, images are segmented using the maximum variance intra cluster and Firefly algorithm, which is a swarm-based approach with high convergence speed, less computation rate and search for the optimal multiple threshold. Feature extraction techniques based on color features and region properties are applied to obtain the representative features. After that, the images are annotated using translation model based on the Net Bayes system, which is efficient for multi-label learning with high precision and less complexity. Experiments are performed using Corel Database. The results show that the proposed system is better than traditional ones for automatic image annotation and retrieval.Keywords: feature extraction, feature selection, image annotation, classification
Procedia PDF Downloads 586