Search results for: fuzzy logic based analysis
45146 Sensitive Analysis of the ZF Model for ABC Multi Criteria Inventory Classification
Authors: Makram Ben Jeddou
Abstract:
The ABC classification is widely used by managers for inventory control. The classical ABC classification is based on the Pareto principle and according to the criterion of the annual use value only. Single criterion classification is often insufficient for a closely inventory control. Multi-criteria inventory classification models have been proposed by researchers in order to take into account other important criteria. From these models, we will consider the ZF model in order to make a sensitive analysis on the composite score calculated for each item. In fact, this score based on a normalized average between a good and a bad optimized index can affect the ABC items classification. We will then focus on the weights assigned to each index and propose a classification compromise.Keywords: ABC classification, multi criteria inventory classification models, ZF-model
Procedia PDF Downloads 50845145 A Novel Heuristic for Analysis of Large Datasets by Selecting Wrapper-Based Features
Authors: Bushra Zafar, Usman Qamar
Abstract:
Large data sample size and dimensions render the effectiveness of conventional data mining methodologies. A data mining technique are important tools for collection of knowledgeable information from variety of databases and provides supervised learning in the form of classification to design models to describe vital data classes while structure of the classifier is based on class attribute. Classification efficiency and accuracy are often influenced to great extent by noisy and undesirable features in real application data sets. The inherent natures of data set greatly masks its quality analysis and leave us with quite few practical approaches to use. To our knowledge first time, we present a new approach for investigation of structure and quality of datasets by providing a targeted analysis of localization of noisy and irrelevant features of data sets. Machine learning is based primarily on feature selection as pre-processing step which offers us to select few features from number of features as a subset by reducing the space according to certain evaluation criterion. The primary objective of this study is to trim down the scope of the given data sample by searching a small set of important features which may results into good classification performance. For this purpose, a heuristic for wrapper-based feature selection using genetic algorithm and for discriminative feature selection an external classifier are used. Selection of feature based on its number of occurrence in the chosen chromosomes. Sample dataset has been used to demonstrate proposed idea effectively. A proposed method has improved average accuracy of different datasets is about 95%. Experimental results illustrate that proposed algorithm increases the accuracy of prediction of different diseases.Keywords: data mining, generic algorithm, KNN algorithms, wrapper based feature selection
Procedia PDF Downloads 31645144 Finite Difference Method of the Seismic Analysis of Earth Dam
Authors: Alaoua Bouaicha, Fahim Kahlouche, Abdelhamid Benouali
Abstract:
Many embankment dams have suffered failures during earthquakes due to the increase of pore water pressure under seismic loading. After analyzing of the behavior of embankment dams under severe earthquakes, major advances have been attained in the understanding of the seismic action on dams. The present study concerns numerical analysis of the seismic response of earth dams. The procedure uses a nonlinear stress-strain relation incorporated into the code FLAC2D based on the finite difference method. This analysis provides the variation of the pore water pressure and horizontal displacement.Keywords: Earthquake, Numerical Analysis, FLAC2D, Displacement, Embankment Dam, Pore Water Pressure
Procedia PDF Downloads 37845143 The History of Sambipitu Formation Temperature during the Early Miocene Epooch at Kali Ngalang, Nglipar, Gunung Kidul Regency
Authors: R. Harman Dwi, Ryan Avirsa, P. Abraham Ivan
Abstract:
Understanding of temperatures in the past, present, and future temperatures can be possible to do by analysis abundance of fossil foraminifera. This research was conducted in Sambipitu Formation, Ngalang River, Nglipar, Gunung Kidul Regency. The research method is divided into 3 stages: 1) study of literature, research based on previous researchers, 2) spatial, observation and sampling every 5-10 meters, 3) descriptive, analyzing samples consisting of a 10-gram sample weight, washing sample using 30% peroxide, biostratigraphy analysis, paleotemperature analysis using abundance of fossil, diversity analysis using Simpson diversity index method, and comparing current temperature data. There are two phases based on the appearance of Globorotalia menardii and Pulleniatina obliqueculata pointed to Phase Tropical Area, and the appearance of fossil Globigerinoides ruber and Orbulina universa fossil shows the phase of Subtropical Area. Paleotemperatur based on the appearance of Globorotalia menardii, Globigerinoides trilobus, Globigerinoides ruber, Orbulina universa, and Pulleniatina obliqueculata pointed to Warm Water Area and Warm Water Area (average surface water approximate 25°C).Keywords: abundance, biostratigraphy, Simpson diversity index method, paleotemperature
Procedia PDF Downloads 17245142 Feature Evaluation Based on Random Subspace and Multiple-K Ensemble
Authors: Jaehong Yu, Seoung Bum Kim
Abstract:
Clustering analysis can facilitate the extraction of intrinsic patterns in a dataset and reveal its natural groupings without requiring class information. For effective clustering analysis in high dimensional datasets, unsupervised dimensionality reduction is an important task. Unsupervised dimensionality reduction can generally be achieved by feature extraction or feature selection. In many situations, feature selection methods are more appropriate than feature extraction methods because of their clear interpretation with respect to the original features. The unsupervised feature selection can be categorized as feature subset selection and feature ranking method, and we focused on unsupervised feature ranking methods which evaluate the features based on their importance scores. Recently, several unsupervised feature ranking methods were developed based on ensemble approaches to achieve their higher accuracy and stability. However, most of the ensemble-based feature ranking methods require the true number of clusters. Furthermore, these algorithms evaluate the feature importance depending on the ensemble clustering solution, and they produce undesirable evaluation results if the clustering solutions are inaccurate. To address these limitations, we proposed an ensemble-based feature ranking method with random subspace and multiple-k ensemble (FRRM). The proposed FRRM algorithm evaluates the importance of each feature with the random subspace ensemble, and all evaluation results are combined with the ensemble importance scores. Moreover, FRRM does not require the determination of the true number of clusters in advance through the use of the multiple-k ensemble idea. Experiments on various benchmark datasets were conducted to examine the properties of the proposed FRRM algorithm and to compare its performance with that of existing feature ranking methods. The experimental results demonstrated that the proposed FRRM outperformed the competitors.Keywords: clustering analysis, multiple-k ensemble, random subspace-based feature evaluation, unsupervised feature ranking
Procedia PDF Downloads 33945141 Human Capital Discourse and Higher Education Policy
Authors: Tien-Hui Chiang
Abstract:
Human capital discourse encourages many countries to expand the capacity of HEIs. Along with this expansion, the higher education system is redefined as a free market and in turn it is privatized and commercialized. However, the state’s role in education is to balance social justice and capital accumulation. This role is further regulated by a specific form of neoliberalism constituted by social contexts. These correlations call for exploring the influence of human capital discourse on interwoven issues, such as the state’s role in education, higher education policy, and employability. Method: According to the perspective of neoliberal governmentality, answers to the above four research questions are likely to be embedded within discourses in documents related to higher education policies. Consequently, this study adopts a qualitative approach by analyzing official documents, including government reports, official statistics, circulars and official statements. Documents were collected and subjected to content analysis, with a particular focus on the period from 2005 to 2021. The technique of content analysis was applied to decode keywords and core concepts of these documents. Findings: Neoliberalism is exerted through human capital discourse in China particularly in the changes in higher education policies moving from quantitative expansion to quality control via employment or employability. Such changes highlight that the principle of “n”eoliberalism is more suitable for illustrating the practice of free market logic in different social contexts. The modifications of neoliberalism adopted by the Chinese government reflect that the state’s mission is to secure social security or the common good, so that public managerialism - in the form of programs for employment, internship and entrepreneurship - is adopted in the name of the public interest and the collective mission. Public managerialism now is not only targeted towards social institutions but the population more generally, incarnated here by college graduates. Its practice is not only to renovate organizational cultures but to activate people’s commitment to national development.Keywords: employability, higher education expansion, neoliberalism, human capital discourse
Procedia PDF Downloads 7845140 Effectiveness of Damping Devices on Coupling Beams of 15-story Building Based on Nonlinear Analysis Procedures
Authors: Galih Permana, Yuskar Lase
Abstract:
In recent years, damping device has been experimentally studied to replace diagonally reinforced coupling beams, to mitigate rebar congestion problem. This study focuses on evaluating the effectiveness of various damping devices in a high-rise building. The type of damping devices evaluated is Viscoelastic Damper (VCD) and Rotational Friction Damper (RFD), with study case of a 15-story reinforced concrete apartment building with a dual system (column-beam and shear walls). The analysis used is a nonlinear time history analysis with 11 pairs of ground motions matched to the Indonesian response spectrum based on ASCE 41-17 and ASCE 7-16. In this analysis, each damper will be varied with a different position, namely the first model, the damper will be installed on the entire floor and in the second model, the damper will be installed on the 5th floor to the 9th floor, which is the floor with the largest drift. The results show that the model using both dampers increases the level of structural performance both globally and locally in the building, which will reduce the level of damage to the structural elements. But between the two dampers, the coupling beam that uses RFD is more effective than using VCD in improving building performance. The damper on the coupling beam has a good role in dissipating earthquakes and also in terms of ease of installation.Keywords: building, coupling beam, damper, nonlinear time history analysis
Procedia PDF Downloads 17245139 Prediction of Mechanical Strength of Multiscale Hybrid Reinforced Cementitious Composite
Authors: Salam Alrekabi, A. B. Cundy, Mohammed Haloob Al-Majidi
Abstract:
Novel multiscale hybrid reinforced cementitious composites based on carbon nanotubes (MHRCC-CNT), and carbon nanofibers (MHRCC-CNF) are new types of cement-based material fabricated with micro steel fibers and nanofilaments, featuring superior strain hardening, ductility, and energy absorption. This study focused on established models to predict the compressive strength, and direct and splitting tensile strengths of the produced cementitious composites. The analysis was carried out based on the experimental data presented by the previous author’s study, regression analysis, and the established models that available in the literature. The obtained models showed small differences in the predictions and target values with experimental verification indicated that the estimation of the mechanical properties could be achieved with good accuracy.Keywords: multiscale hybrid reinforced cementitious composites, carbon nanotubes, carbon nanofibers, mechanical strength prediction
Procedia PDF Downloads 16145138 Analysis of the Secondary Stationary Flow Around an Oscillating Circular Cylinder
Authors: Artem Nuriev, Olga Zaitseva
Abstract:
This paper is devoted to the study of a viscous incompressible flow around a circular cylinder performing harmonic oscillations, especially the steady streaming phenomenon. The research methodology is based on the asymptotic explanation method combined with the computational bifurcation analysis. Present studies allow to identify several regimes of the secondary streaming with different flow structures. The results of the research are in good agreement with experimental and numerical simulation data.Keywords: oscillating cylinder, secondary streaming, flow regimes, asymptotic and bifurcation analysis
Procedia PDF Downloads 43545137 Using Analytical Hierarchy Process and TOPSIS Approaches in Designing a Finite Element Analysis Automation Program
Authors: Ming Wen, Nasim Nezamoddini
Abstract:
Sophisticated numerical simulations like finite element analysis (FEA) involve a complicated process from model setup to post-processing tasks that require replication of time-consuming steps. Utilizing FEA automation program simplifies the complexity of the involved steps while minimizing human errors in analysis set up, calculations, and results processing. One of the main challenges in designing FEA automation programs is to identify user requirements and link them to possible design alternatives. This paper presents a decision-making framework to design a Python based FEA automation program for modal analysis, frequency response analysis, and random vibration fatigue (RVF) analysis procedures. Analytical hierarchy process (AHP) and technique for order preference by similarity to ideal solution (TOPSIS) are applied to evaluate design alternatives considering the feedback received from experts and program users.Keywords: finite element analysis, FEA, random vibration fatigue, process automation, analytical hierarchy process, AHP, TOPSIS, multiple-criteria decision-making, MCDM
Procedia PDF Downloads 11245136 Planning for Sustainable Tourism in Chabahar Coastal Zone Using Swot Analysis
Authors: R. Karami, A. Gharaei
Abstract:
The aim of this study was to investigate ecotourism status in Chabahar coastal zone using swot analysis and strategic planning. Firstly, the current status of region was studied by literature review, field survey and statistical analysis. Then strengths and weaknesses (internal factors) were identified as well as opportunities and threats (external factors) using Delphi Method. Based on the obtained results, the total score of 2.46 in IFE matrix and 2.33 in the EFE matrix represents poor condition related to the internal and external factors respectively. This condition means both external and internal factors have not been utilized properly and the zone needs defensive plan; thus appropriate planning and organizational management practices are required to deal with these factors. Furthermore strategic goals, objectives and action plans in short, medium and long term schedule were formulated in attention to swot analysis.Keywords: tourism, SWOT analysis, strategic planning, Chabahar
Procedia PDF Downloads 51345135 The Trajectory of the Ball in Football Game
Authors: Mahdi Motahari, Mojtaba Farzaneh, Ebrahim Sepidbar
Abstract:
Tracking of moving and flying targets is one of the most important issues in image processing topic. Estimating of trajectory of desired object in short-term and long-term scale is more important than tracking of moving and flying targets. In this paper, a new way of identifying and estimating of future trajectory of a moving ball in long-term scale is estimated by using synthesis and interaction of image processing algorithms including noise removal and image segmentation, Kalman filter algorithm in order to estimating of trajectory of ball in football game in short-term scale and intelligent adaptive neuro-fuzzy algorithm based on time series of traverse distance. The proposed system attain more than 96% identify accuracy by using aforesaid methods and relaying on aforesaid algorithms and data base video in format of synthesis and interaction. Although the present method has high precision, it is time consuming. By comparing this method with other methods we realize the accuracy and efficiency of that.Keywords: tracking, signal processing, moving targets and flying, artificial intelligent systems, estimating of trajectory, Kalman filter
Procedia PDF Downloads 45945134 Web Page Design Optimisation Based on Segment Analytics
Authors: Varsha V. Rohini, P. R. Shreya, B. Renukadevi
Abstract:
In the web analytics the information delivery and the web usage is optimized and the analysis of data is done. The analytics is the measurement, collection and analysis of webpage data. Page statistics and user metrics are the important factor in most of the web analytics tool. This is the limitation of the existing tools. It does not provide design inputs for the optimization of information. This paper aims at providing an extension for the scope of web analytics to provide analysis and statistics of each segment of a webpage. The number of click count is calculated and the concentration of links in a web page is obtained. Its user metrics are used to help in proper design of the displayed content in a webpage by Vision Based Page Segmentation (VIPS) algorithm. When the algorithm is applied on the web page it divides the entire web page into the visual block tree. The visual block tree generated will further divide the web page into visual blocks or segments which help us to understand the usage of each segment in a page and its content. The dynamic web pages and deep web pages are used to extend the scope of web page segment analytics. Space optimization concept is used with the help of the output obtained from the Vision Based Page Segmentation (VIPS) algorithm. This technique provides us the visibility of the user interaction with the WebPages and helps us to place the important links in the appropriate segments of the webpage and effectively manage space in a page and the concentration of links.Keywords: analytics, design optimization, visual block trees, vision based technology
Procedia PDF Downloads 26645133 Design Improvement of Dental Implant-Based on Bone Remodelling
Authors: Solehuddin Shuib, Koay Boon Aik, Zainul Ahmad Rajion
Abstract:
There are many types of mechanical failure on the dental implant. In this project, the failure that needs to take into consideration is the bone resorption on the dental implant. Human bone has its ability to remodel after the implantation. As the dental implant is installed into the bone, the bone will detect and change the bone structure to achieve new biomechanical environment. This phenomenon is known as bone remodeling. The objective of the project is to improve the performance of dental implant by using different types of design. These designs are used to analyze and predict the failure of the dental implant by using finite element analysis (FEA) namely ANSYS. The bone is assumed to be fully attached to the implant or cement. Hence, results are then compared with other researchers. The results were presented in the form of Von Mises stress, normal stress, shear stress analysis, and displacement. The selected design will be analyzed further based on a theoretical calculation of bone remodeling on the dental implant. The results have shown that the design constructed passed the failure analysis. Therefore, the selected design is proven to have a stable performance at the recovery stage.Keywords: dental implant, FEA, bone remodeling, design
Procedia PDF Downloads 50145132 Vehicle Gearbox Fault Diagnosis Based on Cepstrum Analysis
Authors: Mohamed El Morsy, Gabriela Achtenová
Abstract:
Research on damage of gears and gear pairs using vibration signals remains very attractive, because vibration signals from a gear pair are complex in nature and not easy to interpret. Predicting gear pair defects by analyzing changes in vibration signal of gears pairs in operation is a very reliable method. Therefore, a suitable vibration signal processing technique is necessary to extract defect information generally obscured by the noise from dynamic factors of other gear pairs. This article presents the value of cepstrum analysis in vehicle gearbox fault diagnosis. Cepstrum represents the overall power content of a whole family of harmonics and sidebands when more than one family of sidebands is present at the same time. The concept for the measurement and analysis involved in using the technique are briefly outlined. Cepstrum analysis is used for detection of an artificial pitting defect in a vehicle gearbox loaded with different speeds and torques. The test stand is equipped with three dynamometers; the input dynamometer serves as the internal combustion engine, the output dynamometers introduce the load on the flanges of the output joint shafts. The pitting defect is manufactured on the tooth side of a gear of the fifth speed on the secondary shaft. Also, a method for fault diagnosis of gear faults is presented based on order cepstrum. The procedure is illustrated with the experimental vibration data of the vehicle gearbox. The results show the effectiveness of cepstrum analysis in detection and diagnosis of the gear condition.Keywords: cepstrum analysis, fault diagnosis, gearbox, vibration signals
Procedia PDF Downloads 37945131 Exponential Spline Solution for Singularly Perturbed Boundary Value Problems with an Uncertain-But-Bounded Parameter
Authors: Waheed Zahra, Mohamed El-Beltagy, Ashraf El Mhlawy, Reda Elkhadrawy
Abstract:
In this paper, we consider singular perturbation reaction-diffusion boundary value problems, which contain a small uncertain perturbation parameter. To solve these problems, we propose a numerical method which is based on an exponential spline and Shishkin mesh discretization. While interval analysis principle is used to deal with the uncertain parameter, sensitivity analysis has been conducted using different methods. Numerical results are provided to show the applicability and efficiency of our method, which is ε-uniform convergence of almost second order.Keywords: singular perturbation problem, shishkin mesh, two small parameters, exponential spline, interval analysis, sensitivity analysis
Procedia PDF Downloads 27445130 Case Analysis of Bamboo Based Social Enterprises in India-Improving Profitability and Sustainability
Authors: Priyal Motwani
Abstract:
The current market for bamboo products in India is about Rs. 21000 crores and is highly unorganised and fragmented. In this study, we have closely analysed the structure and functions of a major bamboo craft based organisation in Kerela, India and elaborated about its value chain, product mix, pricing strategy and supply chain, collaborations and competitive landscape. We have identified six major bottlenecks that are prevalent in such organisations, based on the Indian context, in relation to their product mix, asset management, and supply chain- corresponding waste management and retail network. The study has identified that the target customers for the bamboo based products and alternative revenue streams (eco-tourism, microenterprises, training), by carrying out secondary and primary research (5000 sample space), that can boost the existing revenue by 150%. We have then recommended an optimum product mix-covering premium, medium and low valued processing, for medium sized bamboo based organisations, in accordance with their capacity to maximize their revenue potential. After studying such organisations and their counter parts, the study has established an optimum retail network, considering B2B, B2C physical and online retail, to maximize their sales to their target groups. On the basis of the results obtained from the analysis of the future and present trends, our study gives recommendations to improve the revenue potential of bamboo based organisation in India and promote sustainability.Keywords: bamboo, bottlenecks, optimization, product mix, retail network, value chain
Procedia PDF Downloads 21745129 Comparative Analysis of Dissimilarity Detection between Binary Images Based on Equivalency and Non-Equivalency of Image Inversion
Authors: Adnan A. Y. Mustafa
Abstract:
Image matching is a fundamental problem that arises frequently in many aspects of robot and computer vision. It can become a time-consuming process when matching images to a database consisting of hundreds of images, especially if the images are big. One approach to reducing the time complexity of the matching process is to reduce the search space in a pre-matching stage, by simply removing dissimilar images quickly. The Probabilistic Matching Model for Binary Images (PMMBI) showed that dissimilarity detection between binary images can be accomplished quickly by random pixel mapping and is size invariant. The model is based on the gamma binary similarity distance that recognizes an image and its inverse as containing the same scene and hence considers them to be the same image. However, in many applications, an image and its inverse are not treated as being the same but rather dissimilar. In this paper, we present a comparative analysis of dissimilarity detection between PMMBI based on the gamma binary similarity distance and a modified PMMBI model based on a similarity distance that does distinguish between an image and its inverse as being dissimilar.Keywords: binary image, dissimilarity detection, probabilistic matching model for binary images, image mapping
Procedia PDF Downloads 15345128 Automatic Reporting System for Transcriptome Indel Identification and Annotation Based on Snapshot of Next-Generation Sequencing Reads Alignment
Authors: Shuo Mu, Guangzhi Jiang, Jinsa Chen
Abstract:
The analysis of Indel for RNA sequencing of clinical samples is easily affected by sequencing experiment errors and software selection. In order to improve the efficiency and accuracy of analysis, we developed an automatic reporting system for Indel recognition and annotation based on image snapshot of transcriptome reads alignment. This system includes sequence local-assembly and realignment, target point snapshot, and image-based recognition processes. We integrated high-confidence Indel dataset from several known databases as a training set to improve the accuracy of image processing and added a bioinformatical processing module to annotate and filter Indel artifacts. Subsequently, the system will automatically generate data, including data quality levels and images results report. Sanger sequencing verification of the reference Indel mutation of cell line NA12878 showed that the process can achieve 83% sensitivity and 96% specificity. Analysis of the collected clinical samples showed that the interpretation accuracy of the process was equivalent to that of manual inspection, and the processing efficiency showed a significant improvement. This work shows the feasibility of accurate Indel analysis of clinical next-generation sequencing (NGS) transcriptome. This result may be useful for RNA study for clinical samples with microsatellite instability in immunotherapy in the future.Keywords: automatic reporting, indel, next-generation sequencing, NGS, transcriptome
Procedia PDF Downloads 19145127 Exploiting Fast Independent Component Analysis Based Algorithm for Equalization of Impaired Baseband Received Signal
Authors: Muhammad Umair, Syed Qasim Gilani
Abstract:
A technique using Independent Component Analysis (ICA) for blind receiver signal processing is investigated. The problem of the receiver signal processing is viewed as of signal equalization and implementation imperfections compensation. Based on this, a model similar to a general ICA problem is developed for the received signal. Then, the use of ICA technique for blind signal equalization in the time domain is presented. The equalization is regarded as a signal separation problem, since the desired signal is separated from interference terms. This problem is addressed in the paper by over-sampling of the received signal. By using ICA for equalization, besides channel equalization, other transmission imperfections such as Direct current (DC) bias offset, carrier phase and In phase Quadrature phase imbalance will also be corrected. Simulation results for a system using 16-Quadraure Amplitude Modulation(QAM) are presented to show the performance of the proposed scheme.Keywords: blind equalization, blind signal separation, equalization, independent component analysis, transmission impairments, QAM receiver
Procedia PDF Downloads 21445126 We Are the 99 percent – the Occupy-Movement in Social Media
Authors: Wolfram Karg
Abstract:
The Occupy-Movement came into in 2011 existence in the US as a reaction to one of the worst economic crisis since World War II. With cuts in benefits and social services, with people being evicted from their homes on the one hand and high bonuses granted to their managers of the very same companies, a strong feeling of injustice besieged people in the US and caused them to voice their anger peacefully in social media and on the streets. Due to the world-wide-web, users all around the world read about this movement and recognized the same injustice in their own countries, making Occupy a global movement. The vast array of topics covered by Occupy offers a unique chance to carry out a corpus-based discourse analysis based on the DIMEAN-Model. The focus on this paper is limited to two aspects of DIMEAN: intertextual references and the use of connectors in texts. Because the discourse is to a large extent carried out via posts in blogs, online-articles and comments, the paper also analyses, in how far modern (i.e. computer-based media) there is a correlation between the use of connectors in different communicative types used by the Occupy-Movement.Keywords: discourse, new media, occupy, corpus analysis
Procedia PDF Downloads 49345125 Nature of Cities: Ontological Dimension of the Urban
Authors: Ana Cristina García-Luna Romero
Abstract:
This document seeks to reflect on the urban project from its conceptual identity root. In the first instance, a proposal is made on how the city project is sustained from the conceptual root, from the logos: it opens a way to assimilate the imagination; what we imagine becomes a reality. In this way, firstly, the need to use language as a vehicle for transmitting the stories that sustain us as humanity can be deemed as an important social factor that enables us to social behavior. Secondly, the need to attend to the written language as a mechanism of power, as a means to consolidate a dominant ideology or a political position, is raised; as it served to carry out the modernization project, it is therefore addressed differences between the real and the literate city. Thus, the consolidated urban-architectural project is based on logos, the project, and planning. Considering the importance of materiality and its relation to subjective well-being contextualized from a socio-urban approach, we question ourselves into how we can look at something that is doubtful. From a philosophy perspective, the truth is considered to be nothing more than a matter of correspondence between the observer and the observed. To understand beyond the relative of the gaze, it is necessary to expose different perspectives since it depends on the understanding of what is observed and how it is critically analyzed. Therefore, the analysis of materiality, as a political field, takes a proposal based on this research in the principles in transgenesis: principle of communication, representativeness, security, health, malleability, availability of potentiality or development, conservation, sustainability, economy, harmony, stability, accessibility, justice, legibility, significance, consistency, joint responsibility, connectivity, beauty, among others. The (urban) human being acts because he wants to live in a certain way: in a community, in a fair way, with opportunity for development, with the possibility of managing the environment according to their needs, etc. In order to comply with this principle, it is necessary to design strategies from the principles in transgenesis, which must be named, defined, understood, and socialized by the urban being, the companies, and from themselves. In this way, the technical status of the city in the neoliberal present determines extraordinary conditions for reflecting on an almost emergency scenario created by the impact of cities that, far from being limited to resilient proposals, must aim at the reflection of the urban process that the present social model has generated. Therefore, can we rethink the paradigm of the perception of life quality in the current neoliberal model in the production of the character of public space related to the practices of being urban. What we are trying to do within this document is to build a framework to study under what logic the practices of the social system that make sense of the public space are developed, what the implications of the phenomena of the inscription of action and materialization (and its results over political action between the social and the technical system) are and finally, how we can improve the quality of life of individuals from the urban space.Keywords: cities, nature, society, urban quality of life
Procedia PDF Downloads 12445124 Estimation of Energy Losses of Photovoltaic Systems in France Using Real Monitoring Data
Authors: Mohamed Amhal, Jose Sayritupac
Abstract:
Photovoltaic (PV) systems have risen as one of the modern renewable energy sources that are used in wide ranges to produce electricity and deliver it to the electrical grid. In parallel, monitoring systems have been deployed as a key element to track the energy production and to forecast the total production for the next days. The reliability of the PV energy production has become a crucial point in the analysis of PV systems. A deeper understanding of each phenomenon that causes a gain or a loss of energy is needed to better design, operate and maintain the PV systems. This work analyzes the current losses distribution in PV systems starting from the available solar energy, going through the DC side and AC side, to the delivery point. Most of the phenomena linked to energy losses and gains are considered and modeled, based on real time monitoring data and datasheets of the PV system components. An analysis of the order of magnitude of each loss is compared to the current literature and commercial software. To date, the analysis of PV systems performance based on a breakdown structure of energy losses and gains is not covered enough in the literature, except in some software where the concept is very common. The cutting-edge of the current analysis is the implementation of software tools for energy losses estimation in PV systems based on several energy losses definitions and estimation technics. The developed tools have been validated and tested on some PV plants in France, which are operating for years. Among the major findings of the current study: First, PV plants in France show very low rates of soiling and aging. Second, the distribution of other losses is comparable to the literature. Third, all losses reported are correlated to operational and environmental conditions. For future work, an extended analysis on further PV plants in France and abroad will be performed.Keywords: energy gains, energy losses, losses distribution, monitoring, photovoltaic, photovoltaic systems
Procedia PDF Downloads 17645123 Video Based Ambient Smoke Detection By Detecting Directional Contrast Decrease
Authors: Omair Ghori, Anton Stadler, Stefan Wilk, Wolfgang Effelsberg
Abstract:
Fire-related incidents account for extensive loss of life and material damage. Quick and reliable detection of occurring fires has high real world implications. Whereas a major research focus lies on the detection of outdoor fires, indoor camera-based fire detection is still an open issue. Cameras in combination with computer vision helps to detect flames and smoke more quickly than conventional fire detectors. In this work, we present a computer vision-based smoke detection algorithm based on contrast changes and a multi-step classification. This work accelerates computer vision-based fire detection considerably in comparison with classical indoor-fire detection.Keywords: contrast analysis, early fire detection, video smoke detection, video surveillance
Procedia PDF Downloads 44745122 Synthesis and Characterization of Mass Catalysts Based on Cobalt and Molybdenum
Authors: Nassira Ouslimani
Abstract:
The electronic structure of transition metals gives them many catalytic possibilities in many types of reactions, particularly cobalt and molybdenum. It is in this context that this study is part of the synthesis and characterization of mass catalysts based on cobalt and molybdenum Co1₋xMoO4 (X=0 and X=0.5 and X=1). The two catalysts were prepared by Co-precipitation using ammonia as a precipitating agent and one by precipitation. The samples obtained were analyzed by numerous physic-chemical analysis techniques: ATG-ATD-DSC, DRX-HT, SEM-EDX, and the elemental composition of the catalysts was verified by SAA as well as the FTIR. The ATG-DSC shows a mass loss for all the catalysts of approximately 8%, corresponding to the loss of water and the decomposition of nitrates. The DRX-HT analysis allows the detection of the two CoMoO4 phases with diffraction peaks which increase with the increase in temperature. The results of the FTIR analysis made it possible to highlight the vibration modes of the bonds of the structure of the prepared catalysts. The SEM images of the solids show very different textures with almost homogeneous surfaces with a more regular particle size distribution and a more defined grain shape. The EDX analysis showed the presence of the elements Co, Mo, and O in proportions very close to the nominal proportions. Finally, the actual composition, evaluated by SAA, is close to the theoretical composition fixed during the preparation. This testifies to the good conditions for the preparation of the catalysts by the co-precipitation method.Keywords: catalytic, molybdenum, coprecipitation, cobalt, ammonia
Procedia PDF Downloads 9045121 The Effect of Artificial Intelligence on Construction Development
Authors: Shady Gamal Aziz Shehata
Abstract:
Difficulty in defining construction quality arises due to perception based on the nature and requirements of the market, the different partners themselves and the results they want. Quantitative research was used in this constructivist research. A case-based study was conducted to assess the structures of positive attitudes and expectations in the context of quality improvement. A survey based on expert opinions was analyzed among construction organizations/companies operating in the construction industry in Pakistan. The financial strength, management structure and construction experience of the construction companies formed the basis of their selection. A good concept is visible at the project level and is seen as the most valuable part of the construction project. Each quality improvement technique was expected to increase the user's profits by improving the efficiency of the construction project. The Survey is useful for construction professionals to evaluate current construction concepts and expectations for the application of quality improvement techniques in construction projects.Keywords: correlation analysis, lean construction tools, lean construction, logistic regression analysis, risk management, safety construction quality, expectation, improvement, perception
Procedia PDF Downloads 5945120 A Review of Spatial Analysis as a Geographic Information Management Tool
Authors: Chidiebere C. Agoha, Armstong C. Awuzie, Chukwuebuka N. Onwubuariri, Joy O. Njoku
Abstract:
Spatial analysis is a field of study that utilizes geographic or spatial information to understand and analyze patterns, relationships, and trends in data. It is characterized by the use of geographic or spatial information, which allows for the analysis of data in the context of its location and surroundings. It is different from non-spatial or aspatial techniques, which do not consider the geographic context and may not provide as complete of an understanding of the data. Spatial analysis is applied in a variety of fields, which includes urban planning, environmental science, geosciences, epidemiology, marketing, to gain insights and make decisions about complex spatial problems. This review paper explores definitions of spatial analysis from various sources, including examples of its application and different analysis techniques such as Buffer analysis, interpolation, and Kernel density analysis (multi-distance spatial cluster analysis). It also contrasts spatial analysis with non-spatial analysis.Keywords: aspatial technique, buffer analysis, epidemiology, interpolation
Procedia PDF Downloads 31845119 Analyzing the Effectiveness of Communication Practices and Processes within Project-Based Firms
Authors: Paul Saah, Charles Mbohwa, Nelson Sizwe Madonsela
Abstract:
The capacity to deliver projects on schedule, within budget, and to the pleasure of the client depends on effective communication, which is the lifeblood of project-based businesses. In order to pinpoint areas for development and shed light on the crucial role that communication plays in project success, the aim of this study is to evaluate the efficacy of communication practises and processes inside project-based organisations. In order to analyse concepts and get a greater grasp of their theoretical basis, this study's methodology combines a careful review of the relevant literature with a conceptual analysis of the subject. Data from a varied sample of project-based businesses spanning all industries and sizes were collected via document analysis. The relationship between communication practises, and processes were investigated in connection to key performance measures such as project outcomes, client satisfaction, and team dynamics. According to the study's findings, project-based businesses that adopt effective communication practises, and procedures experience a reduction in unfavourable experiences, stronger integration, and coordination, clarity of purpose, and practises that can hasten problem resolution. However, failing to adopt effective communication practises and procedures in project-based company result in counter issues, including project derailment from the schedule, failure to meet goals, inefficient use of existing resources, and failure to meet organisational goals. Therefore, optimising their communication practises, and procedures are crucial for sustainable growth and competitive advantage as project-based enterprises continue to play a crucial part in today's dynamic business scene.Keywords: effective communication, project-based firms, communication practices, project success, communication strategies
Procedia PDF Downloads 7045118 Image-Based (RBG) Technique for Estimating Phosphorus Levels of Different Crops
Authors: M. M. Ali, Ahmed Al- Ani, Derek Eamus, Daniel K. Y. Tan
Abstract:
In this glasshouse study, we developed the new image-based non-destructive technique for detecting leaf P status of different crops such as cotton, tomato and lettuce. Plants were allowed to grow on nutrient media containing different P concentrations, i.e. 0%, 50% and 100% of recommended P concentration (P0 = no P, L; P1 = 2.5 mL 10 L-1 of P and P2 = 5 mL 10 L-1 of P as NaH2PO4). After 10 weeks of growth, plants were harvested and data on leaf P contents were collected using the standard destructive laboratory method and at the same time leaf images were collected by a handheld crop image sensor. We calculated leaf area, leaf perimeter and RGB (red, green and blue) values of these images. This data was further used in the linear discriminant analysis (LDA) to estimate leaf P contents, which successfully classified these plants on the basis of leaf P contents. The data indicated that P deficiency in crop plants can be predicted using the image and morphological data. Our proposed non-destructive imaging method is precise in estimating P requirements of different crop species.Keywords: image-based techniques, leaf area, leaf P contents, linear discriminant analysis
Procedia PDF Downloads 38045117 Analysis of the Decoupling Relationship between Urban Green Development and the Level of Regional Integration Based on the Tapio Model
Authors: Ruoyu Mao
Abstract:
Exploring the relationship between urban green development and regional integration level is of great significance for realising regional high quality and sustainable development. Based on the Tapio decoupling model and the theoretical framework of urban green development and regional integration, this paper builds an analysis system, makes a quantitative analysis of urban green development and regional integration level in a certain period, and discusses the relationship between the two. It also takes China's Yangtze River Delta urban agglomeration as an example to study the degree of decoupling, the type of decoupling, and the trend of the evolution of the spatio-temporal pattern of decoupling between the level of urban green development and the level of regional integration in the period of 2014-2021, with the aim of providing a useful reference for the future development of the region.Keywords: regional integration, urban green development, Tapio decoupling model, Yangtze River Delta urban agglomeration
Procedia PDF Downloads 43