Search results for: neural net works
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3471

Search results for: neural net works

711 Carbon Footprint of Road Project for Sustainable Development: Lessons Learnt from Traffic Management of a Developing Urban Centre

Authors: Sajjad Shukur Ullah, Syed Shujaa Safdar Gardezi

Abstract:

Road infrastructure plays a vital role in the economic activities of any economy. Besides derived benefits from these facilities, the utilization of extensive energy resources, fuels, and materials results in a negative impact on the environment in terms of carbon footprint; carbon footprint is the overall amount of greenhouse gas (GHG) generated from any action. However, this aspect of environmental impact from road structure is not seriously considered during such developments, thus undermining a critical factor of sustainable development, which usually remains unaddressed, especially in developing countries. The current work investigates the carbon footprint impact of a small road project (0.8 km, dual carriageway) initiated for traffic management in an urban centre. Life cycle assessment (LCA) with boundary conditions of cradle to the site has been adopted. The only construction phase of the life cycle has been assessed at this stage. An impact of 10 ktons-CO2 (6260 ton-CO2/km) has been assessed. The rigid pavement dominated the contributions as compared to a flexible component. Among the structural elements, the underpass works shared the major portion. Among the materials, the concrete and steel utilized for various structural elements resulted in more than 90% of the impact. The earth-moving equipment was dominant in operational carbon. The results have highlighted that road infrastructure projects pose serious threats to the environment during their construction and which need to be considered during the approval stages. This work provides a guideline for supporting sustainable development that could only be ensured when such endeavours are properly assessed by industry professionals and decide various alternative environmental conscious solutions for the future.

Keywords: construction waste management, kiloton, life cycle assessment, rigid pavement

Procedia PDF Downloads 83
710 The Effects of Damping Devices on Displacements, Velocities and Accelerations of Structures

Authors: Radhwane Boudjelthia

Abstract:

The most recent earthquakes occurred in the world have killed thousands of people and severe damage. For all the actors involved in the building process, the earthquake is the litmus test for construction. The goal we set ourselves is to contribute to the implementation of a thoughtful approach to the seismic protection of structures. For many engineers, the most conventional approach to protection works (buildings and bridges) the effects of earthquakes is to increase rigidity. This approach is not always effective, especially when there is a context that favors the phenomenon of resonance and amplification of seismic forces. Therefore, the field of earthquake engineering has made significant inroads, among others catalyzed by the development of computational techniques in computer form and the use of powerful test facilities. This has led to the emergence of several innovative technologies, such as the introduction of special devices insulation between infrastructure and superstructure. This approach, commonly known as "seismic isolation," to absorb the significant efforts without the structure is damaged and thus ensuring the protection of lives and property. In addition, the restraints to the construction by the ground shaking are located mainly at the supports. With these moves, the natural period of construction is increasing, and seismic loads are reduced. Thus, there is an attenuation of the seismic movement. Likewise, the insulation of the base mechanism may be used in combination with earthquake dampers in order to control the deformation of the insulation system and the absolute displacement of the superstructure located above the isolation interface. On the other hand, only can use these earthquake dampers to reduce the oscillation amplitudes and thus reduce seismic loads. The use of damping devices represents an effective solution for the rehabilitation of existing structures. Given all these acceleration reducing means considered passive, much research has been conducted for several years to develop an active control system of the response of buildings to earthquakes.

Keywords: earthquake, building, seismic forces, displacement, resonance, response.

Procedia PDF Downloads 52
709 Effect of Coaching Related Incompetency to Stand Trial on Symptom Validity Test: Robustness, Sensitivity, and Specificity

Authors: Natthawut Arin

Abstract:

In forensic contexts, competency to stand trial assessments are the most common referrals. The defendants may attempt to endorse psychopathology symptoms and feign incompetent. Coaching, which can be teaching them test-taking strategies to avoid detection of psychopathological symptoms feigning. Recently, the Symptom Validity Testings (SVTs) were created to detect feigning. Moreover, the works of the literature showed that the effects of coaching on SVTs may be more robust to the effects of coaching. Thai Symptom Validity Test (SVT-Th) was designed as SVTs which demonstrated adequate psychometric properties and ability to classify between feigners and honest responders. Thus, the current study to examine the utility as the robustness of SVT-Th in the detection of feigned psychopathology. Participants consisted of 120 were recruited from undergraduate courses in psychology, randomly assigned to one of three groups. The SVT-Th was administered to those three scenario-experimental groups: (a) Uncoached group were asked to respond honestly (n=40), (b) Symptom-coached without warning group were asked to feign psychiatric symptoms to gain incompetency to stand trial (n=40), while (c) Test-coached with warning group were asked to feign psychiatric symptoms to avoid test detection but being incompetency to stand trial (n=40). Group differences were analyzed using one-way ANOVAs. The result revealed an uncoached group (M = 4.23, SD.= 5.20) had significantly lower SVT-Th mean scores than those both coached groups (M =185.00, SD.= 72.88 and M = 132.10, SD.= 54.06, respectively). Classification rates were calculated to determine the classification accuracy. Result indicated that SVT-Th had overall classification accuracy rates of 96.67% with acceptable of 95% sensitivity and 100% specificity rates. Overall, the results of the present study indicate that the SVT-Th yielded high adequate indices of accuracy and these findings suggest that the SVT-Th is robustness against coaching.

Keywords: incompetency to stand trial, coaching, robustness, classification accuracy

Procedia PDF Downloads 121
708 The Nuclear Energy Museum in Brazil: Creative Solutions to Transform Science Education into Meaningful Learning

Authors: Denise Levy, Helen J. Khoury

Abstract:

Nuclear technology is a controversial issue among a great share of the Brazilian population. Misinformation and common wrong beliefs confuse public’s perceptions and the scientific community is expected to offer a wider perspective on the benefits and risks resulting from ionizing radiation in everyday life. Attentive to the need of new approaches between science and society, the Nuclear Energy Museum, in northeast Brazil, is an initiative created to communicate the growing impact of the beneficial applications of nuclear technology in medicine, industry, agriculture and electric power generation. Providing accessible scientific information, the museum offers a rich learning environment, making use of different educational strategies, such as films, interactive panels and multimedia learning tools, which not only increase the enjoyment of visitors, but also maximize their learning potential. Developed according to modern active learning instructional strategies, multimedia materials are designed to present the increasingly role of nuclear science in modern life, transforming science education into a meaningful learning experience. In year 2016, nine different interactive computer-based activities were developed, presenting curiosities about ionizing radiation in different landmarks around the world, such as radiocarbon dating works in Egypt, nuclear power generation in France and X-radiography of famous paintings in Italy. Feedback surveys have reported a high level of visitors’ satisfaction, proving the high quality experience in learning nuclear science at the museum. The Nuclear Energy Museum is the first and, up to the present time, the only permanent museum in Brazil devoted entirely to nuclear science.

Keywords: nuclear technology, multimedia learning tools, science museum, society and education

Procedia PDF Downloads 301
707 Recent Progress in the Uncooled Mid-Infrared Lead Selenide Polycrystalline Photodetector

Authors: Hao Yang, Lei Chen, Ting Mei, Jianbang Zheng

Abstract:

Currently, the uncooled PbSe photodetectors in the mid-infrared range (2-5μm) with sensitization technology extract more photoelectric response than traditional ones, and enable the room temperature (300K) photo-detection with high detectivity, which have attracted wide attentions in many fields. This technology generally contains the film fabrication with vapor phase deposition (VPD) and a sensitizing process with doping of oxygen and iodine. Many works presented in the recent years almost provide and high temperature activation method with oxygen/iodine vapor diffusion, which reveals that oxygen or iodine plays an important role in the sensitization of PbSe material. In this paper, we provide our latest experimental results and discussions in the stoichiometry of oxygen and iodine and its influence on the polycrystalline structure and photo-response. The experimental results revealed that crystal orientation was transformed from (200) to (420) by sensitization, and the responsivity of 5.42 A/W was gained by the optimal stoichiometry of oxygen and iodine with molecular density of I2 of ~1.51×1012 mm-3 and oxygen pressure of ~1Mpa. We verified that I2 plays a role in transporting oxygen into the lattice of crystal, which is actually not its major role. It is revealed that samples sensitized with iodine transform atomic proportion of Pb from 34.5% to 25.0% compared with samples without iodine from XPS data, which result in the proportion of about 1:1 between Pb and Se atoms by sublimation of PbI2 during sensitization process, and Pb/Se atomic proportion is controlled by I/O atomic proportion in the polycrystalline grains, which is very an important factor for improving responsivity of uncooled PbSe photodetector. Moreover, a novel sensitization and dopant activation method is proposed using oxygen ion implantation with low ion energy of < 500eV and beam current of ~120μA/cm2. These results may be helpful to understanding the sensitization mechanism of polycrystalline lead salt materials.

Keywords: polycrystalline PbSe, sensitization, transport, stoichiometry

Procedia PDF Downloads 331
706 Steady and Spatio-Temporal Monitoring of Water Quality Feeding Area Southwest of Great Casablanca (Morocco)

Authors: Hicham Maklache, Rajae Delhi, Fatiha Benzha, Mohamed Tahiri

Abstract:

In Morocco, where semi-arid climate is dominant, the supply of industrial and drink water is provided primarily by surface water. Morocco has currently 118 multi-purpose dams. If the construction of these works was a necessity to ensure in all seasons, the water essential to our country, it is impartial to control and protect the quality of running water. -Most dam reservoir used are threatened by eutrophication due to increased terrigenous and anthropogenic pollutants, coming from an over-fertilization of water by phosphorus and nitrogen nutrients and accelerated by uncontrolled development of microalgae aging. It should also be noted that the daily practices of citizens with respect to the resource, an essential component involved in almost all human activities (agriculture, agro-industries, hydropower, ...), has contributed significantly to the deterioration of water quality despite its treatment in several plants. Therefore, the treated water, provides a legacy of bad tastes and odors unacceptable to the consumer. -The present work exhibits results of water quality watershed Oum Erbia used to supply drinking water to the whole terraced area connecting the city of Khenifra to the city of Azemmour. The area south west of Great Casablanca (metropolis of the kingdom with about 4 million inhabitants) supplied 50% of its water needs by sourcing Dam Sidi Said Maachou located, last anchor point of the watershed before the spill in the Atlantic Ocean. The results were performed in a spatio-temporal scale and helped to establish a history of monitoring water quality during the 2009-2011 cycles, the study also presents the development of quality according to the seasonal rhythmicity and rainfall. It gives also an overview on the concept of watershed stewardship.

Keywords: crude surface water quality, Oum Er Rbia hydraulic basin, spatio-temporal monitoring, Great Casablanca drink water quality, Morocco

Procedia PDF Downloads 426
705 Development of a Fuzzy Logic Based Model for Monitoring Child Pornography

Authors: Mariam Ismail, Kazeem Rufai, Jeremiah Balogun

Abstract:

A study was conducted to apply fuzzy logic to the development of a monitoring model for child pornography based on associated risk factors, which can be used by forensic experts or integrated into forensic systems for the early detection of child pornographic activities. A number of methods were adopted in the study, which includes an extensive review of related works was done in order to identify the factors that are associated with child pornography following which they were validated by an expert sex psychologist and guidance counselor, and relevant data was collected. Fuzzy membership functions were used to fuzzify the associated variables identified alongside the risk of the occurrence of child pornography based on the inference rules that were provided by the experts consulted, and the fuzzy logic expert system was simulated using the Fuzzy Logic Toolbox available in the MATLAB Software Release 2016. The results of the study showed that there were 4 categories of risk factors required for assessing the risk of a suspect committing child pornography offenses. The results of the study showed that 2 and 3 triangular membership functions were used to formulate the risk factors based on the 2 and 3 number of labels assigned, respectively. The results of the study showed that 5 fuzzy logic models were formulated such that the first 4 was used to assess the impact of each category on child pornography while the last one takes the 4 outputs from the 4 fuzzy logic models as inputs required for assessing the risk of child pornography. The following conclusion was made; there were factors that were related to personal traits, social traits, history of child pornography crimes, and self-regulatory deficiency traits by the suspects required for the assessment of the risk of child pornography crimes committed by a suspect. Using the values of the identified risk factors selected for this study, the risk of child pornography can be easily assessed from their values in order to determine the likelihood of a suspect perpetuating the crime.

Keywords: fuzzy, membership functions, pornography, risk factors

Procedia PDF Downloads 108
704 Modeling and Temperature Control of Water-cooled PEMFC System Using Intelligent Algorithm

Authors: Chen Jun-Hong, He Pu, Tao Wen-Quan

Abstract:

Proton exchange membrane fuel cell (PEMFC) is the most promising future energy source owing to its low operating temperature, high energy efficiency, high power density, and environmental friendliness. In this paper, a comprehensive PEMFC system control-oriented model is developed in the Matlab/Simulink environment, which includes the hydrogen supply subsystem, air supply subsystem, and thermal management subsystem. Besides, Improved Artificial Bee Colony (IABC) is used in the parameter identification of PEMFC semi-empirical equations, making the maximum relative error between simulation data and the experimental data less than 0.4%. Operation temperature is essential for PEMFC, both high and low temperatures are disadvantageous. In the thermal management subsystem, water pump and fan are both controlled with the PID controller to maintain the appreciate operation temperature of PEMFC for the requirements of safe and efficient operation. To improve the control effect further, fuzzy control is introduced to optimize the PID controller of the pump, and the Radial Basis Function (RBF) neural network is introduced to optimize the PID controller of the fan. The results demonstrate that Fuzzy-PID and RBF-PID can achieve a better control effect with 22.66% decrease in Integral Absolute Error Criterion (IAE) of T_st (Temperature of PEMFC) and 77.56% decrease in IAE of T_in (Temperature of inlet cooling water) compared with traditional PID. In the end, a novel thermal management structure is proposed, which uses the cooling air passing through the main radiator to continue cooling the secondary radiator. In this thermal management structure, the parasitic power dissipation can be reduced by 69.94%, and the control effect can be improved with a 52.88% decrease in IAE of T_in under the same controller.

Keywords: PEMFC system, parameter identification, temperature control, Fuzzy-PID, RBF-PID, parasitic power

Procedia PDF Downloads 64
703 Y-Y’ Calculus in Physical Sciences and Engineering with Particular Reference to Fundamentals of Soil Consolidation

Authors: Sudhir Kumar Tewatia, Kanishck Tewatia, Anttriksh Tewatia

Abstract:

Advancements in soil consolidation are discussed, and further improvements are proposed with particular reference to Tewatia’s Y-Y’ Approach, which is called the Settlement versus Rate of Settlement Approach in consolidation. A branch of calculus named Y-Y' (or y versus dy/dx) is suggested (as compared to the common X-Y', x versus dy/dx, dy/dx versus x or Newton-Leibniz branch) that solves some complicated/unsolved theoretical and practical problems in physical sciences (Physics, Chemistry, Mathematics, Biology, and allied sciences) and engineering in an amazingly simple and short manner, particularly when independent variable X is unknown and X-Y' Approach can’t be used. Complicated theoretical and practical problems in 1D, 2D, 3D Primary and Secondary consolidations with non-uniform gradual loading and irregularly shaped clays are solved with elementary school level Y-Y' Approach, and it is interesting to note that in X-Y' Approach, equations become more difficult while we move from one to three dimensions, but in Y-Y' Approach even 2D/3D equations are very simple to derive, solve, and use; rather easier sometimes. This branch of calculus will have a far-reaching impact on understanding and solving the problems in different fields of physical sciences and engineering that were hitherto unsolved or difficult to be solved by normal calculus/numerical/computer methods. Some particular cases from soil consolidation that basically creeps and diffusion equations in isolation and in combination with each other are taken for comparison with heat transfer. The Y-Y’ Approach can similarly be applied in wave equations and other fields wherever normal calculus works or fails. Soil mechanics uses mathematical analogies from other fields of physical sciences and engineering to solve theoretical and practical problems; for example, consolidation theory is a replica of the heat equation from thermodynamics with the addition of the effective stress principle. An attempt is made to give them mathematical analogies.

Keywords: calculus, clay, consolidation, creep, diffusion, heat, settlement

Procedia PDF Downloads 73
702 Synthesis and Biological Activities of Novel -1,2,3-Triazoles Derivatives

Authors: Zahra Dehghani, Hoda Dehghani, Elham Zarenezhad

Abstract:

1,2,3-Triazole derivatives are important compounds in medicinal chemistry owing to their wide applications in drug discovery. They can readily associate with biologically targets through the hydrogen bonding and dipole interactions. The 1,2,3-triazole core is a key structural motif in many bioactive compounds, exhibiting a broad spectrum of biological activities, such as antiviral, anticancer, anti-HIV, antibiotic, antibacterial, and antimicrobial. Additionally, they have found significant industrial applications as dyes, agrochemicals, corrosion inhibitors, photo stabilizers, and photographic materials. we disclose the synthesis and characterization of 1-azido-3-(aryl-2-yloxy)propan-2-ol drivatives. The chemistry works well with various ß-azido alcohols involving aryloxy, alkoxy and alkyl residues, and also tolerates a wide spectrum of electron-donating and electron-withdrawing functional groups in both alkyne and azide molecules. Most of ß-azidoalcohols used in these experiments were pre-synthesized by the regioselective ring opening reaction of corresponded epoxides with sodium azide, whereas the majority of terminal alkynes were prepared via SN2-type reaction of propargyl bromide and corresponded nucleophiles. To evaluate the bioactivity of title compounds, the in vitro antifungal activity of all compound was investigated against several pathogenic fungi including Candida albicans, Candida krusei, Aspergillus niger, and Trichophyton rubrum , clotrimazole and fluconazole was used as standard antifungal drugs, also To understand the antibacterial activity of synthesized compounds, they were in vitro screened against E. coli and S. aureus as Gram-negative and Gram-positive bacteria, respectively. The in vitro tests have shown the promising antifungal but marginal antibacterial activity against tested fungi and bacteria.

Keywords: biological activities, antibacterial, antifungal, 1, 2, 3-Triazole

Procedia PDF Downloads 416
701 The Evaluation of the Performance of CaCO3/Polymer Nano-Composites for the Preservation of Historic Limestone Monuments

Authors: Mohammed Badereldien, Rezk Diab, Mohamoud Ali, Ayman Aboelkassem

Abstract:

The stone surfaces of historical architectural heritage in Egypt are under threat from of various environmental factors such as temperature fluctuation, humidity, pollution, and microbes. Due to these factors, the facades of buildings are deteriorating deformation and disfiguration of external decoration and the formation of black accretion also often from the stone works. The aim of this study is to evaluate the effectiveness of CaCO₃ nano-particles as consolidation and protection material for calcareous stone monuments. Selected tests were carried out in order to estimate the superficial consolidating and protective effect of the treatment. When applied the nanoparticles dispersed in the acrylic copolymer; poly ethylmethacrylate (EMA)/methylacrylate (MA) (70/30, respectively) (EMA)/methylacrylate (MA) (70/30, respectively). The synthesis process of CaCO₃ nanoparticles/polymer nano-composite was prepared using in situ emulsion polymerization system. The consolidation and protection were characterized by TEM, while the penetration depth, re-aggregating effects of the deposited phase, and the surface morphology before and after treatment were examined by SEM (Scanning Electron Microscopy). Improvement of the stones' mechanical properties was evaluated by compressive strength tests. Changes in water-interaction properties were evaluated by water absorption capillarity measurements, and colorimetric measurements were used to evaluate the optical appearance. Together the results appear to demonstrate that CaCO₃/polymer nanocomposite is an efficient material for the consolidation of limestone architecture and monuments. As compared with samples treated with pure acrylic copolymer without Calcium carbonate nanoparticles, for example, CaCO₃ nanoparticles are completely compatible, strengthening limestone against thermal aging and improving its mechanical properties.

Keywords: calcium carbonate nanoparticles, consolidation, nanocomposites, calcareous stone, colorimetric measurements, compressive strength

Procedia PDF Downloads 116
700 Sentiment Analysis of Fake Health News Using Naive Bayes Classification Models

Authors: Danielle Shackley, Yetunde Folajimi

Abstract:

As more people turn to the internet seeking health-related information, there is more risk of finding false, inaccurate, or dangerous information. Sentiment analysis is a natural language processing technique that assigns polarity scores to text, ranging from positive, neutral, and negative. In this research, we evaluate the weight of a sentiment analysis feature added to fake health news classification models. The dataset consists of existing reliably labeled health article headlines that were supplemented with health information collected about COVID-19 from social media sources. We started with data preprocessing and tested out various vectorization methods such as Count and TFIDF vectorization. We implemented 3 Naive Bayes classifier models, including Bernoulli, Multinomial, and Complement. To test the weight of the sentiment analysis feature on the dataset, we created benchmark Naive Bayes classification models without sentiment analysis, and those same models were reproduced, and the feature was added. We evaluated using the precision and accuracy scores. The Bernoulli initial model performed with 90% precision and 75.2% accuracy, while the model supplemented with sentiment labels performed with 90.4% precision and stayed constant at 75.2% accuracy. Our results show that the addition of sentiment analysis did not improve model precision by a wide margin; while there was no evidence of improvement in accuracy, we had a 1.9% improvement margin of the precision score with the Complement model. Future expansion of this work could include replicating the experiment process and substituting the Naive Bayes for a deep learning neural network model.

Keywords: sentiment analysis, Naive Bayes model, natural language processing, topic analysis, fake health news classification model

Procedia PDF Downloads 81
699 A Radiomics Approach to Predict the Evolution of Prostate Imaging Reporting and Data System Score 3/5 Prostate Areas in Multiparametric Magnetic Resonance

Authors: Natascha C. D'Amico, Enzo Grossi, Giovanni Valbusa, Ala Malasevschi, Gianpiero Cardone, Sergio Papa

Abstract:

Purpose: To characterize, through a radiomic approach, the nature of areas classified PI-RADS (Prostate Imaging Reporting and Data System) 3/5, recognized in multiparametric prostate magnetic resonance with T2-weighted (T2w), diffusion and perfusion sequences with paramagnetic contrast. Methods and Materials: 24 cases undergoing multiparametric prostate MR and biopsy were admitted to this pilot study. Clinical outcome of the PI-RADS 3/5 was found through biopsy, finding 8 malignant tumours. The analysed images were acquired with a Philips achieva 1.5T machine with a CE- T2-weighted sequence in the axial plane. Semi-automatic tumour segmentation was carried out on MR images using 3DSlicer image analysis software. 45 shape-based, intensity-based and texture-based features were extracted and represented the input for preprocessing. An evolutionary algorithm (a TWIST system based on KNN algorithm) was used to subdivide the dataset into training and testing set and select features yielding the maximal amount of information. After this pre-processing 20 input variables were selected and different machine learning systems were used to develop a predictive model based on a training testing crossover procedure. Results: The best machine learning system (three-layers feed-forward neural network) obtained a global accuracy of 90% ( 80 % sensitivity and 100% specificity ) with a ROC of 0.82. Conclusion: Machine learning systems coupled with radiomics show a promising potential in distinguishing benign from malign tumours in PI-RADS 3/5 areas.

Keywords: machine learning, MR prostate, PI-Rads 3, radiomics

Procedia PDF Downloads 173
698 Adverse Curing Conditions and Performance of Concrete: Bangladesh Perspective

Authors: T. Manzur

Abstract:

Concrete is the predominant construction material in Bangladesh. In large projects, stringent quality control procedures are usually followed under the supervision of experienced engineers and skilled labors. However, in the case of small projects and particularly at distant locations from major cities, proper quality control is often an issue. It has been found from experience that such quality related issues mainly arise from inappropriate proportioning of concrete mixes and improper curing conditions. In most cases external curing method is followed which requires supply of adequate quantity of water along with proper protection against evaporation. Often these conditions are found missing in the general construction sites and eventually lead to production of weaker concrete both in terms of strength and durability. In this study, an attempt has been made to investigate the performance of general concreting works of the country when subjected to several adverse curing conditions that are quite common in various small to medium construction sites. A total of six different types of adverse curing conditions were simulated in the laboratory and samples were kept under those conditions for several days. A set of samples was also submerged in normal curing condition having proper supply of curing water. Performance of concrete was evaluated in terms of compressive strength, tensile strength, chloride permeability and drying shrinkage. About 37% and 25% reduction in 28-day compressive and tensile strength were observed respectively, for samples subjected to most adverse curing condition as compared to the samples under normal curing conditions. Normal curing concrete exhibited moderate permeability (close to low permeability) whereas concrete under adverse curing conditions showed very high permeability values. Similar results were also obtained for shrinkage tests. This study, thus, will assist concerned engineers and supervisors to understand the importance of quality assurance during the curing period of concrete.

Keywords: adverse, concrete, curing, compressive strength, drying shrinkage, permeability, tensile strength

Procedia PDF Downloads 180
697 Conception of Increasing the Efficiency of Excavation Shoring by Prestressing Diaphragm Walls

Authors: Mateusz Frydrych

Abstract:

The construction of diaphragm walls as excavation shoring as well as part of deep foundations is widely used in geotechnical engineering. Today's design challenges lie in the optimal dimensioning of the cross-section, which is demanded by technological considerations. Also in force is the issue of optimization and sustainable use of construction materials, including reduction of carbon footprint, which is currently a relevant challenge for the construction industry. The author presents the concept of an approach to achieving increased efficiency of diaphragm wall excavation shoring by using structural compression technology. The author proposes to implement prestressed tendons in a non-linear manner in the reinforcement cage. As a result bending moment is reduced, which translates into a reduction in the amount of steel needed in the section, a reduction in displacements, and a reduction in the scratching of the casing, including the achievement of better tightness. This task is rarely seen and has not yet been described in a scientific way in the literature. The author has developed a dynamic numerical model that allows the dimensioning of the cross-section of a prestressed shear wall, as well as the study of casing displacements and cross-sectional forces in any defined computational situation. Numerical software from the Sofistik - open source development environment - was used for the study, and models were validated in Plaxis software . This is an interesting idea that allows for optimizing the execution of construction works and reducing the required resources by using fewer materials and saving time. The author presents the possibilities of a prestressed diaphragm wall, among others, using. The example of a diaphragm wall working as a cantilever at the height of two underground floors without additional strutting or stability protection by using ground anchors. This makes the execution of the work more criminal for the contractor and, as a result, cheaper for the investor.

Keywords: prestressed diaphragm wall, Plaxis, Sofistik, innovation, FEM, optimisation

Procedia PDF Downloads 57
696 Filling the Gaps with Representation: Netflix’s Anne with an E as a Way to Reveal What the Text Hid

Authors: Arkadiusz Adam Gardaś

Abstract:

In his theory of gaps, Wolfgang Iser states that literary texts often lack direct messages. Instead of using straightforward descriptions, authors leave the gaps or blanks, i.e., the spaces within the text that come into existence only when readers fill them with their understanding and experiences. This paper’s aim is to present Iser’s literary theory in an intersectional way by comparing it to the idea of intersemiotic translation. To be more precise, the author uses the example of Netflix’s adaption of Lucy Maud Montgomery’s Anne of Green Gables as a form of rendering a book into a film in such a way that certain textual gaps are filled with film images. Intersemiotic translation is a rendition in which signs of one kind of media are translated into the signs of the other media. Film adaptions are the most common, but not the only, type of intersemiotic translation. In this case, the role of the translator is taken by a screenwriter. A screenwriter’s role can reach beyond the direct meaning presented by the author, and instead, it can delve into the source material (here – a novel) in a deeper way. When it happens, a screenwriter is able to spot the gaps in the text and fill them with images that can later be presented to the viewers. Anne with an E, the Netflix adaption of Montgomery’s novel, may be used as a highly meaningful example of such a rendition. It is due to the fact that the 2017 series was broadcasted more than a hundred years after the first edition of the novel was published. This means that what the author might not have been able to show in her text can now be presented in a more open way. The screenwriter decided to use this opportunity to represent certain groups in the film, i.e., racial and sexual minorities, and women. Nonetheless, the series does not alter the novel; in fact, it adds to it by filling the blanks with more direct images. In the paper, fragments of the first season of Anne with an E are analysed in comparison to its source, the novel by Montgomery. The main purpose of that is to show how intersemiotic translation connected with the Iser’s literary theory can enrich the understanding of works of art, culture, media, and literature.

Keywords: intersemiotic translation, film, literary gaps, representation

Procedia PDF Downloads 296
695 Design and Development of an Optimal Fault Tolerant 3 Degree of Freedom Robotic Manipulator

Authors: Ramish, Farhan Khalique Awan

Abstract:

Kinematic redundancy within the manipulators presents extended dexterity and manipulability to the manipulators. Redundant serial robotic manipulators are very popular in industries due to its competencies to keep away from singularities during normal operation and fault tolerance because of failure of one or more joints. Such fault tolerant manipulators are extraordinarily beneficial in applications where human interference for repair and overhaul is both impossible or tough; like in case of robotic arms for space programs, nuclear applications and so on. The design of this sort of fault tolerant serial 3 DoF manipulator is presented in this paper. This work was the extension of the author’s previous work of designing the simple 3R serial manipulator. This work is the realization of the previous design with optimizing the link lengths for incorporating the feature of fault tolerance. Various measures have been followed by the researchers to quantify the fault tolerance of such redundant manipulators. The fault tolerance in this work has been described in terms of the worst-case measure of relative manipulability that is, in fact, a local measure of optimization that works properly for certain configuration of the manipulators. An optimum fault tolerant Jacobian matrix has been determined first based on prescribed null space properties after which the link parameters have been described to meet the given Jacobian matrix. A solid model of the manipulator was then developed to realize the mathematically rigorous design. Further work was executed on determining the dynamic properties of the fault tolerant design and simulations of the movement for various trajectories have been carried out to evaluate the joint torques. The mathematical model of the system was derived via the Euler-Lagrange approach after which the same has been tested using the RoboAnalyzer© software. The results have been quite in agreement. From the CAD model and dynamic simulation data, the manipulator was fabricated in the workshop and Advanced Machining lab of NED University of Engineering and Technology.

Keywords: fault tolerant, Graham matrix, Jacobian, kinematics, Lagrange-Euler

Procedia PDF Downloads 205
694 Influence of Glenohumeral Joint Approximation Technique on the Cardiovascular System in the Acute Phase after Stroke

Authors: Iva Hereitova, Miroslav Svatek, Vit Novacek

Abstract:

Background and Aim: Autonomic imbalance is one of the complications for immobilized patients in the acute stage after a stroke. The predominance of sympathetic activity significantly increases cardiac activity. The technique of glenohumeral joint approximation may contribute in a non-pharmacological way to the regulation of blood pressure and heart rate in patients in this risk group. The aim of the study was to evaluate the effect of glenohumeral joint approximation on the change in heart rate and blood pressure in immobilized patients in the acute phase after a stroke. Methods: The experimental study bilaterally evaluated heart rate, systolic and diastolic pressure values before and after glenohumeral joint approximation in 40 immobilized participants (72.6 ± 10.2 years) in the acute phase after stroke. The experimental group was compared with 40 healthy participants in the control group (68.6 ± 14.2 years). An SpO2 vital signs monitor and a validated Microlife WatchBP Office blood pressure monitor were used for evaluation. Statistical processing and evaluation were performed in MATLAB R2019 (The Math Works®, Inc., Natick, MA, USA). Results: Approximation of the glenohumeral joint resulted in a statistically significant decrease in systolic and diastolic pressure. An average decrease in systolic pressure for individual groups ranged from 8.2 to 11.3 mmHg (p <0.001). For diastolic pressure, the average decrease ranged from 5.0 - 14.2 mmHg (p <0.001). There was a statistically significant reduction in heart rate (p <0.01) only in patients after ischemic stroke in the inferior cerebral artery. There was the average decrease in heart rate of 3.9 beats per minute (median 4 beats per minute). Conclusion: Approximation of the glenohumeral joint leads to a statistically significant decrease in systolic and diastolic pressure in immobilized patients in the acute phase after stroke.

Keywords: Aproximation technique, Cardiovaskular system, Glenohumeral joint, Stroke

Procedia PDF Downloads 189
693 Identification of miRNA-miRNA Interactions between Virus and Host in Human Cytomegalovirus Infection

Authors: Kai-Yao Huang, Tzong-Yi Lee, Pin-Hao Ho, Tzu-Hao Chang, Cheng-Wei Chang

Abstract:

Background: Human cytomegalovirus (HCMV) infects much people around the world, and there were many researches mention that many diseases were caused by HCMV. To understand the mechanism of HCMV lead to diseases during infection. We observe a microRNA (miRNA) – miRNA interaction between HCMV and host during infection. We found HCMV miRNA sequence component complementary with host miRNA precursors, and we also found that the host miRNA abundances were decrease in HCMV infection. Hence, we focus on the host miRNA which may target by the other HCMV miRNA to find theirs target mRNAs expression and analysis these mRNAs affect what kind of signaling pathway. Interestingly, we found the affected mRNA play an important role in some diseases related pathways, and these diseases had been annotated by HCMV infection. Results: From our analysis procedure, we found 464 human miRNAs might be targeted by 26 HCMV miRNAs and there were 291 human miRNAs shows the concordant decrease trend during HCMV infection. For case study, we found hcmv-miR-US22-5p may regulate hsa-mir-877 and we analysis the KEGG pathway which built by hsa-mir-877 validate target mRNA. Additionally, through survey KEGG Disease database found that these mRNA co-regulate some disease related pathway for instance cancer, nerve disease. However, there were studies annotated that HCMV infection casuse cancer and Alzheimer. Conclusions: This work supply a different scenario of miRNA target interactions(MTIs). In previous study assume miRNA only target to other mRNA. Here we wonder there is possibility that miRNAs might regulate non-mRNA targets, like other miRNAs. In this study, we not only consider the sequence similarity with HCMV miRNAs and human miRNA precursors but also the expression trend of these miRNAs. Then we analysis the human miRNAs validate target mRNAs and its associated KEGG pathway. Finally, we survey related works to validate our investigation.

Keywords: human cytomegalovirus, HCMV, microRNA, miRNA

Procedia PDF Downloads 414
692 A One-Dimensional Modeling Analysis of the Influence of Swirl and Tumble Coefficient in a Single-Cylinder Research Engine

Authors: Mateus Silva Mendonça, Wender Pereira de Oliveira, Gabriel Heleno de Paula Araújo, Hiago Tenório Teixeira Santana Rocha, Augusto César Teixeira Malaquias, José Guilherme Coelho Baeta

Abstract:

The stricter legislation and the greater demand of the population regard to gas emissions and their effects on the environment as well as on human health make the automotive industry reinforce research focused on reducing levels of contamination. This reduction can be achieved through the implementation of improvements in internal combustion engines in such a way that they promote the reduction of both specific fuel consumption and air pollutant emissions. These improvements can be obtained through numerical simulation, which is a technique that works together with experimental tests. The aim of this paper is to build, with support of the GT-Suite software, a one-dimensional model of a single-cylinder research engine to analyze the impact of the variation of swirl and tumble coefficients on the performance and on the air pollutant emissions of an engine. Initially, the discharge coefficient is calculated through the software Converge CFD 3D, given that it is an input parameter in GT-Power. Mesh sensitivity tests are made in 3D geometry built for this purpose, using the mass flow rate in the valve as a reference. In the one-dimensional simulation is adopted the non-predictive combustion model called Three Pressure Analysis (TPA) is, and then data such as mass trapped in cylinder, heat release rate, and accumulated released energy are calculated, aiming that the validation can be performed by comparing these data with those obtained experimentally. Finally, the swirl and tumble coefficients are introduced in their corresponding objects so that their influences can be observed when compared to the results obtained previously.

Keywords: 1D simulation, single-cylinder research engine, swirl coefficient, three pressure analysis, tumble coefficient

Procedia PDF Downloads 85
691 A Flexible Real-Time Eco-Drive Strategy for Electric Minibus

Authors: Felice De Luca, Vincenzo Galdi, Piera Stella, Vito Calderaro, Adriano Campagna, Antonio Piccolo

Abstract:

Sustainable mobility has become one of the major issues of recent years. The challenge in reducing polluting emissions as much as possible has led to the production and diffusion of vehicles with internal combustion engines that are less polluting and to the adoption of green energy vectors, such as vehicles powered by natural gas or LPG and, more recently, with hybrid and electric ones. While on the one hand, the spread of electric vehicles for private use is becoming a reality, albeit rather slowly, not the same is happening for vehicles used for public transport, especially those that operate in the congested areas of the cities. Even if the first electric buses are increasingly being offered on the market, it remains central to the problem of autonomy for battery fed vehicles with high daily routes and little time available for recharging. In fact, at present, solid-state batteries are still too large in size, heavy, and unable to guarantee the required autonomy. Therefore, in order to maximize the energy management on the vehicle, the optimization of driving profiles offer a faster and cheaper contribution to improve vehicle autonomy. In this paper, following the authors’ precedent works on electric vehicles in public transport and energy management strategies in the electric mobility area, an eco-driving strategy for electric bus is presented and validated. Particularly, the characteristics of the prototype bus are described, and a general-purpose eco-drive methodology is briefly presented. The model is firstly simulated in MATLAB™ and then implemented on a mobile device installed on-board of a prototype bus developed by the authors in a previous research project. The solution implemented furnishes the bus-driver suggestions on the guide style to adopt. The result of the test in a real case will be shown to highlight the effectiveness of the solution proposed in terms of energy saving.

Keywords: eco-drive, electric bus, energy management, prototype

Procedia PDF Downloads 116
690 Support for Reporting Guidelines in Surgical Journals Needs Improvement: A Systematic Review

Authors: Riaz A. Agha, Ishani Barai, Shivanchan Rajmohan, Seon Lee, Mohammed O. Anwar, Alex J. Fowler, Dennis P. Orgill, Douglas G. Altman

Abstract:

Introduction: Medical knowledge is growing fast. Evidence-based medicine works best if the evidence is reported well. Past studies have shown reporting quality to be lacking in the field of surgery. Reporting guidelines are an important tool for authors to optimize the reporting of their research. The objective of this study was to analyse the frequency and strength of recommendation for such reporting guidelines within surgical journals. Methods: A systematic review of the 198 journals within the Journal Citation Report 2014 (surgery category) published by Thomson Reuters was undertaken. The online guide for authors for each journal was screened by two independent groups and results were compared. Data regarding the presence and strength of recommendation to use reporting guidelines was extracted. Results: 193 journals were included (as five appeared twice having changed their name). These had a median impact factor of 1.526 (range 0.047 to 8.327), with a median of 145 articles published per journal (range 29-659), with 34,036 articles published in total over the two-year window 2012-2013. The majority (62%) of surgical journals made no mention of reporting guidelines within their guidelines for authors. Of the journals (38%) that did mention them, only 14% (10/73) required the use of all relevant reporting guidelines. The most frequently mentioned reporting guideline was CONSORT (46 journals). Conclusion: The mention of reporting guidelines within the guide for authors of surgical journals needs improvement. Authors, reviewers and editors should work to ensure that research is reported in line with the relevant reporting guidelines. Journals should consider hard-wiring adherence to them. This will allow peer-reviewers to focus on what is present, not what is missing, raising the level of scholarly discourse between authors and the scientific community and reducing frustration amongst readers.

Keywords: CONSORT, guide for authors, PRISMA, reporting guidelines, journal impact factor, citation analysis

Procedia PDF Downloads 452
689 Nurturing of Children with Results from Their Nature (DNA) Using DNA-MILE

Authors: Tan Lay Cheng (Cheryl), Low Huiqi

Abstract:

Background: All children learn at different pace. Individualized learning is an approach that tailors to the individual learning needs of each child. When implementing this approach, educators have to base their lessons on the understanding that all students learn differently and that what works for one student may not work for another. In the current early childhood environment, individualized learning is for children with diverse needs. However, a typical developing child is also able to benefit from individualized learning. This research abstract explores the concept of utilizing DNA-MILE, a patented (in Singapore) DNA-based assessment tool that can be used to measure a variety of factors that can impact learning. The assessment report includes the dominant intelligence of the user or, in this case, the child. From the result, a personalized learning plan that is tailored to each individual student's needs. Methods: A study will be conducted to investigate the effectiveness of DNA-MILE in supporting individualized learning. The study will involve a group of 20 preschoolers who were randomly assigned to either a DNA-MILE-assessed group (experimental group) or a control group. 10 children in each group. The experimental group will receive DNA Mile assessments and personalized learning plans, while the control group will not. The children in the experimental group will be taught using the dominant intelligence (as shown in the DNA-MILE report) to enhance their learning in other domains. The children in the control group will be taught using the curriculum and lesson plan set by their teacher for the whole class. Parents’ and teachers’ interviews will be conducted to provide information about the children before the study and after the study. Results: The results of the study will show the difference in the outcome of the learning, which received DNA Mile assessments and personalized learning plans, significantly outperformed the control group on a variety of measures, including standardized tests, grades, and motivation. Conclusion: The results of this study suggest that DNA Mile can be an effective tool for supporting individualized learning. By providing personalized learning plans, DNA Mile can help to improve learning outcomes for all students.

Keywords: individualized, DNA-MILE, learning, preschool, DNA, multiple intelligence

Procedia PDF Downloads 96
688 Shadows and Symbols: The Tri-Level Importance of Memory in Jane Yolen's 'the Devil's Arithmetic' and Soon-To-Be-Published 'Mapping the Bones'

Authors: Kirsten A. Bartels

Abstract:

'Never again' and 'Lest we forget' have long been messages associated with the events of the Shoah. Yet as we attempt to learn from the past, we must find new ways to engage with its memories. The preservation of the culture and the value of tradition are critical factors in Jane Yolen's works of Holocaust fiction, The Devil's Arithmetic and Mapping the Bones, emphasized through the importance of remembering. That word, in its multitude of forms (remember, remembering, memories), occurs no less than ten times in the first four pages and over one hundred times in the one hundred and sixty-four-page narrative The Devil’s Arithmetic. While Yolen takes a different approach to showcasing the importance of memory in Mapping the Bones, it is of equal import in this work and arguably to the future of Holocaust knowing. The idea of remembering, the desire to remember, and the ability to remember, are explored in three divergent ways in The Devil’s Arithmetic. First, in the importance to remember a past which is not her own – to understand history or acquired memories. Second, in the protagonist's actual or initial memories, those of her life in modern-day New York. Third, in a reverse mode of forgetting and trying to reacquire that which has been lost -- as Hannah is processed in the camp and she forgets everything, all worlds prior to the camp are lost to her. As numbers replace names, Yolen stresses the importance of self-identity or owned memories. In addition, the importance of relaying memory, the transitions of memory from perspective, and the ideas of reflective telling are explored in Mapping the Bones -- through the telling of the story through the lens of one of the twins as the events are unfolding; and then the through the reflective telling from the lens of the other twin. Parallel to the exploration of the intersemiosis of memory is the discussion of literary shadows (foreshadowing, backshadowing, and side-shadowing) and their impact on the reader's experience with Yolen's narrative. For in this type of exploration, one cannot look at the events described in Yolen's work and not also contemplate the figurative shadows cast.

Keywords: holocaust literature, memory, narrative, Yolen

Procedia PDF Downloads 218
687 Detect Critical Thinking Skill in Written Text Analysis. The Use of Artificial Intelligence in Text Analysis vs Chat/Gpt

Authors: Lucilla Crosta, Anthony Edwards

Abstract:

Companies and the market place nowadays struggle to find employees with adequate skills in relation to anticipated growth of their businesses. At least half of workers will need to undertake some form of up-skilling process in the next five years in order to remain aligned with the requests of the market . In order to meet these challenges, there is a clear need to explore the potential uses of AI (artificial Intelligence) based tools in assessing transversal skills (critical thinking, communication and soft skills of different types in general) of workers and adult students while empowering them to develop those same skills in a reliable trustworthy way. Companies seek workers with key transversal skills that can make a difference between workers now and in the future. However, critical thinking seems to be the one of the most imprtant skill, bringing unexplored ideas and company growth in business contexts. What employers have been reporting since years now, is that this skill is lacking in the majority of workers and adult students, and this is particularly visible trough their writing. This paper investigates how critical thinking and communication skills are currently developed in Higher Education environments through use of AI tools at postgraduate levels. It analyses the use of a branch of AI namely Machine Learning and Big Data and of Neural Network Analysis. It also examines the potential effect the acquisition of these skills through AI tools and what kind of effects this has on employability This paper will draw information from researchers and studies both at national (Italy & UK) and international level in Higher Education. The issues associated with the development and use of one specific AI tool Edulai, will be examined in details. Finally comparisons will be also made between these tools and the more recent phenomenon of Chat GPT and forthcomings and drawbacks will be analysed.

Keywords: critical thinking, artificial intelligence, higher education, soft skills, chat GPT

Procedia PDF Downloads 86
686 The Korean Neo-Confucian Ideal of Pluralism and Han

Authors: Hyeon Sop Baek

Abstract:

This paper will investigate the Korean concept of han and suggest that the feeling of han is essentially inseparable from the central project of the Korean neo-Confucian philosophical tradition. Han is a complex sentiment, but one may characterize it as an internally directed complex of sentiments of frustration, sadness, and anger. In particular, this paper aims to demonstrate that the Korean neo-Confucian project's ultimate objective was to build a pluralistic world – where different people can coexist together in harmony and participate in building the ideal world. Nevertheless, the confrontation between the neo-Confucian idea – that every person has the intrinsic potential to be moral – and the bleakness of reality that made their objective virtually impossible to achieve led to the formation and development of the feeling of han. The paper will first examine the concept of han and what it entails and then investigate the core elements of Korean neo-Confucianism, examining the works of Korean neo-Confucians, including Toegye, Yulgok, and Jeong Dojeon. Furthermore, the concept of plurality will be drawn from the political theory of Hannah Arendt. While the Arendtian and Korean neo-Confucian philosophies are ultimately different, this paper will contend that the two philosophies' broader aims share many resonating points. Specifically, within both philosophies, the human plurality – that all humans are equal but not the same – underlies the foundation of an ideal political realm. From there, an argument that the difficulty faced by the neo-Confucians in Korea in constructing a polity based on the ideal of respect and human moral capacity ultimately contributed to the emergence of the sentiment han will be presented. In conclusion, this paper will demonstrate that the ultimate objectives of Korean Confucianism lie in closing the gap between the ideal and reality in moral cultivation as well as its political project of building an ideal, pluralistic world, and han emerges from the realization of the difficulty of achieving that goal. Finally, this paper will contest that han needs not be perceived negatively, and han can be a driving force for political participation in the contemporary democratic, pluralistic society.

Keywords: Korea, Confucianism, neo-Confucianism, philosophy, han, Korean philosophy

Procedia PDF Downloads 120
685 Integrating Best Practices for Construction Waste in Quality Management Systems

Authors: Paola Villoria Sáez, Mercedes Del Río Merino, Jaime Santa Cruz Astorqui, Antonio Rodríguez Sánchez

Abstract:

The Spanish construction industry generates large volumes of waste. However, despite the legislative improvements introduced for construction and demolition waste (CDW), construction waste recycling rate remains well below other European countries and also below the target set for 2020. This situation can be due to many difficulties. i.e.: The difficulty of onsite segregation or the estimation in advance of the total amount generated. Despite these difficulties, the proper management of CDW must be one of the main aspects to be considered by the construction companies. In this sense, some large national companies are implementing Integrated Management Systems (IMS) including not only quality and safety aspects, but also environment issues. However, although this fact is a reality for large construction companies still the vast majority of companies need to adopt this trend. In short, it is common to find in small and medium enterprises a decentralized management system: A single system of quality management, another for system safety management and a third one for environmental management system (EMS). In addition, the EMSs currently used address CDW superficially and are mainly focus on other environmental concerns such as carbon emissions. Therefore, this research determines and implements a specific best practice management system for CDW based on eight procedures in a Spanish Construction company. The main advantages and drawbacks of its implementation are highlighted. Results of this study show that establishing and implementing a CDW management system in building works, improve CDW quantification as the company obtains their own CDW generation ratio. This helps construction stakeholders when developing CDW Management Plans and also helps to achieve a higher adjustment of CDW management costs. Finally, integrating this CDW system with the EMS of the company favors the cohesion of the construction process organization at all stages, establishing responsibilities in the field of waste and providing a greater control over the process.

Keywords: construction and demolition waste, waste management, best practices, waste minimization, building, quality management systems

Procedia PDF Downloads 514
684 Human-Machine Cooperation in Facial Comparison Based on Likelihood Scores

Authors: Lanchi Xie, Zhihui Li, Zhigang Li, Guiqiang Wang, Lei Xu, Yuwen Yan

Abstract:

Image-based facial features can be classified into category recognition features and individual recognition features. Current automated face recognition systems extract a specific feature vector of different dimensions from a facial image according to their pre-trained neural network. However, to improve the efficiency of parameter calculation, an algorithm generally reduces the image details by pooling. The operation will overlook the details concerned much by forensic experts. In our experiment, we adopted a variety of face recognition algorithms based on deep learning, compared a large number of naturally collected face images with the known data of the same person's frontal ID photos. Downscaling and manual handling were performed on the testing images. The results supported that the facial recognition algorithms based on deep learning detected structural and morphological information and rarely focused on specific markers such as stains and moles. Overall performance, distribution of genuine scores and impostor scores, and likelihood ratios were tested to evaluate the accuracy of biometric systems and forensic experts. Experiments showed that the biometric systems were skilled in distinguishing category features, and forensic experts were better at discovering the individual features of human faces. In the proposed approach, a fusion was performed at the score level. At the specified false accept rate, the framework achieved a lower false reject rate. This paper contributes to improving the interpretability of the objective method of facial comparison and provides a novel method for human-machine collaboration in this field.

Keywords: likelihood ratio, automated facial recognition, facial comparison, biometrics

Procedia PDF Downloads 113
683 Local Directional Encoded Derivative Binary Pattern Based Coral Image Classification Using Weighted Distance Gray Wolf Optimization Algorithm

Authors: Annalakshmi G., Sakthivel Murugan S.

Abstract:

This paper presents a local directional encoded derivative binary pattern (LDEDBP) feature extraction method that can be applied for the classification of submarine coral reef images. The classification of coral reef images using texture features is difficult due to the dissimilarities in class samples. In coral reef image classification, texture features are extracted using the proposed method called local directional encoded derivative binary pattern (LDEDBP). The proposed approach extracts the complete structural arrangement of the local region using local binary batten (LBP) and also extracts the edge information using local directional pattern (LDP) from the edge response available in a particular region, thereby achieving extra discriminative feature value. Typically the LDP extracts the edge details in all eight directions. The process of integrating edge responses along with the local binary pattern achieves a more robust texture descriptor than the other descriptors used in texture feature extraction methods. Finally, the proposed technique is applied to an extreme learning machine (ELM) method with a meta-heuristic algorithm known as weighted distance grey wolf optimizer (GWO) to optimize the input weight and biases of single-hidden-layer feed-forward neural networks (SLFN). In the empirical results, ELM-WDGWO demonstrated their better performance in terms of accuracy on all coral datasets, namely RSMAS, EILAT, EILAT2, and MLC, compared with other state-of-the-art algorithms. The proposed method achieves the highest overall classification accuracy of 94% compared to the other state of art methods.

Keywords: feature extraction, local directional pattern, ELM classifier, GWO optimization

Procedia PDF Downloads 146
682 Investigation of Different Machine Learning Algorithms in Large-Scale Land Cover Mapping within the Google Earth Engine

Authors: Amin Naboureh, Ainong Li, Jinhu Bian, Guangbin Lei, Hamid Ebrahimy

Abstract:

Large-scale land cover mapping has become a new challenge in land change and remote sensing field because of involving a big volume of data. Moreover, selecting the right classification method, especially when there are different types of landscapes in the study area is quite difficult. This paper is an attempt to compare the performance of different machine learning (ML) algorithms for generating a land cover map of the China-Central Asia–West Asia Corridor that is considered as one of the main parts of the Belt and Road Initiative project (BRI). The cloud-based Google Earth Engine (GEE) platform was used for generating a land cover map for the study area from Landsat-8 images (2017) by applying three frequently used ML algorithms including random forest (RF), support vector machine (SVM), and artificial neural network (ANN). The selected ML algorithms (RF, SVM, and ANN) were trained and tested using reference data obtained from MODIS yearly land cover product and very high-resolution satellite images. The finding of the study illustrated that among three frequently used ML algorithms, RF with 91% overall accuracy had the best result in producing a land cover map for the China-Central Asia–West Asia Corridor whereas ANN showed the worst result with 85% overall accuracy. The great performance of the GEE in applying different ML algorithms and handling huge volume of remotely sensed data in the present study showed that it could also help the researchers to generate reliable long-term land cover change maps. The finding of this research has great importance for decision-makers and BRI’s authorities in strategic land use planning.

Keywords: land cover, google earth engine, machine learning, remote sensing

Procedia PDF Downloads 104