Search results for: artificial microRNA approach
13993 Creation of Processes for a Safety Element Out of Context for an Actuator Circuit Control Module
Authors: Hassan Noun, Christian Urban-Seelmann, Mohamed Abdelfattah, Guillaume Zeller, Rajesh G., Iryna Mozgova, Roland Lachmayer
Abstract:
Several modules in automotive are usually modified and adapted for various project-specific applications. Due to a standardized safety concept, high reusability is accessible. A safety element out of context (SEooC) according to ISO 26262 can be a suitable approach. Based on the same safety concept and analysis, common modules can reach high usability. For developing according to a module out of context, an appropriate and detailed development approach is required. This paper shows how to derive these development processes for platform modules. Therefore, the detailed approach to the safety element out of context is derived. The aim is to create a detailed workflow for all phases of the development and integration of any kind of system modules. As an application example, an automotive project for an actuator control module is considered.Keywords: functional safety, engineering processes, system engineering, electronic engineering
Procedia PDF Downloads 14413992 Experiential Learning: A Case Study for Teaching Operating System Using C and Unix
Authors: Shamshuddin K., Nagaraj Vannal, Diwakar Kulkarni, Raghavendra Nakod
Abstract:
In most of the universities and colleges Operating System (OS) course is treated as theoretical and usually taught in a classroom using conventional teaching methods. In this paper we are presenting a new approach of teaching OS through experiential learning, the course is designed to suit the requirement of undergraduate engineering program of Instrumentation Technology. This new approach has benefited us to improve our student’s programming skills, presentation skills and understanding of the operating system concepts.Keywords: pedagogy, interactive learning, experiential learning, OS, C, UNIX
Procedia PDF Downloads 60613991 Radioactivity Assessment of Sediments in Negombo Lagoon Sri Lanka
Authors: H. M. N. L. Handagiripathira
Abstract:
The distributions of naturally occurring and anthropogenic radioactive materials were determined in surface sediments taken at 27 different locations along the bank of Negombo Lagoon in Sri Lanka. Hydrographic parameters of lagoon water and the grain size analyses of the sediment samples were also carried out for this study. The conductivity of the adjacent water was varied from 13.6 mS/cm to 55.4 mS/cm near to the southern end and the northern end of the lagoon, respectively, and equally salinity levels varied from 7.2 psu to 32.1 psu. The average pH in the water was 7.6 and average water temperature was 28.7 °C. The grain size analysis emphasized the mass fractions of the samples as sand (60.9%), fine sand (30.6%) and fine silt+clay (1.3%) in the sampling locations. The surface sediment samples of wet weight, 1 kg each from upper 5-10 cm layer, were oven dried at 105 °C for 24 hours to get a constant weight, homogenized and sieved through a 2 mm sieve (IAEA technical series no. 295). The radioactivity concentrations were determined using gamma spectrometry technique. Ultra Low Background Broad Energy High Purity Ge Detector, BEGe (Model BE5030, Canberra) was used for radioactivity measurement with Canberra Industries' Laboratory Source-less Calibration Software (LabSOCS) mathematical efficiency calibration approach and Geometry composer software. The mean activity concentration was found to be 24 ± 4, 67 ± 9, 181 ± 10, 59 ± 8, 3.5 ± 0.4 and 0.47 ± 0.08 Bq/kg for 238U, 232Th, 40K, 210Pb, 235U and 137Cs respectively. The mean absorbed dose rate in air, radium equivalent activity, external hazard index, annual gonadal dose equivalent and annual effective dose equivalent were 60.8 nGy/h, 137.3 Bq/kg, 0.4, 425.3 mSv/year and 74.6 mSv/year, respectively. The results of this study will provide baseline information on the natural and artificial radioactive isotopes and environmental pollution associated with information on radiological risk.Keywords: gamma spectrometry, lagoon, radioactivity, sediments
Procedia PDF Downloads 13913990 Optical Multicast over OBS Networks: An Approach Based on Code-Words and Tunable Decoders
Authors: Maha Sliti, Walid Abdallah, Noureddine Boudriga
Abstract:
In the frame of this work, we present an optical multicasting approach based on optical code-words. Our approach associates, in the edge node, an optical code-word to a group multicast address. In the core node, a set of tunable decoders are used to send a traffic data to multiple destinations based on the received code-word. The use of code-words, which correspond to the combination of an input port and a set of output ports, allows the implementation of an optical switching matrix. At the reception of a burst, it will be delayed in an optical memory. And, the received optical code-word is split to a set of tunable optical decoders. When it matches a configured code-word, the delayed burst is switched to a set of output ports.Keywords: optical multicast, optical burst switching networks, optical code-words, tunable decoder, virtual optical memory
Procedia PDF Downloads 60713989 Evaluating the Impact of Landscape Values Associated With the Landscape Developemnt Approach of Neighbourhood Gardens; In Tier Two Cities of India; On Users’ Perception Towards the Space. Case: City of Nashik, Maharashtra, India
Authors: Anandi Anant Lale, Pooja Sadananda Patil
Abstract:
Neighbourhood gardens (NGs), in the rapidly growing tier two cities of India, play a pivotal role in maintaining and enhancing the quality of life of the dwellers in terms of mental, physical and socio- cultural well-being. They are the breathing areas which avail the opportunity of accessing nature while being in the close proximity of modern infrastructural provisions of the neighbourhood. In this article, the landscape values (viz. Cultural, Functional, Environmental and Perceptual) associated with the landscape development approach of neighbourhood gardens in the city of Nashik; one of the major tier two cities of Maharashtra; India, are studied through physical survey of selected NGs and the respective neighborhoods. Contextual study of the selected neighbourhood with the emphasis on dwellers' response in terms of physical as well as mental associations with the NGs is recorded through visitors' interviews. Analysis of interrelation of the landscape values and the users' response to the NGs revealed that each landscape value associated with the landscape development approach, has impact of diverse intensity on the users' perception, in different neighbourhoods. Contextual needs of selected neighbourhoods govern the user's perception towards the respective NGs and eventually define the role of landscape value/s associated with the landscape development approach of NG in deciding the competence of the space. The findings of the study can form the basis to redefine the landscape development approach for the future NGs in tier two cities of India that will justify the contextual needs of every neighbourhood through the emphasis of landscape values.Keywords: neighbourhood garden, landscape value, user’s perception, context, landscape development
Procedia PDF Downloads 11313988 A Comprehensive Survey of Artificial Intelligence and Machine Learning Approaches across Distinct Phases of Wildland Fire Management
Authors: Ursula Das, Manavjit Singh Dhindsa, Kshirasagar Naik, Marzia Zaman, Richard Purcell, Srinivas Sampalli, Abdul Mutakabbir, Chung-Horng Lung, Thambirajah Ravichandran
Abstract:
Wildland fires, also known as forest fires or wildfires, are exhibiting an alarming surge in frequency in recent times, further adding to its perennial global concern. Forest fires often lead to devastating consequences ranging from loss of healthy forest foliage and wildlife to substantial economic losses and the tragic loss of human lives. Despite the existence of substantial literature on the detection of active forest fires, numerous potential research avenues in forest fire management, such as preventative measures and ancillary effects of forest fires, remain largely underexplored. This paper undertakes a systematic review of these underexplored areas in forest fire research, meticulously categorizing them into distinct phases, namely pre-fire, during-fire, and post-fire stages. The pre-fire phase encompasses the assessment of fire risk, analysis of fuel properties, and other activities aimed at preventing or reducing the risk of forest fires. The during-fire phase includes activities aimed at reducing the impact of active forest fires, such as the detection and localization of active fires, optimization of wildfire suppression methods, and prediction of the behavior of active fires. The post-fire phase involves analyzing the impact of forest fires on various aspects, such as the extent of damage in forest areas, post-fire regeneration of forests, impact on wildlife, economic losses, and health impacts from byproducts produced during burning. A comprehensive understanding of the three stages is imperative for effective forest fire management and mitigation of the impact of forest fires on both ecological systems and human well-being. Artificial intelligence and machine learning (AI/ML) methods have garnered much attention in the cyber-physical systems domain in recent times leading to their adoption in decision-making in diverse applications including disaster management. This paper explores the current state of AI/ML applications for managing the activities in the aforementioned phases of forest fire. While conventional machine learning and deep learning methods have been extensively explored for the prevention, detection, and management of forest fires, a systematic classification of these methods into distinct AI research domains is conspicuously absent. This paper gives a comprehensive overview of the state of forest fire research across more recent and prominent AI/ML disciplines, including big data, classical machine learning, computer vision, explainable AI, generative AI, natural language processing, optimization algorithms, and time series forecasting. By providing a detailed overview of the potential areas of research and identifying the diverse ways AI/ML can be employed in forest fire research, this paper aims to serve as a roadmap for future investigations in this domain.Keywords: artificial intelligence, computer vision, deep learning, during-fire activities, forest fire management, machine learning, pre-fire activities, post-fire activities
Procedia PDF Downloads 7213987 Using MALDI-TOF MS to Detect Environmental Microplastics (Polyethylene, Polyethylene Terephthalate, and Polystyrene) within a Simulated Tissue Sample
Authors: Kara J. Coffman-Rea, Karen E. Samonds
Abstract:
Microplastic pollution is an urgent global threat to our planet and human health. Microplastic particles have been detected within our food, water, and atmosphere, and found within the human stool, placenta, and lung tissue. However, most spectrometric microplastic detection methods require chemical digestion which can alter or destroy microplastic particles and makes it impossible to acquire information about their in-situ distribution. MALDI TOF MS (Matrix-assisted laser desorption ionization-time of flight mass spectrometry) is an analytical method using a soft ionization technique that can be used for polymer analysis. This method provides a valuable opportunity to both acquire information regarding the in-situ distribution of microplastics and also minimizes the destructive element of chemical digestion. In addition, MALDI TOF MS allows for expanded analysis of the microplastics including detection of specific additives that may be present within them. MALDI TOF MS is particularly sensitive to sample preparation and has not yet been used to analyze environmental microplastics within their specific location (e.g., biological tissues, sediment, water). In this study, microplastics were created using polyethylene gloves, polystyrene micro-foam, and polyethylene terephthalate cable sleeving. Plastics were frozen using liquid nitrogen and ground to obtain small fragments. An artificial tissue was created using a cellulose sponge as scaffolding coated with a MaxGel Extracellular Matrix to simulate human lung tissue. Optimal preparation techniques (e.g., matrix, cationization reagent, solvent, mixing ratio, laser intensity) were first established for each specific polymer type. The artificial tissue sample was subsequently spiked with microplastics, and specific polymers were detected using MALDI-TOF-MS. This study presents a novel method for the detection of environmental polyethylene, polyethylene terephthalate, and polystyrene microplastics within a complex sample. Results of this study provide an effective method that can be used in future microplastics research and can aid in determining the potential threats to environmental and human health that they pose.Keywords: environmental plastic pollution, MALDI-TOF MS, microplastics, polymer identification
Procedia PDF Downloads 25613986 Artificial Intelligence and Development: The Missing Link
Authors: Driss Kettani
Abstract:
ICT4D actors are naturally attempted to include AI in the range of enabling technologies and tools that could support and boost the Development process, and to refer to these as AI4D. But, doing so, assumes that AI complies with the very specific features of ICT4D context, including, among others, affordability, relevance, openness, and ownership. Clearly, none of these is fulfilled, and the enthusiastic posture that AI4D is a natural part of ICT4D is not grounded and, to certain extent, does not serve the purpose of Technology for Development at all. In the context of Development, it is important to emphasize and prioritize ICT4D, in the national digital transformation strategies, instead of borrowing "trendy" waves of the IT Industry that are motivated by business considerations, with no specific care/consideration to Development.Keywords: AI, ICT4D, technology for development, position paper
Procedia PDF Downloads 8813985 Surveillance of Super-Extended Objects: Bimodal Approach
Authors: Andrey V. Timofeev, Dmitry Egorov
Abstract:
This paper describes an effective solution to the task of a remote monitoring of super-extended objects (oil and gas pipeline, railways, national frontier). The suggested solution is based on the principle of simultaneously monitoring of seismoacoustic and optical/infrared physical fields. The principle of simultaneous monitoring of those fields is not new but in contrast to the known solutions the suggested approach allows to control super-extended objects with very limited operational costs. So-called C-OTDR (Coherent Optical Time Domain Reflectometer) systems are used to monitor the seismoacoustic field. Far-CCTV systems are used to monitor the optical/infrared field. A simultaneous data processing provided by both systems allows effectively detecting and classifying target activities, which appear in the monitored objects vicinity. The results of practical usage had shown high effectiveness of the suggested approach.Keywords: C-OTDR monitoring system, bimodal processing, LPboost, SVM
Procedia PDF Downloads 47013984 Developing University EFL Students’ Communicative Competence by Using Communicative Approach
Authors: Mutwakel Abdalla Ali Garalzain
Abstract:
The aim of this study is to develop university EFL students’ communicative competence. The descriptive, analytical method was used in this study. To collect the data, the researcher designed two questionnaires, one for university EFL students and the other for English language teachers. The respondents of the study were eighty-eight; 76 university EFL students, and 12 English language teachers. The data obtained were analyzed by using statistical package for social science (SPSS). The findings of the study have revealed that most of the university EFL students are unable to express their ideas properly, although they have an abundance of vocabulary. The findings of the study have also shown that most of the university EFL students have positive attitudes towards communicative competence. The results of the study also identified the best strategies that can be used to enhance university EFL students’ communicative competence in English language teaching. The study recommends that English language textbooks should be compatible with the requirements of the student-centered approach. It also recommends that English language teachers should adopt the communicative approach’s strategies in the EFL classroom.Keywords: applied linguistics, communicative competence , English language teaching, university EFL students
Procedia PDF Downloads 19813983 An Intelligence-Led Methodologly for Detecting Dark Actors in Human Trafficking Networks
Authors: Andrew D. Henshaw, James M. Austin
Abstract:
Introduction: Human trafficking is an increasingly serious transnational criminal enterprise and social security issue. Despite ongoing efforts to mitigate the phenomenon and a significant expansion of security scrutiny over past decades, it is not receding. This is true for many nations in Southeast Asia, widely recognized as the global hub for trafficked persons, including men, women, and children. Clearly, human trafficking is difficult to address because there are numerous drivers, causes, and motivators for it to persist, such as non-military and non-traditional security challenges, i.e., climate change, global warming displacement, and natural disasters. These make displaced persons and refugees particularly vulnerable. The issue is so large conservative estimates put a dollar value at around $150 billion-plus per year (Niethammer, 2020) spanning sexual slavery and exploitation, forced labor, construction, mining and in conflict roles, and forced marriages of girls and women. Coupled with corruption throughout military, police, and civil authorities around the world, and the active hands of powerful transnational criminal organizations, it is likely that such figures are grossly underestimated as human trafficking is misreported, under-detected, and deliberately obfuscated to protect those profiting from it. For example, the 2022 UN report on human trafficking shows a 56% reduction in convictions in that year alone (UNODC, 2022). Our Approach: To better understand this, our research utilizes a bespoke methodology. Applying a JAM (Juxtaposition Assessment Matrix), which we previously developed to detect flows of dark money around the globe (Henshaw, A & Austin, J, 2021), we now focus on the human trafficking paradigm. Indeed, utilizing a JAM methodology has identified key indicators of human trafficking not previously explored in depth. Being a set of structured analytical techniques that provide panoramic interpretations of the subject matter, this iteration of the JAM further incorporates behavioral and driver indicators, including the employment of Open-Source Artificial Intelligence (OS-AI) across multiple collection points. The extracted behavioral data was then applied to identify non-traditional indicators as they contribute to human trafficking. Furthermore, as the JAM OS-AI analyses data from the inverted position, i.e., the viewpoint of the traffickers, it examines the behavioral and physical traits required to succeed. This transposed examination of the requirements of success delivers potential leverage points for exploitation in the fight against human trafficking in a new and novel way. Findings: Our approach identified new innovative datasets that have previously been overlooked or, at best, undervalued. For example, the JAM OS-AI approach identified critical 'dark agent' lynchpins within human trafficking that are difficult to detect and harder to connect to actors and agents within a network. Our preliminary data suggests this is in part due to the fact that ‘dark agents’ in extant research have been difficult to detect and potentially much harder to directly connect to the actors and organizations in human trafficking networks. Our research demonstrates that using new investigative techniques such as OS-AI-aided JAM introduces a powerful toolset to increase understanding of human trafficking and transnational crime and illuminate networks that, to date, avoid global law enforcement scrutiny.Keywords: human trafficking, open-source intelligence, transnational crime, human security, international human rights, intelligence analysis, JAM OS-AI, Dark Money
Procedia PDF Downloads 9013982 Use of the SWEAT Analysis Approach to Determine the Effectiveness of a School's Implementation of Its Curriculum
Authors: Prakash Singh
Abstract:
The focus of this study is on the use of the SWEAT analysis approach to determine how effectively a school, as an organization, has implemented its curriculum. To gauge the feelings of the teaching staff, unstructured interviews were employed in this study, asking the participants for their ideas and opinions on each of the three identified aspects of the school: instructional materials, media and technology; teachers’ professional competencies; and the curriculum. This investigation was based on the five key components of the SWEAT model: strengths, weaknesses, expectations, abilities, and tensions. The findings of this exploratory study evoke the significance of the SWEAT achievement model as a tool for strategic analysis to be undertaken in any organization. The findings further affirm the usefulness of this analytical tool for human resource development. Employees have expectations, but competency gaps in their professional abilities may hinder them from fulfilling their tasks in terms of their job description. Also, tensions in the working environment can contribute to their experiences of tobephobia (fear of failure). The SWEAT analysis approach detects such shortcomings in any organization and can therefore culminate in the development of programmes to address such concerns. The strategic SWEAT analysis process can provide a clear distinction between success and failure, and between mediocrity and excellence in organizations. However, more research needs to be done on the effectiveness of the SWEAT analysis approach as a strategic analytical tool.Keywords: SWEAT analysis, strategic analysis, tobephobia, competency gaps
Procedia PDF Downloads 50713981 Chatbots in Education: Case of Development Using a Chatbot Development Platform
Authors: Dulani Jayasuriya
Abstract:
This study outlines the developmental steps of a chatbot for administrative purposes of a large undergraduate course. The chatbot is able to handle student queries about administrative details, including assessment deadlines, course documentation, how to navigate the course, group formation, etc. The development window screenshots are that of a free account on the Snatchbot platform such that this can be adopted by the wider public. While only one connection to an answer based on possible keywords is shown here, one needs to develop multiple connections leading to different answers based on different keywords for the actual chatbot to function. The overall flow of the chatbot showing connections between different interactions is depicted at the end.Keywords: chatbots, education, technology, snatch bot, artificial intelligence
Procedia PDF Downloads 10413980 Snapchat’s Scanning Feature
Authors: Reham Banwair, Lana Alshehri, Sara Hadrawi
Abstract:
The purpose of this project is to identify user satisfaction with the AI functions on Snapchat, in order to generate improvement proposals that allow its development within the app. To achieve this, a qualitative analysis was carried out through interviews to people who usually use the application, revealing their satisfaction or dissatisfaction with the usefulness of the AI. In addition, the background of the company and its introduction in these algorithms were analyzed. Furthermore, the characteristics of the three main functions of AI were explained: identify songs, solve mathematical problems, and recognize plants. As a result, it was obtained that 50% still do not know the characteristics of AI, 50% still believe song recognition is not always correct, 41.7% believe that math problems are usually accurate and 91.7% believes the plant detection tool is working properly.Keywords: artificial intelligence, scanning, Snapchat, machine learning
Procedia PDF Downloads 13413979 Design and Implementation of Wireless Syncronized AI System for Security
Authors: Saradha Priya
Abstract:
Developing virtual human is very important to meet the challenges occurred in many applications where human find difficult or risky to perform the task. A robot is a machine that can perform a task automatically or with guidance. Robotics is generally a combination of artificial intelligence and physical machines (motors). Computational intelligence involves the programmed instructions. This project proposes a robotic vehicle that has a camera, PIR sensor and text command based movement. It is specially designed to perform surveillance and other few tasks in the most efficient way. Serial communication has been occurred between a remote Base Station, GUI Application, and PC.Keywords: Zigbee, camera, pirsensor, wireless transmission, DC motor
Procedia PDF Downloads 34913978 Applying Knowledge Management and Attitude Based on Holistic Approach in Learning Andragogy, as an Effort to Solve Environmental Problems after Mining Activities
Authors: Aloysius Hardoko, Susilo
Abstract:
The root cause of environmental damage post coal mining activities as determined by the province of East Kalimantan as a corridor of economic activity masterplan acceleration of economic development expansion (MP3EI) is the behavior of adults. Adult behavior can be changed through knowledge management and attitude. Based on the root of the problem, the objective of the research is to apply knowledge management and attitude based on holistic approach in learning andragogy as an effort to solve environmental problems after coal mining activities. Research methods to achieve the objective of using quantitative research with pretest posttest group design. Knowledge management and attitudes based on a holistic approach in adult learning are applied through initial learning activities, core and case-based cover of environmental damage. The research instrument is a description of the case of environmental damage. The data analysis uses t-test to see the effect of knowledge management attitude based on holistic approach before and after adult learning. Location and sample of representative research of adults as many as 20 people in Kutai Kertanegara District, one of the districts in East Kalimantan province, which suffered the worst environmental damage. The conclusion of the research result is the application of knowledge management and attitude in adult learning influence to adult knowledge and attitude to overcome environmental problem post coal mining activity.Keywords: knowledge management and attitude, holistic approach, andragogy learning, environmental damage
Procedia PDF Downloads 24113977 Corrosion Interaction Between Steel and Acid Mine Drainage: Use of AI Based on Fuzzy Logic
Authors: Maria Luisa de la Torre, Javier Aroba, Jose Miguel Davila, Aguasanta M. Sarmiento
Abstract:
Steel is one of the most widely used materials in polymetallic sulfide mining installations. One of the main problems suffered by these facilities is the economic losses due to the corrosion of this material, which is accelerated and aggravated by the contact with acid waters generated in these mines when sulfides come into contact with oxygen and water. This generation of acidic water, in turn, is accelerated by the presence of acidophilic bacteria. In order to gain a more detailed understanding of this corrosion process and the interaction between steel and acidic water, a laboratory experiment was carried out in which carbon steel plates were introduced into four different solutions for 27 days: distilled water (BK), which tried to assimilate the effect produced by rain on this material, an acid solution from a mine with a high Fe2+/Fe3+ (PO) content, another acid solution of water from another mine with a high Fe3+/Fe2+ (PH) content and, finally, one that reproduced the acid mine water with a high Fe2+/Fe3+ content but in which there were no bacteria (ST). Every 24 hours, physicochemical parameters were measured, and water samples were taken to carry out an analysis of the dissolved elements. The results of these measurements were processed using an explainable AI model based on fuzzy logic. It could be seen that, in all cases, there was an increase in pH, as well as in the concentrations of Fe and, in particular, Fe(II), as a consequence of the oxidation of the steel plates. Proportionally, the increase in Fe concentration was higher in PO and ST than in PH because Fe precipitates were produced in the latter. The rise of Fe(II) was proportionally much higher in PH, especially in the first hours of exposure, because it started from a lower initial concentration of this ion. Although to a lesser extent than in PH, the greater increase in Fe(II) also occurred faster in PO than in ST, a consequence of the action of the catalytic bacteria. On the other hand, Cu concentrations decreased throughout the experiment (with the exception of distilled water, which initially had no Cu, as a result of an electrochemical process that generates a precipitation of Cu together with Fe hydroxides. This decrease is lower in PH because the high total acidity keeps it in solution for a longer time. With the application of an artificial intelligence tool, it has been possible to evaluate the effects of steel corrosion in mining environments, corroborating and extending what was obtained by means of classical statistics.Keywords: acid mine drainage, artificial intelligence, carbon steel, corrosion, fuzzy logic
Procedia PDF Downloads 813976 A Weighted Approach to Unconstrained Iris Recognition
Authors: Yao-Hong Tsai
Abstract:
This paper presents a weighted approach to unconstrained iris recognition. Nowadays, commercial systems are usually characterized by strong acquisition constraints based on the subject’s cooperation. However, it is not always achievable for real scenarios in our daily life. Researchers have been focused on reducing these constraints and maintaining the performance of the system by new techniques at the same time. With large variation in the environment, there are two main improvements to develop the proposed iris recognition system. For solving extremely uneven lighting condition, statistic based illumination normalization is first used on eye region to increase the accuracy of iris feature. The detection of the iris image is based on Adaboost algorithm. Secondly, the weighted approach is designed by Gaussian functions according to the distance to the center of the iris. Furthermore, local binary pattern (LBP) histogram is then applied to texture classification with the weight. Experiment showed that the proposed system provided users a more flexible and feasible way to interact with the verification system through iris recognition.Keywords: authentication, iris recognition, adaboost, local binary pattern
Procedia PDF Downloads 22513975 Automatic Content Curation of Visual Heritage
Authors: Delphine Ribes Lemay, Valentine Bernasconi, André Andrade, Lara DéFayes, Mathieu Salzmann, FréDéRic Kaplan, Nicolas Henchoz
Abstract:
Digitization and preservation of large heritage induce high maintenance costs to keep up with the technical standards and ensure sustainable access. Creating impactful usage is instrumental to justify the resources for long-term preservation. The Museum für Gestaltung of Zurich holds one of the biggest poster collections of the world from which 52’000 were digitised. In the process of building a digital installation to valorize the collection, one objective was to develop an algorithm capable of predicting the next poster to show according to the ones already displayed. The work presented here describes the steps to build an algorithm able to automatically create sequences of posters reflecting associations performed by curator and professional designers. The exposed challenge finds similarities with the domain of song playlist algorithms. Recently, artificial intelligence techniques and more specifically, deep-learning algorithms have been used to facilitate their generations. Promising results were found thanks to Recurrent Neural Networks (RNN) trained on manually generated playlist and paired with clusters of extracted features from songs. We used the same principles to create the proposed algorithm but applied to a challenging medium, posters. First, a convolutional autoencoder was trained to extract features of the posters. The 52’000 digital posters were used as a training set. Poster features were then clustered. Next, an RNN learned to predict the next cluster according to the previous ones. RNN training set was composed of poster sequences extracted from a collection of books from the Gestaltung Museum of Zurich dedicated to displaying posters. Finally, within the predicted cluster, the poster with the best proximity compared to the previous poster is selected. The mean square distance between features of posters was used to compute the proximity. To validate the predictive model, we compared sequences of 15 posters produced by our model to randomly and manually generated sequences. Manual sequences were created by a professional graphic designer. We asked 21 participants working as professional graphic designers to sort the sequences from the one with the strongest graphic line to the one with the weakest and to motivate their answer with a short description. The sequences produced by the designer were ranked first 60%, second 25% and third 15% of the time. The sequences produced by our predictive model were ranked first 25%, second 45% and third 30% of the time. The sequences produced randomly were ranked first 15%, second 29%, and third 55% of the time. Compared to designer sequences, and as reported by participants, model and random sequences lacked thematic continuity. According to the results, the proposed model is able to generate better poster sequencing compared to random sampling. Eventually, our algorithm is sometimes able to outperform a professional designer. As a next step, the proposed algorithm should include a possibility to create sequences according to a selected theme. To conclude, this work shows the potentiality of artificial intelligence techniques to learn from existing content and provide a tool to curate large sets of data, with a permanent renewal of the presented content.Keywords: Artificial Intelligence, Digital Humanities, serendipity, design research
Procedia PDF Downloads 18413974 Wasting Human and Computer Resources
Authors: Mária Csernoch, Piroska Biró
Abstract:
The legends about “user-friendly” and “easy-to-use” birotical tools (computer-related office tools) have been spreading and misleading end-users. This approach has led us to the extremely high number of incorrect documents, causing serious financial losses in the creating, modifying, and retrieving processes. Our research proved that there are at least two sources of this underachievement: (1) The lack of the definition of the correctly edited, formatted documents. Consequently, end-users do not know whether their methods and results are correct or not. They are not aware of their ignorance. They are so ignorant that their ignorance does not allow them to realize their lack of knowledge. (2) The end-users’ problem-solving methods. We have found that in non-traditional programming environments end-users apply, almost exclusively, surface approach metacognitive methods to carry out their computer related activities, which are proved less effective than deep approach methods. Based on these findings we have developed deep approach methods which are based on and adapted from traditional programming languages. In this study, we focus on the most popular type of birotical documents, the text-based documents. We have provided the definition of the correctly edited text, and based on this definition, adapted the debugging method known in programming. According to the method, before the realization of text editing, a thorough debugging of already existing texts and the categorization of errors are carried out. With this method in advance to real text editing users learn the requirements of text-based documents and also of the correctly formatted text. The method has been proved much more effective than the previously applied surface approach methods. The advantages of the method are that the real text handling requires much less human and computer sources than clicking aimlessly in the GUI (Graphical User Interface), and the data retrieval is much more effective than from error-prone documents.Keywords: deep approach metacognitive methods, error-prone birotical documents, financial losses, human and computer resources
Procedia PDF Downloads 38213973 Investigation of the Factors Influencing the Construction Planning Process Using Participant Observation Method
Authors: Ashokkumar Subbiah
Abstract:
This study investigates the impact of factors that influenced the success of construction planning for a major construction project in Qatar. An approach of participant observation is adopted which is informed by the principles of ethnography: one that reports the participants’ view of their world rather than imposing an artificial theoretical framework upon it. As participant observant, key factors were observed and identified that had an impact on the management and execution of the construction planning. It is found that a ‘shadow culture’ exists between the project participants which, it is argued, is only observable from the perspective of an embedded participant observer. The shadow culture acts to enable the management of the planning process, and its efficacy relates to the ‘quality’ of human inter-relationships amongst immediate stakeholders. Whilst this study uses the concept of shadow culture, it is treated as both a methodological stance and one of the findings of this research in the context of the major construction project in Qatar. The concept of shadow culture is not imposed upon the findings, but instead is used as a research tool: respondents report their own worldview and this is reported from the view of a participant observant in a manner that is understandable and useful to those who are not part of the construction project. The findings of this study identify similar factors influencing the planning process of the Qatar project, but the shadow culture predominantly influences these factors towards the failure of planning process. The research concludes by questioning the assumption that construction planning is a mechanistic process that has to be conducted solely by the planning team. Instead, it is a highly social phenomenon in which the seemingly mechanistic process is made workable by the quality of relationships that exist in the project. Drawing on this the final section provides a series of recommendations that may be helpful in enhancing the efficacy of project planning; these include better training/education at the pre-construction phase; recognition of the importance of shadow processes at management levels, and better appreciation of the impact of contract type and chosen procurement route.Keywords: construction planning, participant observation, project participants, shadow culture
Procedia PDF Downloads 29813972 River Stage-Discharge Forecasting Based on Multiple-Gauge Strategy Using EEMD-DWT-LSSVM Approach
Authors: Farhad Alizadeh, Alireza Faregh Gharamaleki, Mojtaba Jalilzadeh, Houshang Gholami, Ali Akhoundzadeh
Abstract:
This study presented hybrid pre-processing approach along with a conceptual model to enhance the accuracy of river discharge prediction. In order to achieve this goal, Ensemble Empirical Mode Decomposition algorithm (EEMD), Discrete Wavelet Transform (DWT) and Mutual Information (MI) were employed as a hybrid pre-processing approach conjugated to Least Square Support Vector Machine (LSSVM). A conceptual strategy namely multi-station model was developed to forecast the Souris River discharge more accurately. The strategy used herein was capable of covering uncertainties and complexities of river discharge modeling. DWT and EEMD was coupled, and the feature selection was performed for decomposed sub-series using MI to be employed in multi-station model. In the proposed feature selection method, some useless sub-series were omitted to achieve better performance. Results approved efficiency of the proposed DWT-EEMD-MI approach to improve accuracy of multi-station modeling strategies.Keywords: river stage-discharge process, LSSVM, discrete wavelet transform, Ensemble Empirical Decomposition Mode, multi-station modeling
Procedia PDF Downloads 17513971 Relative Effectiveness of Inquiry: Approach and Expository Instructional Methods in Fostering Students’ Retention in Chemistry
Authors: Joy Johnbest Egbo
Abstract:
The study was designed to investigate the relative effectiveness of inquiry role approach and expository instructional methods in fostering students’ retention in chemistry. Two research questions were answered and three null hypotheses were formulated and tested at 0.05 level of significance. A quasi-experimental (the non-equivalent pretest, posttest control group) design was adopted for the study. The population for the study comprised all senior secondary school class two (SS II) students who were offering Chemistry in single sex schools in Enugu Education Zone. The instrument for data collection was a self-developed Chemistry Retention Test (CRT). Relevant data were collected from a sample of one hundred and forty–one (141) students drawn from two secondary schools (1 male and 1 female schools) using simple random sampling technique. A reliability co-efficient of 0.82 was obtained for the instrument using Kuder Richardson formular20 (K-R20). Mean and Standard deviation scores were used to answer the research questions while two–way analysis of covariance (ANCOVA) was used to test the hypotheses. The findings showed that the students taught with Inquiry role approach retained the chemistry concept significantly higher than their counterparts taught with expository method. Female students retained slightly higher than their male counterparts. There is significant interaction between instructional packages and gender on Chemistry students’ retention. It was recommended, among others, that teachers should be encouraged to employ the use of Inquiry-role approach more in the teaching of chemistry and other subjects in general. By so doing, students’ retention of the subject could be increased.Keywords: inquiry role approach, retention, exposition method, chemistry
Procedia PDF Downloads 51313970 Copula-Based Estimation of Direct and Indirect Effects in Path Analysis Model
Authors: Alam Ali, Ashok Kumar Pathak
Abstract:
Path analysis is a statistical technique used to evaluate the strength of the direct and indirect effects of variables. One or more structural regression equations are used to estimate a series of parameters in order to find the better fit of data. Sometimes, exogenous variables do not show a significant strength of their direct and indirect effect when the assumption of classical regression (ordinary least squares (OLS)) are violated by the nature of the data. The main motive of this article is to investigate the efficacy of the copula-based regression approach over the classical regression approach and calculate the direct and indirect effects of variables when data violates the OLS assumption and variables are linked through an elliptical copula. We perform this study using a well-organized numerical scheme. Finally, a real data application is also presented to demonstrate the performance of the superiority of the copula approach.Keywords: path analysis, copula-based regression models, direct and indirect effects, k-fold cross validation technique
Procedia PDF Downloads 7213969 Classification Earthquake Distribution in the Banda Sea Collision Zone with Point Process Approach
Authors: H. J. Wattimanela, U. S. Passaribu, N. T. Puspito, S. W. Indratno
Abstract:
Banda Sea collision zone (BSCZ) of is the result of the interaction and convergence of Indo-Australian plate, Eurasian plate and Pacific plate. This location in the eastern part of Indonesia. This zone has a very high seismic activity. In this research, we will be calculated rate (λ) and Mean Square Eror (MSE). By this result, we will identification of Poisson distribution of earthquakes in the BSCZ with the point process approach. Chi-square test approach and test Anscombe made in the process of identifying a Poisson distribution in the partition area. The data used are earthquakes with Magnitude ≥ 6 SR and its period 1964-2013 and sourced from BMKG Jakarta. This research is expected to contribute to the Moluccas Province and surrounding local governments in performing spatial plan document related to disaster management.Keywords: molluca banda sea collision zone, earthquakes, mean square error, poisson distribution, chi-square test, anscombe test
Procedia PDF Downloads 30013968 Laser Registration and Supervisory Control of neuroArm Robotic Surgical System
Authors: Hamidreza Hoshyarmanesh, Hosein Madieh, Sanju Lama, Yaser Maddahi, Garnette R. Sutherland, Kourosh Zareinia
Abstract:
This paper illustrates the concept of an algorithm to register specified markers on the neuroArm surgical manipulators, an image-guided MR-compatible tele-operated robot for microsurgery and stereotaxy. Two range-finding algorithms, namely time-of-flight and phase-shift, are evaluated for registration and supervisory control. The time-of-flight approach is implemented in a semi-field experiment to determine the precise position of a tiny retro-reflective moving object. The moving object simulates a surgical tool tip. The tool is a target that would be connected to the neuroArm end-effector during surgery inside the magnet bore of the MR imaging system. In order to apply flight approach, a 905-nm pulsed laser diode and an avalanche photodiode are utilized as the transmitter and receiver, respectively. For the experiment, a high frequency time to digital converter was designed using a field-programmable gate arrays. In the phase-shift approach, a continuous green laser beam with a wavelength of 530 nm was used as the transmitter. Results showed that a positioning error of 0.1 mm occurred when the scanner-target point distance was set in the range of 2.5 to 3 meters. The effectiveness of this non-contact approach exhibited that the method could be employed as an alternative for conventional mechanical registration arm. Furthermore, the approach is not limited by physical contact and extension of joint angles.Keywords: 3D laser scanner, intraoperative MR imaging, neuroArm, real time registration, robot-assisted surgery, supervisory control
Procedia PDF Downloads 28613967 Anatomically-Based Oropharyngeal Rehabilitation for the Patients with Obstructive Sleep Apnea Using a Multilevel Approach
Authors: Hsin-Yu Lin, Ching-Hsia Hung
Abstract:
Obstructive sleep apnea (OSA) is characterized by a complete or partial obstruction of the upper airway during sleep. The vulnerable sites of upper airway collapses are consequences of sleep state-dependent reductions in tone in specific pharyngeal dilators. Clinical examinations reveal multilevel collapses of the upper airway among the patients with OSA. Therefore, an anatomically-based oropharyngeal rehabilitation should comprise a multilevel approach, including retropalatal, retroglossal, hypopharyngeal, temporomandibular, and facial levels, all of which involve different muscle groups and contribute to multifunctional interaction and coordination, such as swallowing, breathing, and phonation. The purpose of the study was to exam the effects of this rehabilitation program with a multilevel approach. In this study, fifteen subjects with newly diagnosed moderate or severe OSA (Apnea-Hypopnea-Index≥15) were randomized into an intervention group and control group. The intervention group (N=8) underwent a 12-week-intervention of a hospital-based rehabilitation program, while the control group (N=7) was kept on the waiting list. The 12-week-intervention comprised an anatomically based multilevel approach. The primary outcome was Polysomnography (PSG) data, and the secondary outcome was oropharyngeal and respiratory muscle function. In the intervention group, Apnea-Hypopnea-Index significantly improved (46.96±19.45 versus 32.78±10.78 events/h, p=0.017) compared with control group (35.77±17.49 versus 42.96±17.32 events/h, p=0.043). While the control group remained no change, the intervention group demonstrated other PSG outcomes significantly improvement, including arousal index (46.04±18.9 versus 32.98±8.35/h, p=0.035), mean SpO2 (92.88±2.1 versus 94.13±1.46%, p=0.039). Besides, the intervention group demonstrated significant improvement in oropharyngeal and respiratory muscle function compared to the control group. This anatomically-based oropharyngeal rehabilitation with a multilevel approach can be proven as a non-invasive therapy for patients with OSA.Keywords: obstructive sleep apnea, upper airway, oropharyngeal rehabilitation, multilevel approach
Procedia PDF Downloads 10213966 New Approach for Constructing a Secure Biometric Database
Authors: A. Kebbeb, M. Mostefai, F. Benmerzoug, Y. Chahir
Abstract:
The multimodal biometric identification is the combination of several biometric systems. The challenge of this combination is to reduce some limitations of systems based on a single modality while significantly improving performance. In this paper, we propose a new approach to the construction and the protection of a multimodal biometric database dedicated to an identification system. We use a topological watermarking to hide the relation between face image and the registered descriptors extracted from other modalities of the same person for more secure user identification.Keywords: biometric databases, multimodal biometrics, security authentication, digital watermarking
Procedia PDF Downloads 39113965 Route Planning for Optimization Approach PSO_GA Sharing System (Scooter Sharing-Public Transportation) with Hybrid Optimization Approach PSO_GA
Authors: Mohammad Ali Farrokhpour
Abstract:
In the current decade and sustainable transportation systems, scooter sharing has attracted widespread attention as an environmentally-friendly means of public transportation which can help develop public transportation. The combination of scooters and subway in the area of sustainable transportation systems can provide a great many opportunities for developing access to public transportation. Of the challenges which have arisen and initiated discussions of interest about the implementation of a scooter-subway system to replace personal vehicles is the issue of routing in the aforementioned system. This has been chosen as the main subject of the present paper. Thus, the present paper provides an account for routing in this system. Because the issue of routing includes multiple factors such as time, costs, traffic, green spaces, etc., the above-mentioned problem is considered to be a multi-objective NP-hard optimization problem. For this purpose, the hybrid optimization approach of PSO-GA has been put forward in the present paper for the provided answers to be of higher accuracy and validity than those of normal optimization methods. The results obtained from modeling and problem solving for the case study in the MATLAB software are indicative of the efficiency and desirability of the model and the proposed approach for solving the modelKeywords: route planning, scooter sharing, public transportation, sharing system
Procedia PDF Downloads 8413964 Stacking Ensemble Approach for Combining Different Methods in Real Estate Prediction
Authors: Sol Girouard, Zona Kostic
Abstract:
A home is often the largest and most expensive purchase a person makes. Whether the decision leads to a successful outcome will be determined by a combination of critical factors. In this paper, we propose a method that efficiently handles all the factors in residential real estate and performs predictions given a feature space with high dimensionality while controlling for overfitting. The proposed method was built on gradient descent and boosting algorithms and uses a mixed optimizing technique to improve the prediction power. Usually, a single model cannot handle all the cases thus our approach builds multiple models based on different subsets of the predictors. The algorithm was tested on 3 million homes across the U.S., and the experimental results demonstrate the efficiency of this approach by outperforming techniques currently used in forecasting prices. With everyday changes on the real estate market, our proposed algorithm capitalizes from new events allowing more efficient predictions.Keywords: real estate prediction, gradient descent, boosting, ensemble methods, active learning, training
Procedia PDF Downloads 277