Search results for: spectral domain
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2467

Search results for: spectral domain

967 Drought Detection and Water Stress Impact on Vegetation Cover Sustainability Using Radar Data

Authors: E. Farg, M. M. El-Sharkawy, M. S. Mostafa, S. M. Arafat

Abstract:

Mapping water stress provides important baseline data for sustainable agriculture. Recent developments in the new Sentinel-1 data which allow the acquisition of high resolution images and varied polarization capabilities. This study was conducted to detect and quantify vegetation water content from canopy backscatter for extracting spatial information to encourage drought mapping activities throughout new reclaimed sandy soils in western Nile delta, Egypt. The performance of radar imagery in agriculture strongly depends on the sensor polarization capability. The dual mode capabilities of Sentinel-1 improve the ability to detect water stress and the backscatter from the structure components improves the identification and separation of vegetation types with various canopy structures from other features. The fieldwork data allowed identifying of water stress zones based on land cover structure; those classes were used for producing harmonious water stress map. The used analysis techniques and results show high capability of active sensors data in water stress mapping and monitoring especially when integrated with multi-spectral medium resolution images. Also sub soil drip irrigation systems cropped areas have lower drought and water stress than center pivot sprinkler irrigation systems. That refers to high level of evaporation from soil surface in initial growth stages. Results show that high relationship between vegetation indices such as Normalized Difference Vegetation Index NDVI the observed radar backscattering. In addition to observational evidence showed that the radar backscatter is highly sensitive to vegetation water stress, and essentially potential to monitor and detect vegetative cover drought.

Keywords: canopy backscatter, drought, polarization, NDVI

Procedia PDF Downloads 145
966 Multiple Linear Regression for Rapid Estimation of Subsurface Resistivity from Apparent Resistivity Measurements

Authors: Sabiu Bala Muhammad, Rosli Saad

Abstract:

Multiple linear regression (MLR) models for fast estimation of true subsurface resistivity from apparent resistivity field measurements are developed and assessed in this study. The parameters investigated were apparent resistivity (ρₐ), horizontal location (X) and depth (Z) of measurement as the independent variables; and true resistivity (ρₜ) as the dependent variable. To achieve linearity in both resistivity variables, datasets were first transformed into logarithmic domain following diagnostic checks of normality of the dependent variable and heteroscedasticity to ensure accurate models. Four MLR models were developed based on hierarchical combination of the independent variables. The generated MLR coefficients were applied to another data set to estimate ρₜ values for validation. Contours of the estimated ρₜ values were plotted and compared to the observed data plots at the colour scale and blanking for visual assessment. The accuracy of the models was assessed using coefficient of determination (R²), standard error (SE) and weighted mean absolute percentage error (wMAPE). It is concluded that the MLR models can estimate ρₜ for with high level of accuracy.

Keywords: apparent resistivity, depth, horizontal location, multiple linear regression, true resistivity

Procedia PDF Downloads 276
965 End-to-End Pyramid Based Method for Magnetic Resonance Imaging Reconstruction

Authors: Omer Cahana, Ofer Levi, Maya Herman

Abstract:

Magnetic Resonance Imaging (MRI) is a lengthy medical scan that stems from a long acquisition time. Its length is mainly due to the traditional sampling theorem, which defines a lower boundary for sampling. However, it is still possible to accelerate the scan by using a different approach such as Compress Sensing (CS) or Parallel Imaging (PI). These two complementary methods can be combined to achieve a faster scan with high-fidelity imaging. To achieve that, two conditions must be satisfied: i) the signal must be sparse under a known transform domain, and ii) the sampling method must be incoherent. In addition, a nonlinear reconstruction algorithm must be applied to recover the signal. While the rapid advances in Deep Learning (DL) have had tremendous successes in various computer vision tasks, the field of MRI reconstruction is still in its early stages. In this paper, we present an end-to-end method for MRI reconstruction from k-space to image. Our method contains two parts. The first is sensitivity map estimation (SME), which is a small yet effective network that can easily be extended to a variable number of coils. The second is reconstruction, which is a top-down architecture with lateral connections developed for building high-level refinement at all scales. Our method holds the state-of-art fastMRI benchmark, which is the largest, most diverse benchmark for MRI reconstruction.

Keywords: magnetic resonance imaging, image reconstruction, pyramid network, deep learning

Procedia PDF Downloads 91
964 Equilibrium, Kinetics, and Thermodynamic Studies on Heavy Metal Biosorption by Trichoderma Species

Authors: Sobia Mushtaq, Firdaus E. Bareen, Asma Tayyeb

Abstract:

This study conducted to investigate the metal biosorption potential of indigenous Trichoderma species (T. harzianum KS05T01, T. longibrachiatum KS09T03, Trichoderma sp KS17T09., T. viridi KS17T011, T. atrobruneo KS21T014, and T. citrinoviride) that have been isolated from contaminated soil of Kasur Tannery Waste Management Agency. The effect of different biosorption parameters as initial metal ion concentration, pH, contact time , and temperature of incubation was investigated on the biosorption potential of these species. The metal removal efficiency and (E%) and metal uptake capacity (mg/g) increased along with the increase of initial metal concentration in media. The Trichoderma species can tolerate and survive under heavy metal stress up to 800mg/L. Among the two isotherm models were applied on the biosorption data, Langmuir isotherm model and Freundlich isotherm model, maximum correlation coefficients values (R 2 ) of 1was found for Langmuir model, which showed the better fitted model for the Trichoderma biosorption. The metal biosorption was increased with the increase of temperature and pH of the media. The maximum biosorption was observed between 25-30 o C and at pH 6.-7.5, while the biosorption rate was increased from 3-6 days of incubation, and then the rate of biosorption was slowed down. The biosorption data was better fitted for Pseudo kinetic first order during the initial days of biosorption. Thermodynamic parameters as standard Gibbs free energy (G), standard enthalpy change (H), and standard entropy (S) were calculated. The results confirmed the heavy metal biosorption by Trichoderma species was endothermic and spontaneous reaction in nature. The FTIR spectral analysis and SEM-EDX analysis of the treated and controlled mycelium revealed the changes in the active functional sites and morphological variations of the outer surface. The data analysis envisaged that high metal tolerance exhibited by Trichoderma species indicates its potential as efficacious and successful mediator for bioremediation of the heavy metal polluted environments.

Keywords: heavy metal, fungal biomass, biosorption, kinetics

Procedia PDF Downloads 122
963 Enhancing the Sensitivity of Antigen Based Sandwich ELISA for COVID-19 Diagnosis in Saliva Using Gold Conjugated Nanobodies

Authors: Manal Kamel, Sara Maher

Abstract:

Development of sensitive non-invasive tests for detection of SARS-CoV-2 antigens is imperative to manage the extent of infection throughout the population, yet, it is still challenging. Here, we designed and optimized a sandwich enzyme-linked immunosorbent assay (ELISA) for SARS-CoV-2 S1 antigen detection in saliva. Both saliva samples and nasopharyngeal swapswere collected from 170 PCR-confirmed positive and negative cases. Gold nanoparticles (AuNPs) were conjugated with S1protein receptor binding domain (RBD) nanobodies. Recombinant S1 monoclonal antibodies (S1mAb) as primery antibody and gold conjugated nanobodies as secondary antibody were employed in sandwich ELISA. Our developed system were optimized to achieve 87.5 % sensitivity and 100% specificity for saliva samples compared to 89 % and 100% for nasopharyngeal swaps, respectively. This means that saliva could be a suitable replacement for nasopharyngeal swaps No cross reaction was detected with other corona virus antigens. These results revealed that our developed ELISAcould be establishedas a new, reliable, sensitive, and non-invasive test for diagnosis of SARS-CoV-2 infection, using the easily collected saliva samples.

Keywords: COVID 19, diagnosis, ELISA, nanobodies

Procedia PDF Downloads 134
962 An Automated System for the Detection of Citrus Greening Disease Based on Visual Descriptors

Authors: Sidra Naeem, Ayesha Naeem, Sahar Rahim, Nadia Nawaz Qadri

Abstract:

Citrus greening is a bacterial disease that causes considerable damage to citrus fruits worldwide. Efficient method for this disease detection must be carried out to minimize the production loss. This paper presents a pattern recognition system that comprises three stages for the detection of citrus greening from Orange leaves: segmentation, feature extraction and classification. Image segmentation is accomplished by adaptive thresholding. The feature extraction stage comprises of three visual descriptors i.e. shape, color and texture. From shape feature we have used asymmetry index, from color feature we have used histogram of Cb component from YCbCr domain and from texture feature we have used local binary pattern. Classification was done using support vector machines and k nearest neighbors. The best performances of the system is Accuracy = 88.02% and AUROC = 90.1% was achieved by automatic segmented images. Our experiments validate that: (1). Segmentation is an imperative preprocessing step for computer assisted diagnosis of citrus greening, and (2). The combination of shape, color and texture features form a complementary set towards the identification of citrus greening disease.

Keywords: citrus greening, pattern recognition, feature extraction, classification

Procedia PDF Downloads 184
961 Component Interface Formalization in Robotic Systems

Authors: Anton Hristozov, Eric Matson, Eric Dietz, Marcus Rogers

Abstract:

Components are heavily used in many software systems, including robotics systems. The growth of sophistication and diversity of new capabilities for robotic systems presents new challenges to their architectures. Their complexity is growing exponentially with the advent of AI, smart sensors, and the complex tasks they have to accomplish. Such complexity requires a more rigorous approach to the creation, use, and interoperability of software components. The issue is exacerbated because robotic systems are becoming more and more reliant on third-party components for certain functions. In order to achieve this kind of interoperability, including dynamic component replacement, we need a way to standardize their interfaces. A formal approach is desperately needed to specify what an interface of a robotic software component should contain. This study performs an analysis of the issue and presents a universal and generic approach to standardizing component interfaces for robotic systems. Our approach is inspired by well-established robotic architectures such as ROS, PX4, and Ardupilot. The study is also applicable to other software systems that share similar characteristics with robotic systems. We consider the use of JSON or Domain Specific Languages (DSL) development with tools such as Antlr and automatic code and configuration file generation for frameworks such as ROS and PX4. A case study with ROS2 is presented as a proof of concept for the proposed methodology.

Keywords: CPS, robots, software architecture, interface, ROS, autopilot

Procedia PDF Downloads 92
960 Cloud-Based Multiresolution Geodata Cube for Efficient Raster Data Visualization and Analysis

Authors: Lassi Lehto, Jaakko Kahkonen, Juha Oksanen, Tapani Sarjakoski

Abstract:

The use of raster-formatted data sets in geospatial analysis is increasing rapidly. At the same time, geographic data are being introduced into disciplines outside the traditional domain of geoinformatics, like climate change, intelligent transport, and immigration studies. These developments call for better methods to deliver raster geodata in an efficient and easy-to-use manner. Data cube technologies have traditionally been used in the geospatial domain for managing Earth Observation data sets that have strict requirements for effective handling of time series. The same approach and methodologies can also be applied in managing other types of geospatial data sets. A cloud service-based geodata cube, called GeoCubes Finland, has been developed to support online delivery and analysis of most important geospatial data sets with national coverage. The main target group of the service is the academic research institutes in the country. The most significant aspects of the GeoCubes data repository include the use of multiple resolution levels, cloud-optimized file structure, and a customized, flexible content access API. Input data sets are pre-processed while being ingested into the repository to bring them into a harmonized form in aspects like georeferencing, sampling resolutions, spatial subdivision, and value encoding. All the resolution levels are created using an appropriate generalization method, selected depending on the nature of the source data set. Multiple pre-processed resolutions enable new kinds of online analysis approaches to be introduced. Analysis processes based on interactive visual exploration can be effectively carried out, as the level of resolution most close to the visual scale can always be used. In the same way, statistical analysis can be carried out on resolution levels that best reflect the scale of the phenomenon being studied. Access times remain close to constant, independent of the scale applied in the application. The cloud service-based approach, applied in the GeoCubes Finland repository, enables analysis operations to be performed on the server platform, thus making high-performance computing facilities easily accessible. The developed GeoCubes API supports this kind of approach for online analysis. The use of cloud-optimized file structures in data storage enables the fast extraction of subareas. The access API allows for the use of vector-formatted administrative areas and user-defined polygons as definitions of subareas for data retrieval. Administrative areas of the country in four levels are available readily from the GeoCubes platform. In addition to direct delivery of raster data, the service also supports the so-called virtual file format, in which only a small text file is first downloaded. The text file contains links to the raster content on the service platform. The actual raster data is downloaded on demand, from the spatial area and resolution level required in each stage of the application. By the geodata cube approach, pre-harmonized geospatial data sets are made accessible to new categories of inexperienced users in an easy-to-use manner. At the same time, the multiresolution nature of the GeoCubes repository facilitates expert users to introduce new kinds of interactive online analysis operations.

Keywords: cloud service, geodata cube, multiresolution, raster geodata

Procedia PDF Downloads 135
959 F-VarNet: Fast Variational Network for MRI Reconstruction

Authors: Omer Cahana, Maya Herman, Ofer Levi

Abstract:

Magnetic resonance imaging (MRI) is a long medical scan that stems from a long acquisition time. This length is mainly due to the traditional sampling theorem, which defines a lower boundary for sampling. However, it is still possible to accelerate the scan by using a different approach, such as compress sensing (CS) or parallel imaging (PI). These two complementary methods can be combined to achieve a faster scan with high-fidelity imaging. In order to achieve that, two properties have to exist: i) the signal must be sparse under a known transform domain, ii) the sampling method must be incoherent. In addition, a nonlinear reconstruction algorithm needs to be applied to recover the signal. While the rapid advance in the deep learning (DL) field, which has demonstrated tremendous successes in various computer vision task’s, the field of MRI reconstruction is still in an early stage. In this paper, we present an extension of the state-of-the-art model in MRI reconstruction -VarNet. We utilize VarNet by using dilated convolution in different scales, which extends the receptive field to capture more contextual information. Moreover, we simplified the sensitivity map estimation (SME), for it holds many unnecessary layers for this task. Those improvements have shown significant decreases in computation costs as well as higher accuracy.

Keywords: MRI, deep learning, variational network, computer vision, compress sensing

Procedia PDF Downloads 162
958 Effect of Plasma Radiation on Keratinocyte Cells Involved in the Wound Healing Process

Authors: B. Fazekas, I. Korolov, K. Kutasi

Abstract:

Plasma medicine, which involves the use of gas discharge plasmas for medical applications is a rapidly growing research field. The use of non-thermal atmospheric pressure plasmas in dermatology to assist tissue regeneration by improving the healing of infected and/or chronic wounds is a promising application. It is believed that plasma can activate cells, which are involved in the wound closure. Non-thermal atmospheric plasmas are rich in chemically active species (such as O and N-atoms, O2(a) molecules) and radiative species such as the NO, N2+ and N2 excited molecules, which dominantly radiate in the 200-500 nm spectral range. In order to understand the effect of plasma species, both of chemically active and radiative species on wound healing process, the interaction of physical plasma with the human skin cells is necessary. In order to clarify the effect of plasma radiation on the wound healing process we treated keratinocyte cells – that are one of the main cell types in human skin epidermis – covered with a layer of phosphate-buffered saline (PBS) with a low power atmospheric pressure plasma. For the generation of such plasma we have applied a plasma needle. Here, the plasma is ignited at the tip of the needle in flowing helium gas in contact with the ambient air. To study the effect of plasma radiation we used a plasma needle configuration, where the plasma species – chemically active radicals and charged species – could not reach the treated cells, but only the radiation. For the comparison purposes, we also irradiated the cells using a UV-B light source (FS20 lamp) with a 20 and 40 mJ cm-2 dose of 312 nm. After treatment the viability and the proliferation of the cells have been examined. The proliferation of cells has been studied with a real time monitoring system called Xcelligence. The results have indicated, that the 20 mJ cm-2 dose did not affect cell viability, whereas the 40 mJ cm-2 dose resulted a decrease in cell viability. The results have shown that the plasma radiation have no quantifiable effect on the cell proliferation as compared to the non-treated cells.

Keywords: UV radiation, non-equilibrium gas discharges (non-thermal plasmas), plasma emission, keratinocyte cells

Procedia PDF Downloads 602
957 A Probabilistic Theory of the Buy-Low and Sell-High for Algorithmic Trading

Authors: Peter Shi

Abstract:

Algorithmic trading is a rapidly expanding domain within quantitative finance, constituting a substantial portion of trading volumes in the US financial market. The demand for rigorous and robust mathematical theories underpinning these trading algorithms is ever-growing. In this study, the author establishes a new stock market model that integrates the Efficient Market Hypothesis and the statistical arbitrage. The model, for the first time, finds probabilistic relations between the rational price and the market price in terms of the conditional expectation. The theory consequently leads to a mathematical justification of the old market adage: buy-low and sell-high. The thresholds for “low” and “high” are precisely derived using a max-min operation on Bayes’s error. This explicit connection harmonizes the Efficient Market Hypothesis and Statistical Arbitrage, demonstrating their compatibility in explaining market dynamics. The amalgamation represents a pioneering contribution to quantitative finance. The study culminates in comprehensive numerical tests using historical market data, affirming that the “buy-low” and “sell-high” algorithm derived from this theory significantly outperforms the general market over the long term in four out of six distinct market environments.

Keywords: efficient market hypothesis, behavioral finance, Bayes' decision, algorithmic trading, risk control, stock market

Procedia PDF Downloads 72
956 An Optimal Control Model to Determine Body Forces of Stokes Flow

Authors: Yuanhao Gao, Pin Lin, Kees Weijer

Abstract:

In this paper, we will determine the external body force distribution with analysis of stokes fluid motion using mathematical modelling and numerical approaching. The body force distribution is regarded as the unknown variable and could be determined by the idea of optimal control theory. The Stokes flow motion and its velocity are generated by given forces in a unit square domain. A regularized objective functional is built to match the numerical result of flow velocity with the generated velocity data. So that the force distribution could be determined by minimizing the value of objective functional, which is also the difference between the numerical and experimental velocity. Then after utilizing the Lagrange multiplier method, some partial differential equations are formulated consisting the optimal control system to solve. Finite element method and conjugate gradient method are used to discretize equations and deduce the iterative expression of target body force to compute the velocity numerically and body force distribution. Programming environment FreeFEM++ supports the implementation of this model.

Keywords: optimal control model, Stokes equation, finite element method, conjugate gradient method

Procedia PDF Downloads 405
955 Study of the Mega–Landslide at the Community of Ropoto, Central Greece, and of the Design of Mitigation and Early Warning System Using the Fiber Bragg Grating Technology

Authors: Michael Bellas, George Voulgaridis

Abstract:

This paper refers to the world known mega - landslide induced at the community of Ropoto, belonging to the Municipality of Trikala, in the Central part of Greece. The landslide affected the debris as well as the colluvium mantle of the flysch, and makes up a special case of study in engineering geology and geotechnical engineering not only because of the size of the domain affected by the landslide (approximately 750m long), but also because of the geostructure’s global behavior. Due to the landslide, the whole community’s infrastructure massively collapsed and human lives were put in danger. After the complete simulation of the coupled Seepage - Deformation phenomenon due to the extreme rainfall, and by closely examining the slope’s global behavior, both the mitigation of the landslide, as well as, an advanced surveillance method (Fiber Bragg Grating) using fiber optics were further studied, in order both to retain the geostructure and to monitor its health by creating an early warning system, which would serve as a complete safety net for saving both the community’s infrastructure as well as the lives of its habitats.

Keywords: landslide, remediation measures, the finite element method (FEM), Fiber Bragg Grating (FBG) sensing method

Procedia PDF Downloads 329
954 Understanding the Scope of Architects in Disaster Risk Reduction: The Case of Bhuj

Authors: Sweta Kandari

Abstract:

Predominantly, the conventional role of an architect is to design and construct. However, in a post-disaster scenario, the prevalent role expands and includes many other responsibilities. Agencies collaborating in post-disaster reconstruction face the challenge of building back quickly while requiring them to listen, reflect, develop and deliver as per the needs and requirements of the people. The question of the role of an architect has been extensively discussed in the reconstruction field. Discourses about the role of an architect in post-disaster scenario revolve around the ignorance by the profession, their professional abilities and inabilities. Within this domain, this paper aims at analyzing and recognizing the roles, responsibilities, scope, limitations, skillsets applied and required by an architect while working in a post-disaster situation. Four projects rebuilt after the 2001 Bhuj earthquake in Gujarat, India were examined for this research. Based on the analysis of the case study, areas of intervention of an architect in the various stages of rebuilding were identified. It was reinforced that within the areas of intervention identified, there is a vast gap between the prescribed, the prevalent notion and the performed responsibilities of an architect. This paper brings forth the specific gaps in the rebuilding process while exploring and understanding the relationship between various stakeholders that influence the role of an architect.

Keywords: rebuilding, role of an architect, Bhuj, post-disaster

Procedia PDF Downloads 132
953 Design and Implementation of Medium Access Control Based Routing on Real Wireless Sensor Networks Testbed

Authors: Smriti Agarwal, Ashish Payal, B. V. R. Reddy

Abstract:

IEEE 802.15.4 is a Low Rate Wireless Personal Area Networks (LR-WPAN) standard combined with ZigBee, which is going to enable new applications in Wireless Sensor Networks (WSNs) and Internet of Things (IoT) domain. In recent years, it has become a popular standard for WSNs. Wireless communication among sensor motes, enabled by IEEE 802.15.4 standard, is extensively replacing the existing wired technology in a wide range of monitoring and control applications. Researchers have proposed a routing framework and mechanism that interacts with the IEEE 802.15.4 standard using software platform. In this paper, we have designed and implemented MAC based routing (MBR) based on IEEE 802.15.4 standard using a hardware platform “SENSEnuts”. The experimental results include data through light and temperature sensors obtained from communication between PAN coordinator and source node through coordinator, MAC address of some modules used in the experimental setup, topology of the network created for simulation and the remaining battery power of the source node. Our experimental effort on a WSN Testbed has helped us in bridging the gap between theoretical and practical aspect of implementing IEEE 802.15.4 for WSNs applications.

Keywords: IEEE 802.15.4, routing, WSN, ZigBee

Procedia PDF Downloads 406
952 Blockchain for IoT Security and Privacy in Healthcare Sector

Authors: Umair Shafique, Hafiz Usman Zia, Fiaz Majeed, Samina Naz, Javeria Ahmed, Maleeha Zainab

Abstract:

The Internet of Things (IoT) has become a hot topic for the last couple of years. This innovative technology has shown promising progress in various areas, and the world has witnessed exponential growth in multiple application domains. Researchers are working to investigate its aptitudes to get the best from it by harnessing its true potential. But at the same time, IoT networks open up a new aspect of vulnerability and physical threats to data integrity, privacy, and confidentiality. It's is due to centralized control, data silos approach for handling information, and a lack of standardization in the IoT networks. As we know, blockchain is a new technology that involves creating secure distributed ledgers to store and communicate data. Some of the benefits include resiliency, integrity, anonymity, decentralization, and autonomous control. The potential for blockchain technology to provide the key to managing and controlling IoT has created a new wave of excitement around the idea of putting that data back into the hands of the end-users. In this manuscript, we have proposed a model that combines blockchain and IoT networks to address potential security and privacy issues in the healthcare domain. Then we try to describe various application areas, challenges, and future directions in the healthcare sector where blockchain platforms merge with IoT networks.

Keywords: IoT, blockchain, cryptocurrency, healthcare, consensus, data

Procedia PDF Downloads 180
951 Limitation of Parallel Flow in Three-Dimensional Elongated Porous Domain Subjected to Cross Heat and Mass Flux

Authors: Najwa Mimouni, Omar Rahli, Rachid Bennacer, Salah Chikh

Abstract:

In the present work 2D and 3D numerical simulations of double diffusion natural convection in an elongated enclosure filled with a binary fluid saturating a porous medium are carried out. In the formulation of the problem, the Boussinesq approximation is considered and cross Neumann boundary conditions are specified for heat and mass walls conditions. The numerical method is based on the control volume approach with the third order QUICK scheme. Full approximation storage (FAS) with full multigrid (FMG) method is used to solve the problem. For the explored large range of the controlling parameters, we clearly evidenced that the increase in the depth of the cavity i.e. the lateral aspect ratio has an important effect on the flow patterns. The 2D perfect parallel flows obtained for a small lateral aspect ratio are drastically destabilized by increasing the cavity lateral dimension. This yields a 3D fluid motion with a much more complicated flow pattern and the classically studied 2D parallel flows are impossible.

Keywords: bifurcation, natural convection, heat and mass transfer, parallel flow, porous media

Procedia PDF Downloads 473
950 Speech Emotion Recognition: A DNN and LSTM Comparison in Single and Multiple Feature Application

Authors: Thiago Spilborghs Bueno Meyer, Plinio Thomaz Aquino Junior

Abstract:

Through speech, which privileges the functional and interactive nature of the text, it is possible to ascertain the spatiotemporal circumstances, the conditions of production and reception of the discourse, the explicit purposes such as informing, explaining, convincing, etc. These conditions allow bringing the interaction between humans closer to the human-robot interaction, making it natural and sensitive to information. However, it is not enough to understand what is said; it is necessary to recognize emotions for the desired interaction. The validity of the use of neural networks for feature selection and emotion recognition was verified. For this purpose, it is proposed the use of neural networks and comparison of models, such as recurrent neural networks and deep neural networks, in order to carry out the classification of emotions through speech signals to verify the quality of recognition. It is expected to enable the implementation of robots in a domestic environment, such as the HERA robot from the RoboFEI@Home team, which focuses on autonomous service robots for the domestic environment. Tests were performed using only the Mel-Frequency Cepstral Coefficients, as well as tests with several characteristics of Delta-MFCC, spectral contrast, and the Mel spectrogram. To carry out the training, validation and testing of the neural networks, the eNTERFACE’05 database was used, which has 42 speakers from 14 different nationalities speaking the English language. The data from the chosen database are videos that, for use in neural networks, were converted into audios. It was found as a result, a classification of 51,969% of correct answers when using the deep neural network, when the use of the recurrent neural network was verified, with the classification with accuracy equal to 44.09%. The results are more accurate when only the Mel-Frequency Cepstral Coefficients are used for the classification, using the classifier with the deep neural network, and in only one case, it is possible to observe a greater accuracy by the recurrent neural network, which occurs in the use of various features and setting 73 for batch size and 100 training epochs.

Keywords: emotion recognition, speech, deep learning, human-robot interaction, neural networks

Procedia PDF Downloads 170
949 Conceptual Model Providing More Information on the Contact Situation between Crime Victim and the Police

Authors: M. Inzunza

Abstract:

In contemporary society, victims of crime has been given more recognition, which have contributed to advancing the knowledge on the effects of crime. There exists a complexity of who gets the status of victim and that the typology of good versus bad can interfere with the contact situation of the victim with the police. The aim of this study is to identify the most central areas affecting the contact situation between crime victims and the police to develop a conceptual model to be useful empirically. By considering previously documented problem areas and different theoretical domains, a conceptual model has been developed. Preliminary findings suggest that an area that should be given attention is to get a better understanding of the victim, not only in terms of demographics but also in terms of risk behavior and social network. This area has been considered to influence the status of the crime victim. Another domain of value is the type of crime and the context of the incident in more detail. The police officer approach style in the contact situation is also a pertinent area that is influenced by how the police based victim services are organized and how individual police officers are suited for the mission. Suitability includes constructs from empathy models adapted to the police context and especially focusing on sub-constructs such as perspective taking. Discussion will focus on how these findings can be operationalized in practice and how they are used in ongoing empirical studies.

Keywords: empathy, perspective taking, police contact, victim of crime

Procedia PDF Downloads 137
948 Integrating Assurance and Risk Management of Complex Systems

Authors: Odd Ivar Haugen

Abstract:

This paper explores the relationship between assurance, risk, and risk management in the context of complex safety-related systems. It introduces a nuanced understanding of assurance and argues that the foundation for grounds for justified confidence in claims made about a complex system is related to the system behaviour. It emphasises the importance of knowledge as the cornerstone of assurance. The paper addresses the challenges of epistemic and aleatory uncertainties inherent in safety-critical systems. A systems approach is proposed to model emergent properties and complexity using the composition, environment, structure, mechanisms (CESM) metamodel, offering a structured framework for analysing system behaviour. The interplay between assurance and risk management is conceptualised through two models: the domain model and the control model. Assurance and risk management are mutually dependent on each other to reduce uncertainty and control risk levels. This work highlights the dual roles of assurance in risk management, acting as an epistemic actuator on the one side and providing feedback about the strength of the justification on the other. Assurance and risk management have inseparable roles in ensuring safety in complex systems.

Keywords: assurance, CESM metamodel, confidence, emergent properties, knowledge, objectivity, risk, system behaviour, system safety

Procedia PDF Downloads 4
947 Pulse Method for Investigation of Zr-C Phase Diagram at High Carbon Content Domain under High Temperatures

Authors: Arseniy M. Kondratyev, Sergey V. Onufriev, Alexander I. Savvatimskiy

Abstract:

The microsecond electrical pulse heating technique which provides uniform energy input into an investigated specimen is considered. In the present study we investigated ZrC+C carbide specimens in a form of a thin layer (about 5 microns thick) that were produced using a method of magnetron sputtering on insulating substrates. Specimens contained (at. %): Zr–17.88; C–67.69; N–8.13; O–5.98. Current through the specimen, voltage drop across it and radiation at the wavelength of 856 nm were recorded in the experiments. It enabled us to calculate the input energy, specific heat (from 2300 to 4500 K) and resistivity (referred to the initial dimensions of a specimen). To obtain the true temperature a black body specimen was used. Temperature of the beginning and completion of a phase transition (solid–liquid) was measured.Temperature of the onset of melting was 3150 K at the input energy 2.65 kJ/g; temperature of the completion of melting was 3450 K at the input energy 5.2 kJ/g. The specific heat of the solid phase of investigated carbide calculated using our data on temperature and imparted energy, is close to 0.75 J/gК for temperature range 2100–2800 K. Our results are considered together with the equilibrium Zr-C phase diagram.

Keywords: pulse heating, zirconium carbide, high temperatures, melting

Procedia PDF Downloads 323
946 Derivation of Bathymetry Data Using Worldview-2 Multispectral Images in Shallow, Turbid and Saline Lake Acıgöl

Authors: Muhittin Karaman, Murat Budakoglu

Abstract:

In this study, derivation of lake bathymetry was evaluated using the high resolution Worldview-2 multispectral images in the very shallow hypersaline Lake Acıgöl which does not have a stable water table due to the wet-dry season changes and industrial usage. Every year, a great part of the lake water budget has been consumed for the industrial salt production in the evaporation ponds, which are generally located on the south and north shores of Lake Acıgöl. Therefore, determination of the water level changes from a perspective of remote sensing-based lake water by bathymetry studies has a great importance in the sustainability-control of the lake. While the water table interval is around 1 meter between dry and wet season, dissolved ion concentration, salinity and turbidity also show clear differences during these two distinct seasonal periods. At the same time, with the satellite data acquisition (June 9, 2013), a field study was conducted to collect the salinity values, Secchi disk depths and turbidity levels. Max depth, Secchi disk depth and salinity were determined as 1,7 m, 0,9 m and 43,11 ppt, respectively. Eight-band Worldview-2 image was corrected for atmospheric effects by ATCOR technique. For each sampling point in the image, mean reflectance values in 1*1, 3*3, 5*5, 7*7, 9*9, 11*11, 13*13, 15*15, 17*17, 19*19, 21*21, 51*51 pixel reflectance neighborhoods were calculated separately. A unique image has been derivated for each matrix resolution. Spectral values and depth relation were evaluated for these distinct resolution images. Correlation coefficients were determined for the 1x1 matrix: 0,98, 0,96, 0,95 and 0,90 for the 724 nm, 831 nm, 908 nm and 659 nm, respectively. While 15x5 matrix characteristics with 0,98, 0,97 and 0,97 correlation values for the 724 nm, 908 nm and 831 nm, respectively; 51x51 matrix shows 0,98, 0,97 and 0,96 correlation values for the 724 nm, 831 nm and 659 nm, respectively. Comparison of all matrix resolutions indicates that RedEdge band (724 nm) of the Worldview-2 satellite image has the best correlation with the saline shallow lake of Acıgöl in-situ depth.

Keywords: bathymetry, Worldview-2 satellite image, ATCOR technique, Lake Acıgöl, Denizli, Turkey

Procedia PDF Downloads 447
945 Meditation-Based Interventions in the Workplace

Authors: Louise Fitzgerald, John Allman

Abstract:

Introduction: Having previously engaged in a meditation-based programme (MBP) for staff in general practice, we explore the evidence and extent to which MBPs are employed in the workplace. Aim of the study: We aim to understand the current workplace MBP intervention literature, which will help inform the suitability of these interventions within the workplace domain. Objectives: Uptake of MBPs in the workplace has grown as organizations look to support employee health, wellbeing, and performance. We will discuss the current MBP literature, including the large variability across MBPs and the associated difficulties in evaluating their efficacy. Learning points: 1) MBPs have a positive impact on cognitive function including concentration and memory and as such job performance. MBPs appear to have a positive impact on objective and subjective job satisfaction, productivity, motivation and work engagement. Meditation in the workplace may have positive impacts on mental health issues - including stress reduction and depression. 2) From our review MBPs appear to be implementable in a wide range of professions and work contexts - regardless of individual factors. Given many companies are focusing on health and wellbeing of employees, this could be included in employee wellbeing programmes. 3) Despite the benefits of mindfulness and meditation interventions in psychosocial workplace health and work performance the long-term efficacy has yet to be fully determined.

Keywords: meditation-based programmes, mindfulness, meditation, well-being

Procedia PDF Downloads 140
944 A Research Using Remote Monitoring Technology for Pump Output Monitoring in Distributed Fuel Stations in Nigeria

Authors: Ofoegbu Ositadinma Edward

Abstract:

This research paper discusses a web based monitoring system that enables effective monitoring of fuel pump output and sales volume from distributed fuel stations under the domain of a single company/organization. The traditional method of operation by these organizations in Nigeria is non-automated and accounting for dispensed product is usually approximated and manual as there is little or no technology implemented to presently provide information relating to the state of affairs in the station both to on-ground staff and to supervisory staff that are not physically present in the station. This results in unaccountable losses in product and revenue as well as slow decision making. Remote monitoring technology as a vast research field with numerous application areas incorporating various data collation techniques and sensor networks can be applied to provide information relating to fuel pump status in distributed fuel stations reliably. Thus, the proposed system relies upon a microcontroller, keypad and pump to demonstrate the traditional fuel dispenser. A web-enabled PC with an accompanying graphic user interface (GUI) was designed using virtual basic which is connected to the microcontroller via the serial port which is to provide the web implementation.

Keywords: fuel pump, microcontroller, GUI, web

Procedia PDF Downloads 434
943 Measuring Audit Quality Using Text Analysis: An Empirical Study of Indian Companies

Authors: Leesa Mohanty, Ashok Banerjee

Abstract:

Better audit quality signifies the financial statements of the auditee firm reflect true and fair view of their actual state of affairs, which reduces information asymmetry between management and shareholders, as a result, helps protect interests of shareholders. This study examines the impact of joint audit on audit quality. It is motivated by the ongoing debate where The Institute of Chartered Accountants of India (ICAI), the regulatory body governing auditors, has advocated the finance ministry and the Reserve Bank of India (RBI) for the mandatory use of joint audit in private banks to enhance the quality of audit. Earlier, the Government of India had rejected the plea by ICAI for mandatory joint audits in large companies stating it is not a viable option for promoting domestic firms. We introduce a new measure of audit quality. Drawing from the domain of text analytics, we use relevant phrases in audit reports to gauge audit quality and demonstrate that joint audit improves audit quality. We also, for robustness, use prevalent proxy for audit quality (Big N Auditor, ratio of audit fees to total fees) and find negative effect of joint audit on audit quality. We, therefore highlight that different proxy for audit quality show opposite effect of joint audit.

Keywords: audit fees, audit quality, Big N. Auditor, joint audit

Procedia PDF Downloads 357
942 One-Dimensional Performance Improvement of a Single-Stage Transonic Compressor

Authors: A. Shahsavari, M. Nili-Ahmadabadi

Abstract:

This paper presents an innovative one-dimensional optimization of a transonic compressor based on the radial equilibrium theory by means of increasing blade loading. Firstly, the rotor blade of the transonic compressor is redesigned based on the constant span-wise deHaller number and diffusion. The code is applied to extract compressor meridional plane and blade to blade geometry containing rotor and stator in order to design blade three-dimensional view. A structured grid is generated for the numerical domain of fluid. Finer grids are used for regions near walls to capture boundary layer effects and behavior. RANS equations are solved by finite volume method for rotating zones (rotor) and stationary zones (stator). The experimental data, available for the performance map of NASA Rotor67, is used to validate the results of simulations. Then, the capability of the design method is validated by CFD that is capable of predicting the performance map. The numerical results of new geometry show about 19% increase in pressure ratio and 11% improvement in overall efficiency of the transonic stage; however, the design point mass flow rate of the new compressor is 5.7% less than that of the original compressor.

Keywords: deHaller number, one dimensional design, radial equilibrium, transonic compressor

Procedia PDF Downloads 341
941 Design Guidelines for an Enhanced Interaction Experience in the Domain of Smartphone-Based Applications for Sport and Fitness

Authors: Paolo Pilloni, Fabrizio Mulas, Salvatore Carta

Abstract:

Nowadays, several research studies point up that an active lifestyle is essential for physical and mental health benefits. Mobile phones have greatly influenced people’s habits and attitudes also in the way they exercise. Our research work is mainly focused on investigating how to exploit mobile technologies to favour people’s exertion experience. To this end, we developed an exertion framework users can exploit through a real world mobile application, called BLINDED, designed to act as a virtual personal trainer to support runners during their trainings. In this work, inspired by both previous findings in the field of interaction design for people with visual impairments, feedback gathered from real users of our framework, and positive results obtained from two experimentations, we present some new interaction facilities we designed to enhance the interaction experience during a training. The positive obtained results helped us to derive some interaction design recommendations we believe will be a valid support for designers of future mobile systems conceived to be used in circumstances where there are limited possibilities of interaction.

Keywords: human computer interaction, interaction design guidelines, persuasive mobile technologies for sport and health

Procedia PDF Downloads 532
940 Empirical Green’s Function Technique for Accelerogram Synthesis: The Problem of the Use for Marine Seismic Hazard Assessment

Authors: Artem A. Krylov

Abstract:

Instrumental seismological researches in water areas are complicated and expensive, that leads to the lack of strong motion records in most offshore regions. In the same time the number of offshore industrial infrastructure objects, such as oil rigs, subsea pipelines, is constantly increasing. The empirical Green’s function technique proved to be very effective for accelerograms synthesis under the conditions of poorly described seismic wave propagation medium. But the selection of suitable small earthquake record in offshore regions as an empirical Green’s function is a problem because of short seafloor instrumental seismological investigation results usually with weak micro-earthquakes recordings. An approach based on moving average smoothing in the frequency domain is presented for preliminary processing of weak micro-earthquake records before using it as empirical Green’s function. The method results in significant waveform correction for modeled event. The case study for 2009 L’Aquila earthquake was used to demonstrate the suitability of the method. This work was supported by the Russian Foundation of Basic Research (project № 18-35-00474 mol_a).

Keywords: accelerogram synthesis, empirical Green's function, marine seismology, microearthquakes

Procedia PDF Downloads 324
939 Using of the Fractal Dimensions for the Analysis of Hyperkinetic Movements in the Parkinson's Disease

Authors: Sadegh Marzban, Mohamad Sobhan Sheikh Andalibi, Farnaz Ghassemi, Farzad Towhidkhah

Abstract:

Parkinson's disease (PD), which is characterized by the tremor at rest, rigidity, akinesia or bradykinesia and postural instability, affects the quality of life of involved individuals. The concept of a fractal is most often associated with irregular geometric objects that display self-similarity. Fractal dimension (FD) can be used to quantify the complexity and the self-similarity of an object such as tremor. In this work, we are aimed to propose a new method for evaluating hyperkinetic movements such as tremor, by using the FD and other correlated parameters in patients who are suffered from PD. In this study, we used 'the tremor data of Physionet'. The database consists of fourteen participants, diagnosed with PD including six patients with high amplitude tremor and eight patients with low amplitude. We tried to extract features from data, which can distinguish between patients before and after medication. We have selected fractal dimensions, including correlation dimension, box dimension, and information dimension. Lilliefors test has been used for normality test. Paired t-test or Wilcoxon signed rank test were also done to find differences between patients before and after medication, depending on whether the normality is detected or not. In addition, two-way ANOVA was used to investigate the possible association between the therapeutic effects and features extracted from the tremor. Just one of the extracted features showed significant differences between patients before and after medication. According to the results, correlation dimension was significantly different before and after the patient's medication (p=0.009). Also, two-way ANOVA demonstrates significant differences just in medication effect (p=0.033), and no significant differences were found between subject's differences (p=0.34) and interaction (p=0.97). The most striking result emerged from the data is that correlation dimension could quantify medication treatment based on tremor. This study has provided a technique to evaluate a non-linear measure for quantifying medication, nominally the correlation dimension. Furthermore, this study supports the idea that fractal dimension analysis yields additional information compared with conventional spectral measures in the detection of poor prognosis patients.

Keywords: correlation dimension, non-linear measure, Parkinson’s disease, tremor

Procedia PDF Downloads 244
938 Clustering of Association Rules of ISIS & Al-Qaeda Based on Similarity Measures

Authors: Tamanna Goyal, Divya Bansal, Sanjeev Sofat

Abstract:

In world-threatening terrorist attacks, where early detection, distinction, and prediction are effective diagnosis techniques and for functionally accurate and precise analysis of terrorism data, there are so many data mining & statistical approaches to assure accuracy. The computational extraction of derived patterns is a non-trivial task which comprises specific domain discovery by means of sophisticated algorithm design and analysis. This paper proposes an approach for similarity extraction by obtaining the useful attributes from the available datasets of terrorist attacks and then applying feature selection technique based on the statistical impurity measures followed by clustering techniques on the basis of similarity measures. On the basis of degree of participation of attributes in the rules, the associative dependencies between the attacks are analyzed. Consequently, to compute the similarity among the discovered rules, we applied a weighted similarity measure. Finally, the rules are grouped by applying using hierarchical clustering. We have applied it to an open source dataset to determine the usability and efficiency of our technique, and a literature search is also accomplished to support the efficiency and accuracy of our results.

Keywords: association rules, clustering, similarity measure, statistical approaches

Procedia PDF Downloads 320