Search results for: space time cube
18783 Microwave Imaging by Application of Information Theory Criteria in MUSIC Algorithm
Authors: Majid Pourahmadi
Abstract:
The performance of time-reversal MUSIC algorithm will be dramatically degrades in presence of strong noise and multiple scattering (i.e. when scatterers are close to each other). This is due to error in determining the number of scatterers. The present paper provides a new approach to alleviate such a problem using an information theoretic criterion referred as minimum description length (MDL). The merits of the novel approach are confirmed by the numerical examples. The results indicate the time-reversal MUSIC yields accurate estimate of the target locations with considerable noise and multiple scattering in the received signals.Keywords: microwave imaging, time reversal, MUSIC algorithm, minimum description length (MDL)
Procedia PDF Downloads 33718782 The Language of Risk: Pregnancy and Childbirth in the COVID-19 Era
Authors: Sarah Holdren, Laura Crook, Anne Drapkin Lyerly
Abstract:
Objective: The COVID-19 Pandemic has drawn new attention to long-existing bioethical questions around pregnancy, childbirth, and parenthood. Due to the increased risk of severe COVID-19, pregnant individuals may experience anxiety regarding medical decision-making. Especially in the case of hospital births, questions around the ethics of bringing healthy pregnant individuals into a high-risk environment for viral transmission illuminate gaps in the American maternal and child healthcare system. Limited research has sought to understand the experiences of those who gave birth outside hospitals during this time. This study aims to understand pregnant individuals’ conceptualization of risk during the COVID-19 pandemic. Methods: Individuals who gave birth after March 2020 were recruited through advertisements on social media. Participants completed a 1-hour semi-structured interview and a demographic questionnaire. Interviews were transcribed and coded by members of the research team using thematic narrative analysis. Results: A total of 18 participants were interviewed and completed the demographic questionnaire. The language of risk was utilized in birth narratives in three different ways, which highlighted the multileveled and nuanced ways in which risk is understood and mitigated by pregnant and birthing individuals. These included: 1. The risk of contracting COVID-19 before, during, and after birth, 2. The risk of birth complications requiring medical interventions dependent on selected birthing space (home, birthing center, hospital), and 3. The overall risk of creating life in the middle of a pandemic. The risk of contracting COVID-19 and risk of birth complications were often weighed in paradoxical ways throughout each individual’s pregnancy, while phrases such as “pandemic baby” and “apocalypse” appeared throughout narratives and highlighted the broader implications of pregnancy and childbirth during this momentous time. Conclusions: Healthcare professionals should consider the variety of ways that pregnant and birthing individuals understand the risk when counseling patients on healthcare decisions, especially during times of healthcare crisis such as COVID-19. Future work should look to understand how the language of risk fits into a broader understanding of the human experience of growing life in times of crisis.Keywords: maternal and child health, thematic narrative analysis, COVID-19, risk mitigation
Procedia PDF Downloads 16718781 Discrete Tracking Control of Nonholonomic Mobile Robots: Backstepping Design Approach
Authors: Alexander S. Andreev, Olga A. Peregudova
Abstract:
In this paper, we propose a discrete tracking control of nonholonomic mobile robots with two degrees of freedom. The electro-mechanical model of a mobile robot moving on a horizontal surface without slipping, with two rear wheels controlled by two independent DC electric, and one front roal wheel is considered. We present back-stepping design based on the Euler approximate discrete-time model of a continuous-time plant. Theoretical considerations are verified by numerical simulation. The work was supported by RFFI (15-01-08482).Keywords: actuator dynamics, back stepping, discrete-time controller, Lyapunov function, wheeled mobile robot
Procedia PDF Downloads 41518780 Research on the Strategy of Orbital Avoidance for Optical Remote Sensing Satellite
Authors: Zheng DianXun, Cheng Bo, Lin Hetong
Abstract:
This paper focuses on the orbit avoidance strategies of optical remote sensing satellite. The optical remote sensing satellite, moving along the Sun-synchronous orbit, is equipped with laser warning equipment to alert CCD camera from laser attacks. There are three ways to protect the CCD camera: closing the camera cover, satellite attitude maneuver and satellite orbit avoidance. In order to enhance the safety of optical remote sensing satellite in orbit, this paper explores the strategy of satellite avoidance. The avoidance strategy is expressed as the evasion of pre-determined target points in the orbital coordinates of virtual satellite. The so-called virtual satellite is a passive vehicle which superposes the satellite at the initial stage of avoidance. The target points share the consistent cycle time and the same semi-major axis with the virtual satellite, which ensures the properties of the satellite’s Sun-synchronous orbit remain unchanged. Moreover, to further strengthen the avoidance capability of satellite, it can perform multi-target-points avoid maneuvers. On occasions of fulfilling the satellite orbit tasks, the orbit can be restored back to virtual satellite through orbit maneuvers. Thereinto, the avoid maneuvers adopts pulse guidance. And the fuel consumption is also optimized. The avoidance strategy discussed in this article is applicable to optical remote sensing satellite when it is encountered with hostile attack of space-based laser anti-satellite.Keywords: optical remote sensing satellite, satellite avoidance, virtual satellite, avoid target-point, avoid maneuver
Procedia PDF Downloads 40418779 Use of Hierarchical Temporal Memory Algorithm in Heart Attack Detection
Authors: Tesnim Charrad, Kaouther Nouira, Ahmed Ferchichi
Abstract:
In order to reduce the number of deaths due to heart problems, we propose the use of Hierarchical Temporal Memory Algorithm (HTM) which is a real time anomaly detection algorithm. HTM is a cortical learning algorithm based on neocortex used for anomaly detection. In other words, it is based on a conceptual theory of how the human brain can work. It is powerful in predicting unusual patterns, anomaly detection and classification. In this paper, HTM have been implemented and tested on ECG datasets in order to detect cardiac anomalies. Experiments showed good performance in terms of specificity, sensitivity and execution time.Keywords: cardiac anomalies, ECG, HTM, real time anomaly detection
Procedia PDF Downloads 22818778 Stochastic Model Predictive Control for Linear Discrete-Time Systems with Random Dither Quantization
Authors: Tomoaki Hashimoto
Abstract:
Recently, feedback control systems using random dither quantizers have been proposed for linear discrete-time systems. However, the constraints imposed on state and control variables have not yet been taken into account for the design of feedback control systems with random dither quantization. Model predictive control is a kind of optimal feedback control in which control performance over a finite future is optimized with a performance index that has a moving initial and terminal time. An important advantage of model predictive control is its ability to handle constraints imposed on state and control variables. Based on the model predictive control approach, the objective of this paper is to present a control method that satisfies probabilistic state constraints for linear discrete-time feedback control systems with random dither quantization. In other words, this paper provides a method for solving the optimal control problems subject to probabilistic state constraints for linear discrete-time feedback control systems with random dither quantization.Keywords: optimal control, stochastic systems, random dither, quantization
Procedia PDF Downloads 44518777 Optimizing of the Micro EDM Parameters in Drilling of Titanium Ti-6Al-4V Alloy for Higher Machining Accuracy-Fuzzy Modelling
Authors: Ahmed A. D. Sarhan, Mum Wai Yip, M. Sayuti, Lim Siew Fen
Abstract:
Ti6Al4V alloy is highly used in the automotive and aerospace industry due to its good machining characteristics. Micro EDM drilling is commonly used to drill micro hole on extremely hard material with very high depth to diameter ratio. In this study, the parameters of micro-electrical discharge machining (EDM) in drilling of Ti6Al4V alloy is optimized for higher machining accuracy with less hole-dilation and hole taper ratio. The micro-EDM machining parameters includes, peak current and pulse on time. Fuzzy analysis was developed to evaluate the machining accuracy. The analysis shows that hole-dilation and hole-taper ratio are increased with the increasing of peak current and pulse on time. However, the surface quality deteriorates as the peak current and pulse on time increase. The combination that gives the optimum result for hole dilation is medium peak current and short pulse on time. Meanwhile, the optimum result for hole taper ratio is low peak current and short pulse on time.Keywords: Micro EDM, Ti-6Al-4V alloy, fuzzy logic based analysis, optimization, machining accuracy
Procedia PDF Downloads 49618776 Stereo Camera Based Speed-Hump Detection Process for Real Time Driving Assistance System in the Daytime
Authors: Hyun-Koo Kim, Yong-Hun Kim, Soo-Young Suk, Ju H. Park, Ho-Youl Jung
Abstract:
This paper presents an effective speed hump detection process at the day-time. we focus only on round types of speed humps in the day-time dynamic road environment. The proposed speed hump detection scheme consists mainly of two process as stereo matching and speed hump detection process. Our proposed process focuses to speed hump detection process. Speed hump detection process consist of noise reduction step, data fusion step, and speed hemp detection step. The proposed system is tested on Intel Core CPU with 2.80 GHz and 4 GB RAM tested in the urban road environments. The frame rate of test videos is 30 frames per second and the size of each frame of grabbed image sequences is 1280 pixels by 670 pixels. Using object-marked sequences acquired with an on-vehicle camera, we recorded speed humps and non-speed humps samples. Result of the tests, our proposed method can be applied in real-time systems by computation time is 13 ms. For instance; our proposed method reaches 96.1 %.Keywords: data fusion, round types speed hump, speed hump detection, surface filter
Procedia PDF Downloads 51018775 Hybrid Subspace Approach for Time Delay Estimation in MIMO Systems
Authors: Mojtaba Saeedinezhad, Sarah Yousefi
Abstract:
In this paper, we present a hybrid subspace approach for Time Delay Estimation (TDE) in multivariable systems. While several methods have been proposed for time delay estimation in SISO systems, delay estimation in MIMO systems were always a big challenge. In these systems the existing TDE methods have significant limitations because most of procedures are just based on system response estimation or correlation analysis. We introduce a new hybrid method for TDE in MIMO systems based on subspace identification and explicit output error method; and compare its performance with previously introduced procedures in presence of different noise levels and in a statistical manner. Then the best method is selected with multi objective decision making technique. It is shown that the performance of new approach is much better than the existing methods, even in low signal-to-noise conditions.Keywords: system identification, time delay estimation, ARX, OE, merit ratio, multi variable decision making
Procedia PDF Downloads 34618774 Design and Implementation of Partial Denoising Boundary Image Matching Using Indexing Techniques
Authors: Bum-Soo Kim, Jin-Uk Kim
Abstract:
In this paper, we design and implement a partial denoising boundary image matching system using indexing techniques. Converting boundary images to time-series makes it feasible to perform fast search using indexes even on a very large image database. Thus, using this converting method we develop a client-server system based on the previous partial denoising research in the GUI (graphical user interface) environment. The client first converts a query image given by a user to a time-series and sends denoising parameters and the tolerance with this time-series to the server. The server identifies similar images from the index by evaluating a range query, which is constructed using inputs given from the client, and sends the resulting images to the client. Experimental results show that our system provides much intuitive and accurate matching result.Keywords: boundary image matching, indexing, partial denoising, time-series matching
Procedia PDF Downloads 13718773 Determination of Surface Deformations with Global Navigation Satellite System Time Series
Authors: Ibrahim Tiryakioglu, Mehmet Ali Ugur, Caglar Ozkaymak
Abstract:
The development of GNSS technology has led to increasingly widespread and successful applications of GNSS surveys for monitoring crustal movements. However, multi-period GPS survey solutions have not been applied in monitoring vertical surface deformation. This study uses long-term GNSS time series that are required to determine vertical deformations. In recent years, the surface deformations that are parallel and semi-parallel to Bolvadin fault have occurred in Western Anatolia. These surface deformations have continued to occur in Bolvadin settlement area that is located mostly on alluvium ground. Due to these surface deformations, a number of cracks in the buildings located in the residential areas and breaks in underground water and sewage systems have been observed. In order to determine the amount of vertical surface deformations, two continuous GNSS stations have been established in the region. The stations have been operating since 2015 and 2017, respectively. In this study, GNSS observations from the mentioned two GNSS stations were processed with GAMIT/GLOBK (GNSS Analysis Massachusetts Institute of Technology/GLOBal Kalman) program package to create a coordinate time series. With the time series analyses, the GNSS stations’ behavior models (linear, periodical, etc.), the causes of these behaviors, and mathematical models were determined. The study results from the time series analysis of these two 2 GNSS stations shows approximately 50-80 mm/yr vertical movement.Keywords: Bolvadin fault, GAMIT, GNSS time series, surface deformations
Procedia PDF Downloads 16518772 A Real-Time Simulation Environment for Avionics Software Development and Qualification
Authors: Ferdinando Montemari, Antonio Vitale, Nicola Genito, Luca Garbarino, Urbano Tancredi, Domenico Accardo, Michele Grassi, Giancarmine Fasano, Anna Elena Tirri
Abstract:
The development of guidance, navigation and control algorithms and avionic procedures requires the disposability of suitable analysis and verification tools, such as simulation environments, which support the design process and allow detecting potential problems prior to the flight test, in order to make new technologies available at reduced cost, time and risk. This paper presents a simulation environment for avionic software development and qualification, especially aimed at equipment for general aviation aircrafts and unmanned aerial systems. The simulation environment includes models for short and medium-range radio-navigation aids, flight assistance systems, and ground control stations. All the software modules are able to simulate the modeled systems both in fast-time and real-time tests, and were implemented following component oriented modeling techniques and requirement based approach. The paper describes the specific models features, the architectures of the implemented software systems and its validation process. Performed validation tests highlighted the capability of the simulation environment to guarantee in real-time the required functionalities and performance of the simulated avionics systems, as well as to reproduce the interaction between these systems, thus permitting a realistic and reliable simulation of a complete mission scenario.Keywords: ADS-B, avionics, NAVAIDs, real-time simulation, TCAS, UAS ground control station
Procedia PDF Downloads 22818771 Minimizing Total Completion Time in No-Wait Flowshops with Setup Times
Authors: Ali Allahverdi
Abstract:
The m-machine no-wait flowshop scheduling problem is addressed in this paper. The objective is to minimize total completion time subject to the constraint that the makespan value is not greater than a certain value. Setup times are treated as separate from processing times. Several recent algorithms are adapted and proposed for the problem. An extensive computational analysis has been conducted for the evaluation of the proposed algorithms. The computational analysis indicates that the best proposed algorithm performs significantly better than the earlier existing best algorithm.Keywords: scheduling, no-wait flowshop, algorithm, setup times, total completion time, makespan
Procedia PDF Downloads 34018770 Automated Prepaid Billing Subscription System
Authors: Adekunle K. O, Adeniyi A. E, Kolawole E
Abstract:
One of the most dramatic trends in the communications market in recent years has been the growth of prepaid services. Today, prepaid no longer constitutes the low-revenue, basic-service segment. It is driven by a high margin, value-add service customers who view it as a convenient way of retaining control over their usage and communication spending while expecting high service levels. To service providers, prepaid services offer the advantage of reducing bad accounts while allowing them to predict usage and plan network resources. Yet, the real-time demands of prepaid services require a scalable, real-time platform to manage customers through their entire life cycle. It delivers integrated real-time rating, voucher management, recharge management, customer care and service provisioning for the generation of new prepaid services. It carries high scalability that can handle millions of prepaid customers in real-time through their entire life cycle.Keywords: prepaid billing, voucher management, customers, automated, security
Procedia PDF Downloads 11518769 In vivo Determination of Anticoagulant Property of the Tentacle Extract of Aurelia aurita (Moon Jellyfish) Using Sprague-Dawley Rats
Authors: Bea Carmel H. Casiding, Charmaine A. Guy, Funny Jovis P. Malasan, Katrina Chelsea B. Manlutac, Danielle Ann N. Novilla, Marianne R. Oliveros, Magnolia C. Sibulo
Abstract:
Moon jellyfish, Aurelia aurita, has become a popular research organism for diverse studies. Recent studies have verified the prevention of blood clotting properties of the moon jellyfish tentacle extract through in vitro methods. The purpose of this study was to validate the blood clotting ability of A. aurita tentacle extract using in vivo method of experimentation. The tentacles of A. aurita jellyfish were excised and filtered then centrifuged at 3000xg for 10 minutes. The crude nematocyst extract was suspended in 1:6 ratios with phosphate buffer solution and sonicated for three periods of 20 seconds each at 50 Hz. Protein concentration of the extract was determined using Bradford Assay. Bovine serum albumin was the standard solution used with the following concentrations: 35.0, 70.0, 105.0, 140.0, 175.0, 210.0, 245.0, and 280.0 µg/mL. The absorbance was read at 595 nm. Toxicity testing from OECD guidelines was adapted. The extract suspended in phosphate-buffered saline solution was arbitrarily set into three doses (0.1mg/kg, 0.3mg/kg, 0.5mg/kg) and were administered daily for five days to the experimental groups of five male Sprague-Dawley rats (one dose per group). Before and after the administration period, bleeding time and clotting time tests were performed. The One-way Analysis of Variance (ANOVA) was used to analyze the difference of before and after bleeding time and clotting time from the three treatment groups, time, positive and negative control groups. The average protein concentration of the sonicated crude tentacle extract was 206.5 µg/mL. The highest dose administered (0.5mg/kg) produced significant increase in the time for both bleeding and clotting tests. However, the preceding lower dose (0.3mg/kg) only was significantly effective for clotting time test. The protein contained in the tentacle extract with a concentration of 206.5 mcg/mL and dose of 0.3 mg/kg and 0.5 mg/kg of A. aurita elicited anticoagulating activity.Keywords: anticoagulant, bleeding time test, clotting time test, moon jellyfish
Procedia PDF Downloads 39718768 AniMoveMineR: Animal Behavior Exploratory Analysis Using Association Rules Mining
Authors: Suelane Garcia Fontes, Silvio Luiz Stanzani, Pedro L. Pizzigatti Corrła Ronaldo G. Morato
Abstract:
Environmental changes and major natural disasters are most prevalent in the world due to the damage that humanity has caused to nature and these damages directly affect the lives of animals. Thus, the study of animal behavior and their interactions with the environment can provide knowledge that guides researchers and public agencies in preservation and conservation actions. Exploratory analysis of animal movement can determine the patterns of animal behavior and with technological advances the ability of animals to be tracked and, consequently, behavioral studies have been expanded. There is a lot of research on animal movement and behavior, but we note that a proposal that combines resources and allows for exploratory analysis of animal movement and provide statistical measures on individual animal behavior and its interaction with the environment is missing. The contribution of this paper is to present the framework AniMoveMineR, a unified solution that aggregates trajectory analysis and data mining techniques to explore animal movement data and provide a first step in responding questions about the animal individual behavior and their interactions with other animals over time and space. We evaluated the framework through the use of monitored jaguar data in the city of Miranda Pantanal, Brazil, in order to verify if the use of AniMoveMineR allows to identify the interaction level between these jaguars. The results were positive and provided indications about the individual behavior of jaguars and about which jaguars have the highest or lowest correlation.Keywords: data mining, data science, trajectory, animal behavior
Procedia PDF Downloads 14418767 Estimation of Time Loss and Costs of Traffic Congestion: The Contingent Valuation Method
Authors: Amira Mabrouk, Chokri Abdennadher
Abstract:
The reduction of road congestion which is inherent to the use of vehicles is an obvious priority to public authority. Therefore, assessing the willingness to pay of an individual in order to save trip-time is akin to estimating the change in price which was the result of setting up a new transport policy to increase the networks fluidity and improving the level of social welfare. This study holds an innovative perspective. In fact, it initiates an economic calculation that has the objective of giving an estimation of the monetized time value during the trips made in Sfax. This research is founded on a double-objective approach. The aim of this study is to i) give an estimation of the monetized value of time; an hour dedicated to trips, ii) determine whether or not the consumer considers the environmental variables to be significant, iii) analyze the impact of applying a public management of the congestion via imposing taxation of city tolls on urban dwellers. This article is built upon a rich field survey led in the city of Sfax. With the use of the contingent valuation method, we analyze the “declared time preferences” of 450 drivers during rush hours. Based on the fond consideration of attributed bias of the applied method, we bring to light the delicacy of this approach with regards to the revelation mode and the interrogative techniques by following the NOAA panel recommendations bearing the exception of the valorization point and other similar studies about the estimation of transportation externality.Keywords: willingness to pay, contingent valuation, time value, city toll
Procedia PDF Downloads 43418766 Freight Time and Cost Optimization in Complex Logistics Networks, Using a Dimensional Reduction Method and K-Means Algorithm
Authors: Egemen Sert, Leila Hedayatifar, Rachel A. Rigg, Amir Akhavan, Olha Buchel, Dominic Elias Saadi, Aabir Abubaker Kar, Alfredo J. Morales, Yaneer Bar-Yam
Abstract:
The complexity of providing timely and cost-effective distribution of finished goods from industrial facilities to customers makes effective operational coordination difficult, yet effectiveness is crucial for maintaining customer service levels and sustaining a business. Logistics planning becomes increasingly complex with growing numbers of customers, varied geographical locations, the uncertainty of future orders, and sometimes extreme competitive pressure to reduce inventory costs. Linear optimization methods become cumbersome or intractable due to a large number of variables and nonlinear dependencies involved. Here we develop a complex systems approach to optimizing logistics networks based upon dimensional reduction methods and apply our approach to a case study of a manufacturing company. In order to characterize the complexity in customer behavior, we define a “customer space” in which individual customer behavior is described by only the two most relevant dimensions: the distance to production facilities over current transportation routes and the customer's demand frequency. These dimensions provide essential insight into the domain of effective strategies for customers; direct and indirect strategies. In the direct strategy, goods are sent to the customer directly from a production facility using box or bulk trucks. In the indirect strategy, in advance of an order by the customer, goods are shipped to an external warehouse near a customer using trains and then "last-mile" shipped by trucks when orders are placed. Each strategy applies to an area of the customer space with an indeterminate boundary between them. Specific company policies determine the location of the boundary generally. We then identify the optimal delivery strategy for each customer by constructing a detailed model of costs of transportation and temporary storage in a set of specified external warehouses. Customer spaces help give an aggregate view of customer behaviors and characteristics. They allow policymakers to compare customers and develop strategies based on the aggregate behavior of the system as a whole. In addition to optimization over existing facilities, using customer logistics and the k-means algorithm, we propose additional warehouse locations. We apply these methods to a medium-sized American manufacturing company with a particular logistics network, consisting of multiple production facilities, external warehouses, and customers along with three types of shipment methods (box truck, bulk truck and train). For the case study, our method forecasts 10.5% savings on yearly transportation costs and an additional 4.6% savings with three new warehouses.Keywords: logistics network optimization, direct and indirect strategies, K-means algorithm, dimensional reduction
Procedia PDF Downloads 13918765 Technology in the Calculation of People Health Level: Design of a Computational Tool
Authors: Sara Herrero Jaén, José María Santamaría García, María Lourdes Jiménez Rodríguez, Jorge Luis Gómez González, Adriana Cercas Duque, Alexandra González Aguna
Abstract:
Background: Health concept has evolved throughout history. The health level is determined by the own individual perception. It is a dynamic process over time so that you can see variations from one moment to the next. In this way, knowing the health of the patients you care for, will facilitate decision making in the treatment of care. Objective: To design a technological tool that calculates the people health level in a sequential way over time. Material and Methods: Deductive methodology through text analysis, extraction and logical knowledge formalization and education with expert group. Studying time: September 2015- actually. Results: A computational tool for the use of health personnel has been designed. It has 11 variables. Each variable can be given a value from 1 to 5, with 1 being the minimum value and 5 being the maximum value. By adding the result of the 11 variables we obtain a magnitude in a certain time, the health level of the person. The health calculator allows to represent people health level at a time, establishing temporal cuts being useful to determine the evolution of the individual over time. Conclusion: The Information and Communication Technologies (ICT) allow training and help in various disciplinary areas. It is important to highlight their relevance in the field of health. Based on the health formalization, care acts can be directed towards some of the propositional elements of the concept above. The care acts will modify the people health level. The health calculator allows the prioritization and prediction of different strategies of health care in hospital units.Keywords: calculator, care, eHealth, health
Procedia PDF Downloads 26418764 TessPy – Spatial Tessellation Made Easy
Authors: Jonas Hamann, Siavash Saki, Tobias Hagen
Abstract:
Discretization of urban areas is a crucial aspect in many spatial analyses. The process of discretization of space into subspaces without overlaps and gaps is called tessellation. It helps understanding spatial space and provides a framework for analyzing geospatial data. Tessellation methods can be divided into two groups: regular tessellations and irregular tessellations. While regular tessellation methods, like squares-grids or hexagons-grids, are suitable for addressing pure geometry problems, they cannot take the unique characteristics of different subareas into account. However, irregular tessellation methods allow the border between the subareas to be defined more realistically based on urban features like a road network or Points of Interest (POI). Even though Python is one of the most used programming languages when it comes to spatial analysis, there is currently no library that combines different tessellation methods to enable users and researchers to compare different techniques. To close this gap, we are proposing TessPy, an open-source Python package, which combines all above-mentioned tessellation methods and makes them easily accessible to everyone. The core functions of TessPy represent the five different tessellation methods: squares, hexagons, adaptive squares, Voronoi polygons, and city blocks. By using regular methods, users can set the resolution of the tessellation which defines the finesse of the discretization and the desired number of tiles. Irregular tessellation methods allow users to define which spatial data to consider (e.g., amenity, building, office) and how fine the tessellation should be. The spatial data used is open-source and provided by OpenStreetMap. This data can be easily extracted and used for further analyses. Besides the methodology of the different techniques, the state-of-the-art, including examples and future work, will be discussed. All dependencies can be installed using conda or pip; however, the former is more recommended.Keywords: geospatial data science, geospatial data analysis, tessellations, urban studies
Procedia PDF Downloads 12818763 Accessibility Analysis of Urban Green Space in Zadar Settlement, Croatia
Authors: Silvija Šiljeg, Ivan Marić, Ante Šiljeg
Abstract:
The accessibility of urban green spaces (UGS) is an integral element in the quality of life. Due to rapid urbanization, UGS studies have become a key element in urban planning. The potential benefits of space for its inhabitants are frequently analysed. A functional transport network system and the optimal spatial distribution of urban green surfaces are the prerequisites for maintaining the environmental equilibrium of the urban landscape. An accessibility analysis was conducted as part of the Urban Green Belts Project (UGB). The development of a GIS database for Zadar was the first step in generating the UGS accessibility indicator. Data were collected using the supervised classification method of multispectral LANDSAT images and manual vectorization of digital orthophoto images (DOF). An analysis of UGS accessibility according to the ANGst standard was conducted in the first phase of research. The accessibility indicator was generated on the basis of seven objective measurements, which included average UGS surface per capita and accessibility according to six functional levels of green surfaces. The generated indicator was compared with subjective measurements obtained by conducting a survey (718 respondents) within statistical units. The collected data reflected individual assessments and subjective evaluations of UGS accessibility. This study highlighted the importance of using objective and subjective measures in the process of understanding the accessibility of urban green surfaces. It may be concluded that when evaluating UGS accessibility, residents emphasize the immediate residential environment, ignoring higher UGS functional levels. It was also concluded that large areas of UGS within a city do not necessarily generate similar satisfaction with accessibility. The heterogeneity of output results may serve as guidelines for the further development of a functional UGS city network.Keywords: urban green spaces (UGS), accessibility indicator, subjective and objective measurements, Zadar
Procedia PDF Downloads 25918762 Optimal ECG Sampling Frequency for Multiscale Entropy-Based HRV
Authors: Manjit Singh
Abstract:
Multiscale entropy (MSE) is an extensively used index to provide a general understanding of multiple complexity of physiologic mechanism of heart rate variability (HRV) that operates on a wide range of time scales. Accurate selection of electrocardiogram (ECG) sampling frequency is an essential concern for clinically significant HRV quantification; high ECG sampling rate increase memory requirements and processing time, whereas low sampling rate degrade signal quality and results in clinically misinterpreted HRV. In this work, the impact of ECG sampling frequency on MSE based HRV have been quantified. MSE measures are found to be sensitive to ECG sampling frequency and effect of sampling frequency will be a function of time scale.Keywords: ECG (electrocardiogram), heart rate variability (HRV), multiscale entropy, sampling frequency
Procedia PDF Downloads 27118761 Experimental and Numerical Investigation of Fracture Behavior of Foamed Concrete Based on Three-Point Bending Test of Beams with Initial Notch
Authors: M. Kozłowski, M. Kadela
Abstract:
Foamed concrete is known for its low self-weight and excellent thermal and acoustic properties. For many years, it has been used worldwide for insulation to foundations and roof tiles, as backfill to retaining walls, sound insulation, etc. However, in the last years it has become a promising material also for structural purposes e.g. for stabilization of weak soils. Due to favorable properties of foamed concrete, many interests and studies were involved to analyze its strength, mechanical, thermal and acoustic properties. However, these studies do not cover the investigation of fracture energy which is the core factor governing the damage and fracture mechanisms. Only limited number of publications can be found in literature. The paper presents the results of experimental investigation and numerical campaign of foamed concrete based on three-point bending test of beams with initial notch. First part of the paper presents the results of a series of static loading tests performed to investigate the fracture properties of foamed concrete of varying density. Beam specimens with dimensions of 100×100×840 mm with a central notch were tested in three-point bending. Subsequently, remaining halves of the specimens with dimensions of 100×100×420 mm were tested again as un-notched beams in the same set-up with reduced distance between supports. The tests were performed in a hydraulic displacement controlled testing machine with a load capacity of 5 kN. Apart from measuring the loading and mid-span displacement, a crack mouth opening displacement (CMOD) was monitored. Based on the load – displacement curves of notched beams the values of fracture energy and tensile stress at failure were calculated. The flexural tensile strength was obtained on un-notched beams with dimensions of 100×100×420 mm. Moreover, cube specimens 150×150×150 mm were tested in compression to determine the compressive strength. Second part of the paper deals with numerical investigation of the fracture behavior of beams with initial notch presented in the first part of the paper. Extended Finite Element Method (XFEM) was used to simulate and analyze the damage and fracture process. The influence of meshing and variation of mechanical properties on results was investigated. Numerical models simulate correctly the behavior of beams observed during three-point bending. The numerical results show that XFEM can be used to simulate different fracture toughness of foamed concrete and fracture types. Using the XFEM and computer simulation technology allow for reliable approximation of load–bearing capacity and damage mechanisms of beams made of foamed concrete, which provides some foundations for realistic structural applications.Keywords: foamed concrete, fracture energy, three-point bending, XFEM
Procedia PDF Downloads 30018760 A 15 Minute-Based Approach for Berth Allocation and Quay Crane Assignment
Authors: Hoi-Lam Ma, Sai-Ho Chung
Abstract:
In traditional integrated berth allocation with quay crane assignment models, time dimension is usually assumed in hourly based. However, nowadays, transshipment becomes the main business to many container terminals, especially in Southeast Asia (e.g. Hong Kong and Singapore). In these terminals, vessel arrivals are usually very frequent with small handling volume and very short staying time. Therefore, the traditional hourly-based modeling approach may cause significant berth and quay crane idling, and consequently cannot meet their practical needs. In this connection, a 15-minute-based modeling approach is requested by industrial practitioners. Accordingly, a Three-level Genetic Algorithm (3LGA) with Quay Crane (QC) shifting heuristics is designed to fulfill the research gap. The objective function here is to minimize the total service time. Preliminary numerical results show that the proposed 15-minute-based approach can reduce the berth and QC idling significantly.Keywords: transshipment, integrated berth allocation, variable-in-time quay crane assignment, quay crane assignment
Procedia PDF Downloads 16918759 Implementation of the Canadian Emergency Department Triage and Acuity Scale (CTAS) in an Urgent Care Center in Saudi Arabia
Authors: Abdullah Arafat, Ali Al-Farhan, Amir Omair
Abstract:
Objectives: To review and assess the effectiveness of the implemented modified five-levels triage and acuity scale triage system in AL-Yarmook Urgent Care Center (UCC), King Abdulaziz Residential city, Riyadh, Saudi Arabia. Method: The applied study design was an observational cross sectional design. A data collection sheet was designed and distributed to triage nurses; the data collection was done during triage process and was directly observed by the co-investigator. Triage system was reviewed by measuring three time intervals as quality indicators: time before triage (TBT), time before being seen by physician (TBP) and total length of stay (TLS) taking in consideration timing of presentation and level of triage. Results: During the study period, a total of 187 patients were included in our study. 118 visits were at weekdays and 68 visits at weekends. Overall, 173 patients (92.5%) were seen by the physician in timely manner according to triage guidelines while 14 patients (7.5%) were not seen at appropriate time.Overall, The mean time before seen the triage nurse (TBT) was 5.36 minutes, the mean time to be seen by physician (TBP) was 22.6 minutes and the mean length of stay (TLS) was 59 minutes. The data didn’t showed significant increase in TBT, TBP, and number of patients not seen at the proper time, referral rate and admission rate during weekend. Conclusion: The CTAS is adaptable to countries beyond Canada and worked properly. The applied CTAS triage system in Al-Yarmook UCC is considered to be effective and well applied. Overall, urgent cases have been seen by physician in timely manner according to triage system and there was no delay in the management of urgent cases.Keywords: CTAS, emergency, Saudi Arabia, triage, urgent care
Procedia PDF Downloads 32118758 The Experimental Study on Reducing and Carbonizing Titanium-Containing Slag by Iron-Containing Coke
Authors: Yadong Liu
Abstract:
The experimental study on reduction carbonization of coke containing iron respectively with the particle size of <0.3mm, 0.3-0.6mm and 0.6-0.9mm and synthetic sea sand ore smelting reduction titanium-bearing slag as material were studied under the conditions of holding 6h at most at 1500℃. The effects of coke containing iron particle size and heat preservation time on the formation of TiC and the size of TiC crystal were studied by XRD, SEM and EDS. The results show that it is not good for the formation, concentration and growth of TiC crystal when the particle size of coke containing iron is too small or too large. The suitable particle size is 0.3~0.6mm. The heat preservation time of 2h basically ensures that all the component TiO2 in the slag are reduced and carbonized and converted to TiC. The size of TiC crystal will increase with the prolongation of heat preservation time. The thickness of the TiC layer can reach 20μm when the heat preservation time is 6h.Keywords: coke containing iron, formation and concentration and growth of TiC, reduction and carbonization, titanium-bearing slag
Procedia PDF Downloads 14918757 Tracing the Evolution of English and Urdu Languages: A Linguistic and Cultural Analysis
Authors: Aamna Zafar
Abstract:
Through linguistic and cultural analysis, this study seeks to trace the development of the English and Urdu languages. Along with examining how the vocabulary and syntax of English and Urdu have evolved over time and the linguistic trends that may be seen in these changes, this study will also look at the historical and cultural influences that have shaped the languages throughout time. The study will also look at how English and Urdu have changed over time, both in terms of language use and communication inside each other's cultures and globally. We'll research how these changes affect social relations and cultural identity, as well as how they might affect the future of these languages.Keywords: linguistic and cultural analysis, historical factors, cultural factors, vocabulary, syntax, significance
Procedia PDF Downloads 7518756 Camera Model Identification for Mi Pad 4, Oppo A37f, Samsung M20, and Oppo f9
Authors: Ulrich Wake, Eniman Syamsuddin
Abstract:
The model for camera model identificaiton is trained using pretrained model ResNet43 and ResNet50. The dataset consists of 500 photos of each phone. Dataset is divided into 1280 photos for training, 320 photos for validation and 400 photos for testing. The model is trained using One Cycle Policy Method and tested using Test-Time Augmentation. Furthermore, the model is trained for 50 epoch using regularization such as drop out and early stopping. The result is 90% accuracy for validation set and above 85% for Test-Time Augmentation using ResNet50. Every model is also trained by slightly updating the pretrained model’s weightsKeywords: One Cycle Policy, ResNet34, ResNet50, Test-Time Agumentation
Procedia PDF Downloads 20818755 Travel Time Estimation of Public Transport Networks Based on Commercial Incidence Areas in Quito Historic Center
Authors: M. Fernanda Salgado, Alfonso Tierra, David S. Sandoval, Wilbert G. Aguilar
Abstract:
Public transportation buses usually vary the speed depending on the places with the number of passengers. They require having efficient travel planning, a plan that will help them choose the fast route. Initially, an estimation tool is necessary to determine the travel time of each route, clearly establishing the possibilities. In this work, we give a practical solution that makes use of a concept that defines as areas of commercial incidence. These areas are based on the hypothesis that in the commercial places there is a greater flow of people and therefore the buses remain more time in the stops. The areas have one or more segments of routes, which have an incidence factor that allows to estimate the times. In addition, initial results are presented that verify the hypotheses and that promise adequately the travel times. In a future work, we take this approach to make an efficient travel planning system.Keywords: commercial incidence, planning, public transport, speed travel, travel time
Procedia PDF Downloads 25218754 Donoho-Stark’s and Hardy’s Uncertainty Principles for the Short-Time Quaternion Offset Linear Canonical Transform
Authors: Mohammad Younus Bhat
Abstract:
The quaternion offset linear canonical transform (QOLCT), which isa time-shifted and frequency-modulated version of the quaternion linear canonical transform (QLCT), provides a more general framework of most existing signal processing tools. For the generalized QOLCT, the classical Heisenberg’s and Lieb’s uncertainty principles have been studied recently. In this paper, we first define the short-time quaternion offset linear canonical transform (ST-QOLCT) and drive its relationship with the quaternion Fourier transform (QFT). The crux of the paper lies in the generalization of several well-known uncertainty principles for the ST-QOLCT, including Donoho-Stark’s uncertainty principle, Hardy’s uncertainty principle, Beurling’s uncertainty principle, and the logarithmic uncertainty principle.Keywords: Quaternion Fourier transform, Quaternion offset linear canonical transform, short-time quaternion offset linear canonical transform, uncertainty principle
Procedia PDF Downloads 211