Search results for: data space connector
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 28049

Search results for: data space connector

26219 The Image Redefinition of Urban Destinations: The Case of Madrid and Barcelona

Authors: Montserrat Crespi Vallbona, Marta Domínguez Pérez

Abstract:

Globalization impacts on cities and especially on their centers, especially on those spaces more visible and coveted. Changes are involved in processes such as touristification, gentrification or studentification, in addition of shop trendiness. The city becomes a good of interchange rather than a communal good for its inhabitants and consequently, its value is monetized. So, these different tendencies are analyzed: on one hand, the presence of tourists, the home rental increase, the explosion of businesses related to tourism; on the other hand; the return of middle classes or gentries to the center in a socio-spatial model that has changed highlighting the centers by their culture and their opportunities as well as by the value of public space and centrality; then, the interest of students (national and international) to be part of these city centers as dynamic groups and emerging classes with a higher purchasing power and better cultural capital than in the past; and finally, the conversion of old stores into modern ones, where vintage trend and the renewal of antiquity is the essence. All these transforming processes impact the European cities and redefine their image. All these trends reinforce the impression and brand of the urban center as an attractive space for investment, keeping such nonsense meaningful. These four tendencies have been spreading correlatively impacting the centers and transforming them involving the displacement of former residents of these spaces and revitalizing the center that is financed and commercialized in parallel. The cases of Madrid and Barcelona as spaces of greater evidence in Spain of these tendencies serve to illustrate these processes and represent the spearhead. Useful recommendations are presented to urban planners to find the conciliation of communal and commercialized spaces.

Keywords: gentrification, shop trendiness, studentification, touristification

Procedia PDF Downloads 178
26218 Development of Generalized Correlation for Liquid Thermal Conductivity of N-Alkane and Olefin

Authors: A. Ishag Mohamed, A. A. Rabah

Abstract:

The objective of this research is to develop a generalized correlation for the prediction of thermal conductivity of n-Alkanes and Alkenes. There is a minority of research and lack of correlation for thermal conductivity of liquids in the open literature. The available experimental data are collected covering the groups of n-Alkanes and Alkenes.The data were assumed to correlate to temperature using Filippov correlation. Nonparametric regression of Grace Algorithm was used to develop the generalized correlation model. A spread sheet program based on Microsoft Excel was used to plot and calculate the value of the coefficients. The results obtained were compared with the data that found in Perry's Chemical Engineering Hand Book. The experimental data correlated to the temperature ranged "between" 273.15 to 673.15 K, with R2 = 0.99.The developed correlation reproduced experimental data that which were not included in regression with absolute average percent deviation (AAPD) of less than 7 %. Thus the spread sheet was quite accurate which produces reliable data.

Keywords: N-Alkanes, N-Alkenes, nonparametric, regression

Procedia PDF Downloads 657
26217 Evaluation of Surgical Site Infection in Bile Spillage Cases Compared to Non Bile Spillage Cases Following Laparoscopic Cholecystectomy

Authors: Ishwor Paudel, Pratima Gautam, Sandeep Bhattarai

Abstract:

Bile spillage occurs frequently during laparoscopic cholecystectomy yet its impact on postoperative outcomes remains unknown. It might not be as innocuous as some surgeons tend to believe and in fact, might be associated with post-operative surgical site infections (SSI). It often leads to patient dissatisfaction, emergency department visits, with subsequent readmission and additional procedures. While some authors found no increase in SSI following bile spillage (BS) compared to non-bile spillage cases, others found bile spillage to be associated with SSI. Therefore we sought to examine whether bile spillage is indeed associated with an increased risk of postoperative wound infections after laparoscopic cholecystectomy. I hypothesize that patients who experience BS during operation, have an increased risk of SSI compared to those who do not. This is a prospective observational study conducted in the Department of Surgery, Patan Hospital over a period of one year. Patients undergoing Laparoscopic cholecystectomy were included and bile spillage, if happened was noted. All cases were followed up for 30 days and SSI was diagnosed as per CDC-defined criteria. Fisher’s test was applied to compare SSI in bile spillage versus non-bile spillage cases. A total of 112 patients were included in the final analysis. Bile spillage occurred in 20 cases and was absent in the rest i.e.92 cases. Among bile spillage cases, SSI was found in 4 cases (20%), whereas in nonbile spillage cases, SSI was found in 8 cases (8.7%). However, it was statistically not significant (p-value>0.05). Eleven (92%) cases were superficial SSI and one was an organ-space infection. No mortality or 30-day readmission. Spillage of Gallbladder content does not lead to an increase in SSIs. However as the rate of SSI is still higher, Surgeons should be careful to avoid iatrogenic gallbladder perforation and in case of bile spillage thorough peritoneal irrigation with normal saline should be done.

Keywords: organ space infection, Laparoscopic cholecystectomy, biliary spillage, surgical site infection

Procedia PDF Downloads 81
26216 Survey on Arabic Sentiment Analysis in Twitter

Authors: Sarah O. Alhumoud, Mawaheb I. Altuwaijri, Tarfa M. Albuhairi, Wejdan M. Alohaideb

Abstract:

Large-scale data stream analysis has become one of the important business and research priorities lately. Social networks like Twitter and other micro-blogging platforms hold an enormous amount of data that is large in volume, velocity and variety. Extracting valuable information and trends out of these data would aid in a better understanding and decision-making. Multiple analysis techniques are deployed for English content. Moreover, one of the languages that produce a large amount of data over social networks and is least analyzed is the Arabic language. The proposed paper is a survey on the research efforts to analyze the Arabic content in Twitter focusing on the tools and methods used to extract the sentiments for the Arabic content on Twitter.

Keywords: big data, social networks, sentiment analysis, twitter

Procedia PDF Downloads 585
26215 An Evolutionary Approach for QAOA for Max-Cut

Authors: Francesca Schiavello

Abstract:

This work aims to create a hybrid algorithm, combining Quantum Approximate Optimization Algorithm (QAOA) with an Evolutionary Algorithm (EA) in the place of traditional gradient based optimization processes. QAOA’s were first introduced in 2014, where, at the time, their algorithm performed better than the traditional best known classical algorithm for Max-cut graphs. Whilst classical algorithms have improved since then and have returned to being faster and more efficient, this was a huge milestone for quantum computing, and their work is often used as a benchmarking tool and a foundational tool to explore variants of QAOA’s. This, alongside with other famous algorithms like Grover’s or Shor’s, highlights to the world the potential that quantum computing holds. It also presents the reality of a real quantum advantage where, if the hardware continues to improve, this could constitute a revolutionary era. Given that the hardware is not there yet, many scientists are working on the software side of things in the hopes of future progress. Some of the major limitations holding back quantum computing are the quality of qubits and the noisy interference they generate in creating solutions, the barren plateaus that effectively hinder the optimization search in the latent space, and the availability of number of qubits limiting the scale of the problem that can be solved. These three issues are intertwined and are part of the motivation for using EAs in this work. Firstly, EAs are not based on gradient or linear optimization methods for the search in the latent space, and because of their freedom from gradients, they should suffer less from barren plateaus. Secondly, given that this algorithm performs a search in the solution space through a population of solutions, it can also be parallelized to speed up the search and optimization problem. The evaluation of the cost function, like in many other algorithms, is notoriously slow, and the ability to parallelize it can drastically improve the competitiveness of QAOA’s with respect to purely classical algorithms. Thirdly, because of the nature and structure of EA’s, solutions can be carried forward in time, making them more robust to noise and uncertainty. Preliminary results show that the EA algorithm attached to QAOA can perform on par with the traditional QAOA with a Cobyla optimizer, which is a linear based method, and in some instances, it can even create a better Max-Cut. Whilst the final objective of the work is to create an algorithm that can consistently beat the original QAOA, or its variants, due to either speedups or quality of the solution, this initial result is promising and show the potential of EAs in this field. Further tests need to be performed on an array of different graphs with the parallelization aspect of the work commencing in October 2023 and tests on real hardware scheduled for early 2024.

Keywords: evolutionary algorithm, max cut, parallel simulation, quantum optimization

Procedia PDF Downloads 62
26214 Estimating Current Suicide Rates Using Google Trends

Authors: Ladislav Kristoufek, Helen Susannah Moat, Tobias Preis

Abstract:

Data on the number of people who have committed suicide tends to be reported with a substantial time lag of around two years. We examine whether online activity measured by Google searches can help us improve estimates of the number of suicide occurrences in England before official figures are released. Specifically, we analyse how data on the number of Google searches for the terms “depression” and “suicide” relate to the number of suicides between 2004 and 2013. We find that estimates drawing on Google data are significantly better than estimates using previous suicide data alone. We show that a greater number of searches for the term “depression” is related to fewer suicides, whereas a greater number of searches for the term “suicide” is related to more suicides. Data on suicide related search behaviour can be used to improve current estimates of the number of suicide occurrences.

Keywords: nowcasting, search data, Google Trends, official statistics

Procedia PDF Downloads 363
26213 Analysis of Performance of 3T1D Dynamic Random-Access Memory Cell

Authors: Nawang Chhunid, Gagnesh Kumar

Abstract:

On-chip memories consume a significant portion of the overall die space and power in modern microprocessors. On-chip caches depend on Static Random-Access Memory (SRAM) cells and scaling of technology occurring as per Moore’s law. Unfortunately, the scaling is affecting stability, performance, and leakage power which will become major problems for future SRAMs in aggressive nanoscale technologies due to increasing device mismatch and variations. 3T1D Dynamic Random-Access Memory (DRAM) cell is a non-destructive read DRAM cell with three transistors and a gated diode. In 3T1D DRAM cell gated diode (D1) acts as a storage device and also as an amplifier, which leads to fast read access. Due to its high tolerance to process variation, high density, and low cost of memory as compared to 6T SRAM cell, it is universally used by the advanced microprocessor for on chip data and program memory. In the present paper, it has been shown that 3T1D DRAM cell can perform better in terms of fast read access as compared to 6T, 4T, 3T SRAM cells, respectively.

Keywords: DRAM Cell, Read Access Time, Retention Time, Average Power dissipation

Procedia PDF Downloads 315
26212 Assessing the Spatial Distribution of Urban Parks Using Remote Sensing and Geographic Information Systems Techniques

Authors: Hira Jabbar, Tanzeel-Ur Rehman

Abstract:

Urban parks and open spaces play a significant role in improving physical and mental health of the citizens, strengthen the societies and make the cities more attractive places to live and work. As the world’s cities continue to grow, continuing to value green space in cities is vital but is also a challenge, particularly in developing countries where there is pressure for space, resources, and development. Offering equal opportunity of accessibility to parks is one of the important issues of park distribution. The distribution of parks should allow all inhabitants to have close proximity to their residence. Remote sensing and Geographic information systems (GIS) can provide decision makers with enormous opportunities to improve the planning and management of Park facilities. This study exhibits the capability of GIS and RS techniques to provide baseline knowledge about the distribution of parks, level of accessibility and to help in identification of potential areas for such facilities. For this purpose Landsat OLI imagery for year 2016 was acquired from USGS Earth Explorer. Preprocessing models were applied using Erdas Imagine 2014v for the atmospheric correction and NDVI model was developed and applied to quantify the land use/land cover classes including built up, barren land, water, and vegetation. The parks amongst total public green spaces were selected based on their signature in remote sensing image and distribution. Percentages of total green and parks green were calculated for each town of Lahore City and results were then synchronized with the recommended standards. ANGSt model was applied to calculate the accessibility from parks. Service area analysis was performed using Network Analyst tool. Serviceability of these parks has been evaluated by employing statistical indices like service area, service population and park area per capita. Findings of the study may contribute in helping the town planners for understanding the distribution of parks, demands for new parks and potential areas which are deprived of parks. The purpose of present study is to provide necessary information to planners, policy makers and scientific researchers in the process of decision making for the management and improvement of urban parks.

Keywords: accessible natural green space standards (ANGSt), geographic information systems (GIS), remote sensing (RS), United States geological survey (USGS)

Procedia PDF Downloads 345
26211 Pragmatism in Adaptive Reuse of Obsolete Industrial Land in China

Authors: Yong Li

Abstract:

Major cities in China has experienced a shift from production based on manufacturing industry to tertiary industry. How to make a better use of existing obsolete industrial land within urban cores has become a difficult problem for many policymakers. City governments regard old manufacturing industrial land as an important source of land to facilitate the development of the cities. Despite the announcement of policies in promoting that, a large portion of industrial land is still not properly redeveloped and most of them became obsolete. The study uses the project of Xinyi International Club as a case to examine the process of adaptive reuse of obsolete industrial space in Guangzhou, China. It attempts to elucidate the underlying mechanisms by identifying the key forces from both the government and the private sectors in influencing the process. The study found that market forces in transforming industrial space are exerting a strong impact on the existing land use planning system in Chinese cities. Pragmatic relaxation of the formal land use the regulatory framework and government supportive land-use intervention have also been crucial towards achieving successful implementation of the restructuring project and making it a showcase. This study questions whether these extraordinary measures, in particular, the use of temporary land use permit, are sustainable in facilitating the transformation of derelict industrial land, and in informing future industrial land-use restructuring policies. It concludes that, while the land use regulatory system in China is becoming increasingly dynamic and flexible, it remains ill-equipped in responding positively to the market, which is characterized by an increasing bargaining power of the private sector. A comprehensive appraisal of the overall impacts of these adaptive re-uses on society is wanting.

Keywords: China, land alteration, obsolete industrial properties, urban planning

Procedia PDF Downloads 150
26210 Critical Evaluation of Groundwater Monitoring Networks for Machine Learning Applications

Authors: Pedro Martinez-Santos, Víctor Gómez-Escalonilla, Silvia Díaz-Alcaide, Esperanza Montero, Miguel Martín-Loeches

Abstract:

Groundwater monitoring networks are critical in evaluating the vulnerability of groundwater resources to depletion and contamination, both in space and time. Groundwater monitoring networks typically grow over decades, often in organic fashion, with relatively little overall planning. The groundwater monitoring networks in the Madrid area, Spain, were reviewed for the purpose of identifying gaps and opportunities for improvement. Spatial analysis reveals the presence of various monitoring networks belonging to different institutions, with several hundred observation wells in an area of approximately 4000 km2. This represents several thousand individual data entries, some going back to the early 1970s. Major issues included overlap between the networks, unknown screen depth/vertical distribution for many observation boreholes, uneven time series, uneven monitored species, and potentially suboptimal locations. Results also reveal there is sufficient information to carry out a spatial and temporal analysis of groundwater vulnerability based on machine learning applications. These can contribute to improve the overall planning of monitoring networks’ expansion into the future.

Keywords: groundwater monitoring, observation networks, machine learning, madrid

Procedia PDF Downloads 82
26209 On the Network Packet Loss Tolerance of SVM Based Activity Recognition

Authors: Gamze Uslu, Sebnem Baydere, Alper K. Demir

Abstract:

In this study, data loss tolerance of Support Vector Machines (SVM) based activity recognition model and multi activity classification performance when data are received over a lossy wireless sensor network is examined. Initially, the classification algorithm we use is evaluated in terms of resilience to random data loss with 3D acceleration sensor data for sitting, lying, walking and standing actions. The results show that the proposed classification method can recognize these activities successfully despite high data loss. Secondly, the effect of differentiated quality of service performance on activity recognition success is measured with activity data acquired from a multi hop wireless sensor network, which introduces high data loss. The effect of number of nodes on the reliability and multi activity classification success is demonstrated in simulation environment. To the best of our knowledge, the effect of data loss in a wireless sensor network on activity detection success rate of an SVM based classification algorithm has not been studied before.

Keywords: activity recognition, support vector machines, acceleration sensor, wireless sensor networks, packet loss

Procedia PDF Downloads 479
26208 GIS Data Governance: GIS Data Submission Process for Build-in Project, Replacement Project at Oman electricity Transmission Company

Authors: Rahma Saleh Hussein Al Balushi

Abstract:

Oman Electricity Transmission Company's (OETC) vision is to be a renowned world-class transmission grid by 2025, and one of the indications of achieving the vision is obtaining Asset Management ISO55001 certification, which required setting out a documented Standard Operating Procedures (SOP). Hence, documented SOP for the Geographical information system data process has been established. Also, to effectively manage and improve OETC power transmission, asset data and information need to be governed as such by Asset Information & GIS department. This paper will describe in detail the current GIS data submission process and the journey for developing it. The methodology used to develop the process is based on three main pillars, which are system and end-user requirements, Risk evaluation, data availability, and accuracy. The output of this paper shows the dramatic change in the used process, which results subsequently in more efficient, accurate, and updated data. Furthermore, due to this process, GIS has been and is ready to be integrated with other systems as well as the source of data for all OETC users. Some decisions related to issuing No objection certificates (NOC) for excavation permits and scheduling asset maintenance plans in Computerized Maintenance Management System (CMMS) have been made consequently upon GIS data availability. On the Other hand, defining agreed and documented procedures for data collection, data systems update, data release/reporting and data alterations has also contributed to reducing the missing attributes and enhance data quality index of GIS transmission data. A considerable difference in Geodatabase (GDB) completeness percentage was observed between the years 2017 and year 2022. Overall, concluding that by governance, asset information & GIS department can control the GIS data process; collect, properly record, and manage asset data and information within the OETC network. This control extends to other applications and systems integrated with/related to GIS systems.

Keywords: asset management ISO55001, standard procedures process, governance, CMMS

Procedia PDF Downloads 130
26207 Application of Adaptive Architecture in Building Technologies: A Case Study of Neuhoff Site in Nashville, Tennessee

Authors: Shohreh Moshiri, Hossein Alimohammadi

Abstract:

Building construction has a great impact on climate change. Adaptive design strategies were developed to provide new life and purpose to old buildings and create new environments with economic benefits to meet resident needs. The role of smart material systems is undeniable in providing adaptivity of the architectural environments and their effects on creating better adaptive building environments. In this research, a case study named Neuhoff site located near Cumberland River in the Germantown neighborhood in the city of Nashville, Tennessee, was considered. This building in the early 1920s was constructed as a meat-packing facility and then served as a mixed-use space; however, New City has partnered with world-class architects to reinvent this site to be changed to mixed-use waterfront development. The future office space will be designed with LEED certification as a goal. Environmentally friendly sensitive materials and designs will offer for all adaptive reuse of the building. The smart materials and their applications, especially in the field of building technology and architecture, were emphasized in providing a renovation plan for the site. The advantages and qualities of smart material systems were targeted to explore in this research on the field of architecture. Also, this research helps to understand better the effects of smart material systems on the construction and design processes, exploration of the way to make architecture with better adaptive characteristics, plus provide optimal environmental situations for the users, which reflect on the climatic, structural, and architectural performances.

Keywords: adaptive architecture, building technology, case study, smart material systems

Procedia PDF Downloads 76
26206 Efects of Data Corelation in a Sparse-View Compresive Sensing Based Image Reconstruction

Authors: Sajid Abas, Jon Pyo Hong, Jung-Ryun Le, Seungryong Cho

Abstract:

Computed tomography and laminography are heavily investigated in a compressive sensing based image reconstruction framework to reduce the dose to the patients as well as to the radiosensitive devices such as multilayer microelectronic circuit boards. Nowadays researchers are actively working on optimizing the compressive sensing based iterative image reconstruction algorithm to obtain better quality images. However, the effects of the sampled data’s properties on reconstructed the image’s quality, particularly in an insufficient sampled data conditions have not been explored in computed laminography. In this paper, we investigated the effects of two data properties i.e. sampling density and data incoherence on the reconstructed image obtained by conventional computed laminography and a recently proposed method called spherical sinusoidal scanning scheme. We have found that in a compressive sensing based image reconstruction framework, the image quality mainly depends upon the data incoherence when the data is uniformly sampled.

Keywords: computed tomography, computed laminography, compressive sending, low-dose

Procedia PDF Downloads 470
26205 Pareto System of Optimal Placement and Sizing of Distributed Generation in Radial Distribution Networks Using Particle Swarm Optimization

Authors: Sani M. Lawal, Idris Musa, Aliyu D. Usman

Abstract:

The Pareto approach of optimal solutions in a search space that evolved in multi-objective optimization problems is adopted in this paper, which stands for a set of solutions in the search space. This paper aims at presenting an optimal placement of Distributed Generation (DG) in radial distribution networks with an optimal size for minimization of power loss and voltage deviation as well as maximizing voltage profile of the networks. And these problems are formulated using particle swarm optimization (PSO) as a constraint nonlinear optimization problem with both locations and sizes of DG being continuous. The objective functions adopted are the total active power loss function and voltage deviation function. The multiple nature of the problem, made it necessary to form a multi-objective function in search of the solution that consists of both the DG location and size. The proposed PSO algorithm is used to determine optimal placement and size of DG in a distribution network. The output indicates that PSO algorithm technique shows an edge over other types of search methods due to its effectiveness and computational efficiency. The proposed method is tested on the standard IEEE 34-bus and validated with 33-bus test systems distribution networks. Results indicate that the sizing and location of DG are system dependent and should be optimally selected before installing the distributed generators in the system and also an improvement in the voltage profile and power loss reduction have been achieved.

Keywords: distributed generation, pareto, particle swarm optimization, power loss, voltage deviation

Procedia PDF Downloads 370
26204 Digital Material Characterization Using the Quantum Fourier Transform

Authors: Felix Givois, Nicolas R. Gauger, Matthias Kabel

Abstract:

The efficient digital material characterization is of great interest to many fields of application. It consists of the following three steps. First, a 3D reconstruction of 2D scans must be performed. Then, the resulting gray-value image of the material sample is enhanced by image processing methods. Finally, partial differential equations (PDE) are solved on the segmented image, and by averaging the resulting solutions fields, effective properties like stiffness or conductivity can be computed. Due to the high resolution of current CT images, the latter is typically performed with matrix-free solvers. Among them, a solver that uses the explicit formula of the Green-Eshelby operator in Fourier space has been proposed by Moulinec and Suquet. Its algorithmic, most complex part is the Fast Fourier Transformation (FFT). In our talk, we will discuss the potential quantum advantage that can be obtained by replacing the FFT with the Quantum Fourier Transformation (QFT). We will especially show that the data transfer for noisy intermediate-scale quantum (NISQ) devices can be improved by using appropriate boundary conditions for the PDE, which also allows using semi-classical versions of the QFT. In the end, we will compare the results of the QFT-based algorithm for simple geometries with the results of the FFT-based homogenization method.

Keywords: most likelihood amplitude estimation (MLQAE), numerical homogenization, quantum Fourier transformation (QFT), NISQ devises

Procedia PDF Downloads 82
26203 Determination of the Relative Humidity Profiles in an Internal Micro-Climate Conditioned Using Evaporative Cooling

Authors: M. Bonello, D. Micallef, S. P. Borg

Abstract:

Driven by increased comfort standards, but at the same time high energy consciousness, energy-efficient space cooling has become an essential aspect of building design. Its aims are simple, aiming at providing satisfactory thermal comfort for individuals in an interior space using low energy consumption cooling systems. In this context, evaporative cooling is both an energy-efficient and an eco-friendly cooling process. In the past two decades, several academic studies have been performed to determine the resulting thermal comfort produced by an evaporative cooling system, including studies on temperature profiles, air speed profiles, effect of clothing and personnel activity. To the best knowledge of the authors, no studies have yet considered the analysis of relative humidity (RH) profiles in a space cooled using evaporative cooling. Such a study will determine the effect of different humidity levels on a person's thermal comfort and aid in the consequent improvement designs of such future systems. Under this premise, the research objective is to characterise the resulting different RH profiles in a chamber micro-climate using the evaporative cooling system in which the inlet air speed, temperature and humidity content are varied. The chamber shall be modelled using Computational Fluid Dynamics (CFD) in ANSYS Fluent. Relative humidity shall be modelled using a species transport model while the k-ε RNG formulation is the proposed turbulence model that is to be used. The model shall be validated with measurements taken using an identical test chamber in which tests are to be conducted under the different inlet conditions mentioned above, followed by the verification of the model's mesh and time step. The verified and validated model will then be used to simulate other inlet conditions which would be impractical to conduct in the actual chamber. More details of the modelling and experimental approach will be provided in the full paper The main conclusions from this work are two-fold: the micro-climatic relative humidity spatial distribution within the room is important to consider in the context of investigating comfort at occupant level; and the investigation of a human being's thermal comfort (based on Predicted Mean Vote – Predicted Percentage Dissatisfied [PMV-PPD] values) and its variation with different locations of relative humidity values. The study provides the necessary groundwork for investigating the micro-climatic RH conditions of environments cooled using evaporative cooling. Future work may also target the analysis of ways in which evaporative cooling systems may be improved to better the thermal comfort of human beings, specifically relating to the humidity content around a sedentary person.

Keywords: chamber micro-climate, evaporative cooling, relative humidity, thermal comfort

Procedia PDF Downloads 160
26202 Fuzzy Wavelet Model to Forecast the Exchange Rate of IDR/USD

Authors: Tri Wijayanti Septiarini, Agus Maman Abadi, Muhammad Rifki Taufik

Abstract:

The exchange rate of IDR/USD can be the indicator to analysis Indonesian economy. The exchange rate as a important factor because it has big effect in Indonesian economy overall. So, it needs the analysis data of exchange rate. There is decomposition data of exchange rate of IDR/USD to be frequency and time. It can help the government to monitor the Indonesian economy. This method is very effective to identify the case, have high accurate result and have simple structure. In this paper, data of exchange rate that used is weekly data from December 17, 2010 until November 11, 2014.

Keywords: the exchange rate, fuzzy mamdani, discrete wavelet transforms, fuzzy wavelet

Procedia PDF Downloads 578
26201 Humanising Digital Healthcare to Build Capacity by Harnessing the Power of Patient Data

Authors: Durhane Wong-Rieger, Kawaldip Sehmi, Nicola Bedlington, Nicole Boice, Tamás Bereczky

Abstract:

Patient-generated health data should be seen as the expression of the experience of patients, including the outcomes reflecting the impact a treatment or service had on their physical health and wellness. We discuss how the healthcare system can reach a place where digital is a determinant of health - where data is generated by patients and is respected and which acknowledges their contribution to science. We explore the biggest barriers facing this. The International Experience Exchange with Patient Organisation’s Position Paper is based on a global patient survey conducted in Q3 2021 that received 304 responses. Results were discussed and validated by the 15 patient experts and supplemented with literature research. Results are a subset of this. Our research showed patient communities want to influence how their data is generated, shared, and used. Our study concludes that a reasonable framework is needed to protect the integrity of patient data and minimise abuse, and build trust. Results also demonstrated a need for patient communities to have more influence and control over how health data is generated, shared, and used. The results clearly highlight that the community feels there is a lack of clear policies on sharing data.

Keywords: digital health, equitable access, humanise healthcare, patient data

Procedia PDF Downloads 87
26200 Reduction of Specific Energy Consumption in Microfiltration of Bacillus velezensis Broth by Air Sparging and Turbulence Promoter

Authors: Jovana Grahovac, Ivana Pajcin, Natasa Lukic, Jelena Dodic, Aleksandar Jokic

Abstract:

To obtain purified biomass to be used in the plant pathogen biocontrol or as soil biofertilizer, it is necessary to eliminate residual broth components at the end of the fermentation process. The main drawback of membrane separation techniques is permeate flux decline due to the membrane fouling. Fouling mitigation measures increase the pressure drop along membrane channel due to the increased resistance to flow of the feed suspension, thus increasing the hydraulic power drop. At the same time, these measures lead to an increase in the permeate flux due to the reduced resistance of the filtration cake on the membrane surface. Because of these opposing effects, the energy efficiency of fouling mitigation measures is limited, and the justification of its application is provided by information on a reducing specific energy consumption compared to a case without any measures employed. In this study, the influence of static mixer (Kenics) and air-sparging (two-phase flow) on reduction of specific energy consumption (ER) was investigated. Cultivation Bacillus velezensis was carried out in the 3-L bioreactor (Biostat® Aplus) containing 2 L working volume with two parallel Rushton turbines and without internal baffles. Cultivation was carried out at 28 °C on at 150 rpm with an aeration rate of 0.75 vvm during 96 h. The experiments were carried out in a conventional cross-flow microfiltration unit. During experiments, permeate and retentate were recycled back to the broth vessel to simulate continuous process. The single channel ceramic membrane (TAMI Deutschland) used had a nominal pore size 200 nm with the length of 250 mm and an inner/external diameter of 6/10 mm. The useful membrane channel surface was 4.33×10⁻³ m². Air sparging was brought by the pressurized air connected by a three-way valve to the feed tube by a simple T-connector without diffusor. The different approaches to flux improvement are compared in terms of energy consumption. Reduction of specific energy consumption compared to microfiltration without fouling mitigation is around 49% and 63%, for use of two-phase flow and a static mixer, respectively. In the case of a combination of these two fouling mitigation methods, ER is 60%, i.e., slightly lower compared to the use of turbulence promoter alone. The reason for this result can be found in the fact that flux increase is more affected by the presence of a Kenics static mixer while sparging results in an increase of energy used during microfiltration. By comparing combined method with turbulence promoter flux enhancement method ER is negative (-7%) which can be explained by increased power consumption for air flow with moderate contribution to the flux increase. Another confirmation for this fact can be found by comparing energy consumption values for combined method with energy consumption in the case of two-phase flow. In this instance energy reduction (ER) is 22% that demonstrates that turbulence promoter is more efficient compared to two phase flow. Antimicrobial activity of Bacillus velezensis biomass against phytopathogenic isolates Xanthomonas campestris was preserved under different fouling reduction methods.

Keywords: Bacillus velezensis, microfiltration, static mixer, two-phase flow

Procedia PDF Downloads 121
26199 Triangular Libration Points in the R3bp under Combined Effects of Oblateness, Radiation and Power-Law Profile

Authors: Babatunde James Falaye, Shi Hai Dong, Kayode John Oyewumi

Abstract:

We study the e ffects of oblateness up to J4 of the primaries and power-law density pro file (PDP) on the linear stability of libration location of an in nitesimal mass within the framework of restricted three body problem (R3BP), by using a more realistic model in which a disc with PDP is rotating around the common center of the system mass with perturbed mean motion. The existence and stability of triangular equilibrium points have been explored. It has been shown that triangular equilibrium points are stable for 0 < μ < μc and unstable for μc ≤ μ ≤ 1/2, where c denotes the critical mass parameter. We find that, the oblateness up to J2 of the primaries and the radiation reduces the stability range while the oblateness up to J4 of the primaries increases the size of stability both in the context where PDP is considered and ignored. The PDP has an e ect of about ≈0:01 reduction on the application of c to Earth-Moon and Jupiter-Moons systems. We find that the comprehensive eff ects of the perturbations have a stabilizing proclivity. However, the oblateness up to J2 of the primaries and the radiation of the primaries have tendency for instability, while coecients up to J4 of the primaries have stability predisposition. In the limiting case c = 0, and also by setting appropriate parameter(s) to zero, our results are in excellent agreement with the ones obtained previously. Libration points play a very important role in space mission and as a consequence, our results have a practical application in space dynamics and related areas. The model may be applied to study the navigation and station-keeping operations of spacecraft (in nitesimal mass) around the Jupiter (more massive) -Callisto (less massive) system, where PDP accounts for the circumsolar ring of asteroidal dust, which has a cloud of dust permanently in its wake.

Keywords: libration points, oblateness, power-law density profile, restricted three-body problem

Procedia PDF Downloads 328
26198 Use of Machine Learning in Data Quality Assessment

Authors: Bruno Pinto Vieira, Marco Antonio Calijorne Soares, Armando Sérgio de Aguiar Filho

Abstract:

Nowadays, a massive amount of information has been produced by different data sources, including mobile devices and transactional systems. In this scenario, concerns arise on how to maintain or establish data quality, which is now treated as a product to be defined, measured, analyzed, and improved to meet consumers' needs, which is the one who uses these data in decision making and companies strategies. Information that reaches low levels of quality can lead to issues that can consume time and money, such as missed business opportunities, inadequate decisions, and bad risk management actions. The step of selecting, identifying, evaluating, and selecting data sources with significant quality according to the need has become a costly task for users since the sources do not provide information about their quality. Traditional data quality control methods are based on user experience or business rules limiting performance and slowing down the process with less than desirable accuracy. Using advanced machine learning algorithms, it is possible to take advantage of computational resources to overcome challenges and add value to companies and users. In this study, machine learning is applied to data quality analysis on different datasets, seeking to compare the performance of the techniques according to the dimensions of quality assessment. As a result, we could create a ranking of approaches used, besides a system that is able to carry out automatically, data quality assessment.

Keywords: machine learning, data quality, quality dimension, quality assessment

Procedia PDF Downloads 154
26197 Exploring Data Leakage in EEG Based Brain-Computer Interfaces: Overfitting Challenges

Authors: Khalida Douibi, Rodrigo Balp, Solène Le Bars

Abstract:

In the medical field, applications related to human experiments are frequently linked to reduced samples size, which makes the training of machine learning models quite sensitive and therefore not very robust nor generalizable. This is notably the case in Brain-Computer Interface (BCI) studies, where the sample size rarely exceeds 20 subjects or a few number of trials. To address this problem, several resampling approaches are often used during the data preparation phase, which is an overly critical step in a data science analysis process. One of the naive approaches that is usually applied by data scientists consists in the transformation of the entire database before the resampling phase. However, this can cause model’ s performance to be incorrectly estimated when making predictions on unseen data. In this paper, we explored the effect of data leakage observed during our BCI experiments for device control through the real-time classification of SSVEPs (Steady State Visually Evoked Potentials). We also studied potential ways to ensure optimal validation of the classifiers during the calibration phase to avoid overfitting. The results show that the scaling step is crucial for some algorithms, and it should be applied after the resampling phase to avoid data leackage and improve results.

Keywords: data leackage, data science, machine learning, SSVEP, BCI, overfitting

Procedia PDF Downloads 156
26196 Visual Outcome After 360-Degree Retinectomy in Total Rhegmatogenous Retinal Detachment with Advanced Proliferative Vitreoretinopathy: A Case Series

Authors: Andriati Nadhilah Widyarini, Ezra Margareth

Abstract:

Introduction: Rhegmatogenous retinal detachment is a condition where there’s a break in the retina, which allows the vitreous to directly enter the subretinal space. Proliferative vitreoretinopathy (PVR) may develop due to this condition and can result in a new break, which could cause traction on the previously detached retina. Various methods of therapy can be done to treat this complication. Case: This case series involved 2 eyes of 2 patients who had total retinal detachment with advanced PVR. Pars plana vitrectomy was performed, and a 360-degree retinectomy procedure with perfluorocarbon liquid usage was done. This was followed by endo laser retinopexy to surround the border of retinectomy. 5000 cs silicone oil was used in 1 eye, whereas 12% of perfluoropropane gas was used in the other eye as a tamponade. These procedures were performed with meticulous attention to prevent any fluid from entering the subretinal space. Postoperative examination showed attachment of the retina and improvement of the patient’s visual acuity. Both eyes’ intraocular pressure was in the normal range. One eye developed retinal displacement, but no other complications occurred. Discussion: Rhegmatogenous retinal detachment with advanced PVR is a complex situation for vitreoretinal surgeons. PVR is characterized by the growth and migration of preretinal or subretinal membranes. PVR is the most common cause of retinal reattachment failure. A 360-degree retinectomy is an alternative surgical method to overcome this condition. Objectives of this procedure are releasing retinal traction caused by PVR, reducing the recurrence rate of PVR, and reattaching the retina to the pigment epithelial surface. Conclusion: 360-degree retinectomy provides satisfactory retinal reattachment and visual outcome improvement in rhegmatogenous retinal detachment with advanced PVR.

Keywords: RRD, retinectomy, pars plana, advanced PVR

Procedia PDF Downloads 51
26195 Nuclear Decay Data Evaluation for 217Po

Authors: S. S. Nafee, A. M. Al-Ramady, S. A. Shaheen

Abstract:

Evaluated nuclear decay data for the 217Po nuclide ispresented in the present work. These data include recommended values for the half-life T1/2, α-, β--, and γ-ray emission energies and probabilities. Decay data from 221Rn α and 217Bi β—decays are presented. Q(α) has been updated based on the recent published work of the Atomic Mass Evaluation AME2012. In addition, the logft values were calculated using the Logft program from the ENSDF evaluation package. Moreover, the total internal conversion electrons has been calculated using Bricc program. Meanwhile, recommendation values or the multi-polarities have been assigned based on recently measurement yield a better intensity balance at the 254 keV and 264 keV gamma transitions.

Keywords: nuclear decay data evaluation, mass evaluation, total converison coefficients, atomic mass evaluation

Procedia PDF Downloads 436
26194 Understanding Tactical Urbanisms in Derelict Areas

Authors: Berna Yaylalı, Isin Can Traunmüller

Abstract:

This paper explores the emergent bottom-up practices in the fields of architecture and urban design within comparative perspectives of two cities. As a temporary, easily affordable intervention that gives the possibility of transforming neglected spaces into vibrant public spaces, tactical urbanism, together with creative place-making strategies, presents alternative ways of creating sustainable developments in derelict and underused areas. This study examines the potential of social and physical developments through a reading of case studies of two creative spatial practices: a pop-up garden transformed from an unused derelict space in Favoriten, Vienna, and an urban community garden in Kuzguncuk, Istanbul. Two cities are chosen according to their multicultural population and diversity. Istanbul was selected as a design city by UNESCO Creative Cities Network in 2017, and Vienna was declared an open and livable city by its local government. This research will use media archives and reports, interviews with locals and local governments, site observations, and visual recordings as methods to provide a critical reading on creative public spaces from the view of local users in these neighborhoods. Reflecting on these emergent ways, this study aims at discussing the production process of tactile urbanism with the practices of locals and the decision-making process with cases from İstanbul and Vienna. The comparison between their place-making strategies in tactical urbanism will give important insights for future developments.

Keywords: creative city, tactical urbanism, neglected area, public space

Procedia PDF Downloads 108
26193 Uncovering Geometrical Ideas in Weaving: An Ethnomathematical Approaches to School Pedagogy

Authors: Jaya Bishnu Pradhan

Abstract:

Weaving mat is one of the common activities performed in different community generally in the rural part of Nepal. Mat weavers’ practice mathematical ideas and concepts implicitly in order to perform their job. This study is intended to uncover the mathematical ideas embedded in mat weaving that can help teachers and students for the teaching and learning of school geometry. The ethnographic methodology was used to uncover and describe the beliefs, values, understanding, perceptions, and attitudes of the mat weavers towards mathematical ideas and concepts in the process of mat weaving. A total of 4 mat weavers, two mathematics teachers and 12 students from grade level 6-8, who are used to participate in weaving, were selected for the study. The whole process of the mat weaving was observed in a natural setting. The classroom observation and in-depth interview were taken with the participants with the help of interview guidelines and observation checklist. The data obtained from the field were categorized according to the themes regarding mathematical ideas embedded in the weaving activities, and its possibilities in teaching learning of school geometry. In this study, the mathematical activities in different sectors of their lives, their ways of understanding the natural phenomena, and their ethnomathematical knowledge were analyzed with the notions of pluralism. From the field data, it was found that the mat weaver exhibited sophisticated geometrical ideas in the process of construction of frame of mat. They used x-test method for confirming if the mat is rectangular. Mat also provides a good opportunity to understand the space geometry. A rectangular form of mat may be rolled up when it is not in use and can be converted to a cylindrical form, which usually can be used as larder so as to reserve food grains. From the observation of the situations, this cultural experience enables students to calculate volume, curved surface area and total surface area of the cylinder. The possibilities of incorporation of these cultural activities and its pedagogical use were observed in mathematics classroom. It is argued that it is possible to use mat weaving activities in the teaching and learning of school geometry.

Keywords: ethnography, ethnomathematics, geometry, mat weaving, school pedagogy

Procedia PDF Downloads 161
26192 Geographic Information System Using Google Fusion Table Technology for the Delivery of Disease Data Information

Authors: I. Nyoman Mahayasa Adiputra

Abstract:

Data in the field of health can be useful for the purposes of data analysis, one example of health data is disease data. Disease data is usually in a geographical plot in accordance with the area. Where the data was collected, in the city of Denpasar, Bali. Disease data report is still published in tabular form, disease information has not been mapped in GIS form. In this research, disease information in Denpasar city will be digitized in the form of a geographic information system with the smallest administrative area in the form of district. Denpasar City consists of 4 districts of North Denpasar, East Denpasar, West Denpasar and South Denpasar. In this research, we use Google fusion table technology for map digitization process, where this technology can facilitate from the administrator and from the recipient information. From the administrator side of the input disease, data can be done easily and quickly. From the receiving end of the information, the resulting GIS application can be published in a website-based application so that it can be accessed anywhere and anytime. In general, the results obtained in this study, divided into two, namely: (1) Geolocation of Denpasar and all of Denpasar districts, the process of digitizing the map of Denpasar city produces a polygon geolocation of each - district of Denpasar city. These results can be utilized in subsequent GIS studies if you want to use the same administrative area. (2) Dengue fever mapping in 2014 and 2015. Disease data used in this study is dengue fever case data taken in 2014 and 2015. Data taken from the profile report Denpasar Health Department 2015 and 2016. This mapping can be useful for the analysis of the spread of dengue hemorrhagic fever in the city of Denpasar.

Keywords: geographic information system, Google fusion table technology, delivery of disease data information, Denpasar city

Procedia PDF Downloads 135
26191 Inclusive Practices in Health Sciences: Equity Proofing Higher Education Programs

Authors: Mitzi S. Brammer

Abstract:

Given that the cultural make-up of programs of study in institutions of higher learning is becoming increasingly diverse, much has been written about cultural diversity from a university-level perspective. However, there are little data in the way of specific programs and how they address inclusive practices when teaching and working with marginalized populations. This research study aimed to discover baseline knowledge and attitudes of health sciences faculty, instructional staff, and students related to inclusive teaching/learning and interactions. Quantitative data were collected via an anonymous online survey (one designed for students and another designed for faculty/instructional staff) using a web-based program called Qualtrics. Quantitative data were analyzed amongst the faculty/instructional staff and students, respectively, using descriptive and comparative statistics (t-tests). Additionally, some participants voluntarily engaged in a focus group discussion in which qualitative data were collected around these same variables. Collecting qualitative data to triangulate the quantitative data added trustworthiness to the overall data. The research team analyzed collected data and compared identified categories and trends, comparing those data between faculty/staff and students, and reported results as well as implications for future study and professional practice.

Keywords: inclusion, higher education, pedagogy, equity, diversity

Procedia PDF Downloads 69
26190 The Study of the Socio-Economic and Environmental Impact on the Semi-Arid Environments Using GIS in the Eastern Aurès, Algeria

Authors: Benmessaoud Hassen

Abstract:

We propose in this study to address the impact of socio-economic and environmental impact on the physical environment, especially their spatiotemporal dynamics in semi-arid and arid eastern Aurès. Including 11 municipalities, the study area spreads out over a relatively large surface area of about 60.000 ha. The hindsight is quite important and is determined by 03 days of analysis of environmental variation spread over thirty years (between 1987 and 2007). The multi-source data acquired in this context are integrated into a geographic information system (GIS).This allows, among other indices to calculate areas and classes for each thematic layer of the 4 layers previously defined by a method inspired MEDALUS (Mediterranean Desertification and Land Use).The database created is composed of four layers of information (population, livestock, farming and land use). His analysis in space and time has been supplemented by a validation of the ground truth. Once the database has corrected it used to develop the comprehensive map with the calculation of the index of socio-economic and environmental (ISCE). The map supports and the resulting information does not consist only of figures on the present situation but could be used to forecast future trends.

Keywords: impact of socio-economic and environmental, spatiotemporal dynamics, semi-arid environments, GIS, Eastern Aurès

Procedia PDF Downloads 327