Search results for: spatio-temporal data
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 24452

Search results for: spatio-temporal data

17972 A Geoprocessing Tool for Early Civil Work Notification to Optimize Fiber Optic Cable Installation Cost

Authors: Hussain Adnan Alsalman, Khalid Alhajri, Humoud Alrashidi, Abdulkareem Almakrami, Badie Alguwaisem, Said Alshahrani, Abdullah Alrowaished

Abstract:

Most of the cost of installing a new fiber optic cable is attributed to civil work-trenching-cost. In many cases, information technology departments receive project proposals in their eReview system, but not all projects are visible to everyone. Additionally, if there was no IT scope in the proposed project, it is not likely to be visible to IT. Sometimes it is too late to add IT scope after project budgets have been finalized. Finally, the eReview system is a repository of PDF files for each project, which commits the reviewer to manual work and limits automation potential. This paper details a solution to address the late notification of the eReview system by integrating IT Sites GIS data-sites locations-with land use permit (LUP) data-civil work activity, which is the first step before securing the required land usage authorizations and means no detailed designs for any relevant project before an approved LUP request. To address the manual nature of eReview system, both the LUP System and IT data are using ArcGIS Desktop, which enables the creation of a geoprocessing tool with either Python or Model Builder to automate finding and evaluating potentially usable LUP requests to reduce trenching between two sites in need of a new FOC. To achieve this, a weekly dump was taken from LUP system production data and loaded manually onto ArcMap Desktop. Then a custom tool was developed in model builder, which consisted of a table of two columns containing all the pairs of sites in need of new fiber connectivity. The tool then iterates all rows of this table, taking the sites’ pair one at a time and finding potential LUPs between them, which satisfies the provided search radius. If a group of LUPs was found, an iterator would go through each LUP to find the required civil work between the two sites and the LUP Polyline feature and the distance through the line, which would be counted as cost avoidance if an IT scope had been added. Finally, the tool will export an Excel file named with sites pair, and it will contain as many rows as the number of LUPs, which met the search radius containing trenching and pulling information and cost. As a result, multiple projects have been identified – historical, missed opportunity, and proposed projects. For the proposed project, the savings were about 75% ($750,000) to install a new fiber with the Euclidean distance between Abqaiq GOSP2 and GOSP3 DCOs. In conclusion, the current tool setup identifies opportunities to bundle civil work on single projects at a time and between two sites. More work is needed to allow the bundling of multiple projects between two sites to achieve even more cost avoidance in both capital cost and carbon footprint.

Keywords: GIS, fiber optic cable installation optimization, eliminate redundant civil work, reduce carbon footprint for fiber optic cable installation

Procedia PDF Downloads 203
17971 The “Bright Side” of COVID-19: Effects of Livestream Affordances on Consumer Purchase Willingness: Explicit IT Affordances Perspective

Authors: Isaac Owusu Asante, Yushi Jiang, Hailin Tao

Abstract:

Live streaming marketing, the new electronic commerce element, became an optional marketing channel following the COVID-19 pandemic. Many sellers have leveraged the features presented by live streaming to increase sales. Studies on live streaming have focused on gaming and consumers’ loyalty to brands through live streaming, using interview questionnaires. This study, however, was conducted to measure real-time observable interactions between consumers and sellers. Based on the affordance theory, this study conceptualized constructs representing the interactive features and examined how they drive consumers’ purchase willingness during live streaming sessions using 1238 datasets from Amazon Live, following the manual observation of transaction records. Using structural equation modeling, the ordinary least square regression suggests that live viewers, new followers, live chats, and likes positively affect purchase willingness. The Sobel and Monte Carlo tests show that new followers, live chats, and likes significantly mediate the relationship between live viewers and purchase willingness. The study introduces a new way of measuring interactions in live streaming commerce and proposes a way to manually gather data on consumer behaviors in live streaming platforms when the application programming interface (API) of such platforms does not support data mining algorithms.

Keywords: livestreaming marketing, live chats, live viewers, likes, new followers, purchase willingness

Procedia PDF Downloads 56
17970 Human Resource Management Challenges in Age of Artificial Intelligence: Methodology of Case Analysis

Authors: Olga Leontjeva

Abstract:

In the age of Artificial Intelligence (AI), some organization management approaches need to be adapted or changed. Human Resource Management (HRM) is a part of organization management that is under the managers' focus nowadays, because AI integration into organization activities brings some HRM-connected challenges. The topic became more significant during the crises of many organizations in the world caused by the coronavirus pandemic (COVID-19). The paper presents an approach, which will be used for the study that is going to be focused on the various case analysis. The author of the future study will analyze the cases of the organizations from Latvia and Spain that are grouped by the size, type of activity and area of business. The information for the cases will be collected through structured interviews and online surveys. The main result presented is the questionnaire developed that will be used for the study as well as the definition and description of sampling. The first round of the survey will be based on convenience sampling that is the main limitation of the study. To conclude, the approach developed will help to collect valid data if the organizations participating in the survey are ready to share their cases in depth, so the researchers could draw the right conclusions and generalize compared organizations’ cases. The questionnaire developed for the survey is applicable for both written online data collection as well as for the interviews. The case analysis will help to identify some HRM challenges that are connected to AI integration into organization activities such as management of different generation employees and their training peculiarities.

Keywords: age of artificial intelligence, case analysis, generation Y and Z employees, human resource management

Procedia PDF Downloads 152
17969 Lived Experiences and Perspectives of Adult Survivors of Incest-Related Childhood Sexual Abuse

Authors: Varsha Puri, Sharon Hudson, Ian Kim

Abstract:

Background: Incest-related childhood sexual abuse (IRCSA) is challenging to study due to the shame and secrecy experienced by its survivors. Ramifications of IRCSA worsen when it is unidentified, and interventions are not made. IRCSA perspectives are essential for future prevention and intervention strategies. However, there is limited understanding of this population’s experiences, perspectives, and long-term struggles. To date, research for IRCSA has utilized data from treatment programs and qualitative research with cohorts of 10-20 people, much of the data is from 10-40 years prior. Methods. In June 2018, an anonymous online survey was posted to multiple social media sites (e.g., Facebook IRCSA groups) and sexual abuse resource sites. Survey responses were collected for a year. The survey collected non-identifying demographics, IRCSA experiences, and outcomes data. Results: We obtained 1310 completed surveys. Demographics of all ages, racial backgrounds, financial backgrounds, and genders were obtained; the majority identified as white (81%) and female (76%). Childhood sexual abuse (CSA) started before the age of 6 in 49% and was endured for more than one year in 84% of respondents, and 39% reported ten or more years of abuse. CSA by multiple perpetrators occurred in 58%, while 8% had ten or more perpetrators. CSA by perpetrators under 21 years old was reported by 46%. Female perpetrators were reported by 28% of respondents. Fathers were the highest reported sexual abusers at 60%, and mothers were reported at 17%. Only 16% reported that at least one of their perpetrators was prosecuted for sexual abuse of a minor. Respondents confirmed that 54% of the time, they informed an adult of the abuse; only 2% agreed that “an intervention was made by the family that protected me.” A majority reported that IRCSA has negatively impacted their intimate/sexual relationships (96%) and mental health (96%). A majority reported negative impacts on biological family relationships (88%), physical health (73%), finances (59%), educational achievement (57%), and employment (56%). When asked about suffering from addiction, 85% of respondents answered yes. Prevention strategies selected most by respondents include early school education around CSA prevention (67%), removing the statute of limitations for reporting CSA (69%), and improved laws protecting IRCSA survivors (63%). Conclusion: The data document that IRCSA can be pervasive, and the dearth of intervention and support for survivors have major lasting impacts. Survivors have a unique and valuable perspective on what interventions are needed to prevent IRCSA and support survivors; their voice has long been unheard in crafting prevention and intervention policies and services. These results thus provide an important call to action from these critical stakeholders. Pediatricians should recognize that perpetrators can be pediatric patients, women, and parents. Pediatricians can advocate for more early CSA prevention education and policy changes that remove the statute of limitations for reporting CSA.

Keywords: incest, childhood sexual abuse, incest-related childhood sexual abuse, incest survivor

Procedia PDF Downloads 76
17968 Spiritual Symbols of African Fruits as Responsive Catalysts for Naturopathy

Authors: Orogun Daniel Oghenekevhwe

Abstract:

Africa being an agrarian continent has an abundance of fruits that are both nutritional and medicinal. Regardless of the abundance of these healing elements, Africa leads the statistics of poor healthcare globally. Among others, there are two noticeable challenges in the healthcare system which are ‘Poor access and high cost of medical healthcare’. The effects of both the access and economic implications are (1) Low responsiveness and (2) High mortality rate. While the United Nations and the global health community continue to work towards reduced mortality rates and poor responsiveness to healthcare and wellness, this paper investigates how some Africans use the spiritual symbols of African fruits as responsive catalysts to embrace naturopathy thereby reducing the effects and impacts of poor healthcare challenges in Africa. The main argument is whether there are links between spiritual symbols and fruits that influence Africans' response to naturopathy and low-cost healthcare. Following that is the question of how medical healthcare responds to such development. Bitter Kola (Garcinia) is the case study fruit, and Sunnyside in Pretoria, South Africa, has been spotted as one of the high-traffic selling points of herbal fruits. A mixed research method is applicable with an expected 20 Quantitative data respondents among sellers and nutritionists and 50 Qualitative Data respondents among consumers. Based on the results, it should be clear how spirituality contributes to alternative healthcare and how it can be further encouraged to bridge the gap between the high demand and low supply of healthcare in Africa and beyond.

Keywords: spiritual symbols, naturopathy, African fruits, spirituality, healthcare

Procedia PDF Downloads 43
17967 Access to the Community and Needed Supports among People with Physical Disabilities Receiving Long-Term Services and Supports in the United States

Authors: Stephanie Giordano, Eric Lam, Rosa Plasencia

Abstract:

An important piece of active aging is ensuring people have the right support to meet individual needs. Using NCI-AD data, we will look at measures of satisfaction with community access and needed services among people with physical disabilities receiving LTSS in the US. National Core Indicators—Aging and Disabilities (NCI-AD) is a voluntary effort by State Medicaid, aging, and disability agencies across the US to measure and track their own performance. NCI-AD uses a standardized survey – the Adult Consumer Survey (ACS), to hear directly from people receiving services about the quality of services and supports they receive. Data from the 2018-19 ACS found that compared to people without a physical disability, those with a physical disability were more likely to make choices about the services they receive, including when and how often they receive those services. Yet people with a physical disability were less likely to report they get enough assistance with everyday activities (e.g., shopping, housework, and taking medications) and self-care (e.g., dressing or bathing) and more likely to report that services and supports do not fully meet their needs and goals. A further breakdown by age shows that people 40-65 years old with a physical disability experienced even greater barriers to being as active in the community as they would like to be, indicating a need to better support people as they age with or into a disability. We will explore how these and other outcomes were affected by COVID-19, take a closer look at outcomes by demographics (e.g., race/ethnicity, gender, and mental health diagnoses) and discuss implications on the future needs of service systems.

Keywords: quality-of-life, long-term services and supports, person-centered, community

Procedia PDF Downloads 88
17966 On Consolidated Predictive Model of the Natural History of Breast Cancer Considering Primary Tumor and Primary Distant Metastases Growth

Authors: Ella Tyuryumina, Alexey Neznanov

Abstract:

Finding algorithms to predict the growth of tumors has piqued the interest of researchers ever since the early days of cancer research. A number of studies were carried out as an attempt to obtain reliable data on the natural history of breast cancer growth. Mathematical modeling can play a very important role in the prognosis of tumor process of breast cancer. However, mathematical models describe primary tumor growth and metastases growth separately. Consequently, we propose a mathematical growth model for primary tumor and primary metastases which may help to improve predicting accuracy of breast cancer progression using an original mathematical model referred to CoM-IV and corresponding software. We are interested in: 1) modelling the whole natural history of primary tumor and primary metastases; 2) developing adequate and precise CoM-IV which reflects relations between PT and MTS; 3) analyzing the CoM-IV scope of application; 4) implementing the model as a software tool. The CoM-IV is based on exponential tumor growth model and consists of a system of determinate nonlinear and linear equations; corresponds to TNM classification. It allows to calculate different growth periods of primary tumor and primary metastases: 1) ‘non-visible period’ for primary tumor; 2) ‘non-visible period’ for primary metastases; 3) ‘visible period’ for primary metastases. The new predictive tool: 1) is a solid foundation to develop future studies of breast cancer models; 2) does not require any expensive diagnostic tests; 3) is the first predictor which makes forecast using only current patient data, the others are based on the additional statistical data. Thus, the CoM-IV model and predictive software: a) detect different growth periods of primary tumor and primary metastases; b) make forecast of the period of primary metastases appearance; c) have higher average prediction accuracy than the other tools; d) can improve forecasts on survival of BC and facilitate optimization of diagnostic tests. The following are calculated by CoM-IV: the number of doublings for ‘nonvisible’ and ‘visible’ growth period of primary metastases; tumor volume doubling time (days) for ‘nonvisible’ and ‘visible’ growth period of primary metastases. The CoM-IV enables, for the first time, to predict the whole natural history of primary tumor and primary metastases growth on each stage (pT1, pT2, pT3, pT4) relying only on primary tumor sizes. Summarizing: a) CoM-IV describes correctly primary tumor and primary distant metastases growth of IV (T1-4N0-3M1) stage with (N1-3) or without regional metastases in lymph nodes (N0); b) facilitates the understanding of the appearance period and manifestation of primary metastases.

Keywords: breast cancer, exponential growth model, mathematical modelling, primary metastases, primary tumor, survival

Procedia PDF Downloads 318
17965 Entrepreneurship, Institutional Quality, and Macroeconomic Performance: Evidence from Nigeria

Authors: Cleopatra Oluseye Ibukun

Abstract:

Following the endogenous growth theory, entrepreneurship has been considered pivotal to economic growth and development, particularly in developing countries like Nigeria. Meanwhile, efforts to reduce unemployment has yielded minimal result with over 36% of youth unemployment and a dwindling economic growth despite the country’s natural and human resource endowment. This study, therefore, investigates the effects of entrepreneurship and institutional quality on economic growth and unemployment in Nigeria over the period 1996 to 2018. The data is obtained from the National Bureau of Statistics (NBS), World Bank’s World Development Indicators (WDI), and the World Bank’s World Governance Indicators (WGI). The study period is guided by the availability of data, and the study employs both descriptive and econometric techniques of analysis (specifically, the Auto-regressive Distributed Lag Approach). This approach is preferable given that the variables are stationary at the first difference, while the bounds test suggests the existence of co-integration among the variables. By implication, an increase in entrepreneurship significantly improves economic growth, and it reduces unemployment in both the short-run and the long-run. Besides, institutional quality proxied by the control of corruption, political stability, and government effectiveness significantly mediates the interaction between entrepreneurship and macroeconomic performance. This study concludes that improved institutional quality enhances the effect of entrepreneurship on economic growth and unemployment in Nigeria, and it recommends an improvement in Nigeria’s institutional quality because it can jeopardise or augment the effect of entrepreneurship on macroeconomic performance.

Keywords: entrepreneurship, institutional quality, unemployment, gross domestic product, Nigeria

Procedia PDF Downloads 111
17964 A Machine Learning Approach for Earthquake Prediction in Various Zones Based on Solar Activity

Authors: Viacheslav Shkuratskyy, Aminu Bello Usman, Michael O’Dea, Saifur Rahman Sabuj

Abstract:

This paper examines relationships between solar activity and earthquakes; it applied machine learning techniques: K-nearest neighbour, support vector regression, random forest regression, and long short-term memory network. Data from the SILSO World Data Center, the NOAA National Center, the GOES satellite, NASA OMNIWeb, and the United States Geological Survey were used for the experiment. The 23rd and 24th solar cycles, daily sunspot number, solar wind velocity, proton density, and proton temperature were all included in the dataset. The study also examined sunspots, solar wind, and solar flares, which all reflect solar activity and earthquake frequency distribution by magnitude and depth. The findings showed that the long short-term memory network model predicts earthquakes more correctly than the other models applied in the study, and solar activity is more likely to affect earthquakes of lower magnitude and shallow depth than earthquakes of magnitude 5.5 or larger with intermediate depth and deep depth.

Keywords: k-nearest neighbour, support vector regression, random forest regression, long short-term memory network, earthquakes, solar activity, sunspot number, solar wind, solar flares

Procedia PDF Downloads 54
17963 Numerical Simulation of a Point Absorber Wave Energy Converter Using OpenFOAM in Indian Scenario

Authors: Pooja Verma, Sumana Ghosh

Abstract:

There is a growing need for alternative way of power generation worldwide. The reason can be attributed to limited resources of fossil fuels, environmental pollution, increasing cost of conventional fuels, and lower efficiency of conversion of energy in existing systems. In this context, one of the potential alternatives for power generation is wave energy. However, it is difficult to estimate the amount of electrical energy generation in an irregular sea condition by experiment and or analytical methods. Therefore in this work, a numerical wave tank is developed using the computational fluid dynamics software Open FOAM. In this software a specific utility known as waves2Foam utility is being used to carry out the simulation work. The computational domain is a tank of dimension: 5m*1.5m*1m with a floating object of dimension: 0.5m*0.2m*0.2m. Regular waves are generated at the inlet of the wave tank according to Stokes second order theory. The main objective of the present study is to validate the numerical model against existing experimental data. It shows a good matching with the existing experimental data of floater displacement. Later the model is exploited to estimate energy extraction due to the movement of such a point absorber in real sea conditions. Scale down the wave properties like wave height, wave length, etc. are used as input parameters. Seasonal variations are also considered.

Keywords: OpenFOAM, numerical wave tank, regular waves, floating object, point absorber

Procedia PDF Downloads 340
17962 Role of Pulp Volume Method in Assessment of Age and Gender in Lucknow, India, an Observational Study

Authors: Anurag Tripathi, Sanad Khandelwal

Abstract:

Age and gender determination are required in forensic for victim identification. There is secondary dentine deposition throughout life, resulting in decreased pulp volume and size. Evaluation of pulp volume using Cone Beam Computed Tomography (CBCT)is a noninvasive method to evaluate the age and gender of an individual. The study was done to evaluate the efficacy of pulp volume method in the determination of age and gender.Aims/Objectives: The study was conducted to estimate age and determine sex by measuring tooth pulp volume with the help of CBCT. An observational study of one year duration on CBCT data of individuals was conducted in Lucknow. Maxillary central incisors (CI) and maxillary canine (C) of the randomly selected samples were assessed for measurement of pulp volume using a software. Statistical analysis: Chi Square Test, Arithmetic Mean, Standard deviation, Pearson’s Correlation, Linear & Logistic regression analysis. Results: The CBCT data of Ninety individuals with age range between 18-70 years was evaluated for pulp volume of central incisor and canine (CI & C). The Pearson correlation coefficient between the tooth pulp volume (CI & C) and chronological age suggested that pulp volume decreased with age. The validation of the equations for sex determination showed higher prediction accuracy for CI (56.70%) and lower for C (53.30%).Conclusion: Pulp volume obtained from CBCT is a reliable indicator for age estimation and gender prediction.

Keywords: forensic, dental age, pulp volume, cone beam computed tomography

Procedia PDF Downloads 81
17961 Variability of Hydrological Modeling of the Blue Nile

Authors: Abeer Samy, Oliver C. Saavedra Valeriano, Abdelazim Negm

Abstract:

The Blue Nile Basin is the most important tributary of the Nile River. Egypt and Sudan are almost dependent on water originated from the Blue Nile. This multi-dependency creates conflicts among the three countries Egypt, Sudan, and Ethiopia making the management of these conflicts as an international issue. Good assessment of the water resources of the Blue Nile is an important to help in managing such conflicts. Hydrological models are good tool for such assessment. This paper presents a critical review of the nature and variability of the climate and hydrology of the Blue Nile Basin as a first step of using hydrological modeling to assess the water resources of the Blue Nile. Many several attempts are done to develop basin-scale hydrological modeling on the Blue Nile. Lumped and semi distributed models used averages of meteorological inputs and watershed characteristics in hydrological simulation, to analyze runoff for flood control and water resource management. Distributed models include the temporal and spatial variability of catchment conditions and meteorological inputs to allow better representation of the hydrological process. The main challenge of all used models was to assess the water resources of the basin is the shortage of the data needed for models calibration and validation. It is recommended to use distributed model for their higher accuracy to cope with the great variability and complexity of the Blue Nile basin and to collect sufficient data to have more sophisticated and accurate hydrological modeling.

Keywords: Blue Nile Basin, climate change, hydrological modeling, watershed

Procedia PDF Downloads 341
17960 Forecasting Thermal Energy Demand in District Heating and Cooling Systems Using Long Short-Term Memory Neural Networks

Authors: Kostas Kouvaris, Anastasia Eleftheriou, Georgios A. Sarantitis, Apostolos Chondronasios

Abstract:

To achieve the objective of almost zero carbon energy solutions by 2050, the EU needs to accelerate the development of integrated, highly efficient and environmentally friendly solutions. In this direction, district heating and cooling (DHC) emerges as a viable and more efficient alternative to conventional, decentralized heating and cooling systems, enabling a combination of more efficient renewable and competitive energy supplies. In this paper, we develop a forecasting tool for near real-time local weather and thermal energy demand predictions for an entire DHC network. In this fashion, we are able to extend the functionality and to improve the energy efficiency of the DHC network by predicting and adjusting the heat load that is distributed from the heat generation plant to the connected buildings by the heat pipe network. Two case-studies are considered; one for Vransko, Slovenia and one for Montpellier, France. The data consists of i) local weather data, such as humidity, temperature, and precipitation, ii) weather forecast data, such as the outdoor temperature and iii) DHC operational parameters, such as the mass flow rate, supply and return temperature. The external temperature is found to be the most important energy-related variable for space conditioning, and thus it is used as an external parameter for the energy demand models. For the development of the forecasting tool, we use state-of-the-art deep neural networks and more specifically, recurrent networks with long-short-term memory cells, which are able to capture complex non-linear relations among temporal variables. Firstly, we develop models to forecast outdoor temperatures for the next 24 hours using local weather data for each case-study. Subsequently, we develop models to forecast thermal demand for the same period, taking under consideration past energy demand values as well as the predicted temperature values from the weather forecasting models. The contributions to the scientific and industrial community are three-fold, and the empirical results are highly encouraging. First, we are able to predict future thermal demand levels for the two locations under consideration with minimal errors. Second, we examine the impact of the outdoor temperature on the predictive ability of the models and how the accuracy of the energy demand forecasts decreases with the forecast horizon. Third, we extend the relevant literature with a new dataset of thermal demand and examine the performance and applicability of machine learning techniques to solve real-world problems. Overall, the solution proposed in this paper is in accordance with EU targets, providing an automated smart energy management system, decreasing human errors and reducing excessive energy production.

Keywords: machine learning, LSTMs, district heating and cooling system, thermal demand

Procedia PDF Downloads 123
17959 Preservice EFL Teachers in a Blended Professional Development Program: Learning to Teach Speech Acts

Authors: Mei-Hui Liu

Abstract:

This study examines the effectiveness of a blended professional development program on preservice EFL (English as a foreign language) teachers’ learning to teach speech acts with the advent of Information and Communication Technology, researchers and scholars underscore the significance of integrating online and face-to-face learning opportunities in the teacher education field. Yet, a paucity of evidence has been documented to investigate the extent to which such a blended professional learning model may impact real classroom practice and student learning outcome. This yearlong project involves various stakeholders, including 25 preservice teachers, 5 English professionals, and 45 secondary school students. Multiple data sources collected are surveys, interviews, reflection journals, online discussion messages, artifacts, and discourse completion tests. Relying on the theoretical lenses of Community of Inquiry, data analysis depicts the nature and process of preservice teachers’ professional development in this blended learning community, which triggers and fosters both face-to-face and synchronous/asynchronous online interactions among preservice teachers and English professionals (i.e., university faculty and in-service teachers). Also included is the student learning outcome after preservice teachers put what they learn from the support community into instructional practice. Pedagogical implications and research suggestions are further provided based on the research findings and limitations.

Keywords: blended professional development, preservice EFL teachers, speech act instruction, student learning outcome

Procedia PDF Downloads 201
17958 Times2D: A Time-Frequency Method for Time Series Forecasting

Authors: Reza Nematirad, Anil Pahwa, Balasubramaniam Natarajan

Abstract:

Time series data consist of successive data points collected over a period of time. Accurate prediction of future values is essential for informed decision-making in several real-world applications, including electricity load demand forecasting, lifetime estimation of industrial machinery, traffic planning, weather prediction, and the stock market. Due to their critical relevance and wide application, there has been considerable interest in time series forecasting in recent years. However, the proliferation of sensors and IoT devices, real-time monitoring systems, and high-frequency trading data introduce significant intricate temporal variations, rapid changes, noise, and non-linearities, making time series forecasting more challenging. Classical methods such as Autoregressive integrated moving average (ARIMA) and Exponential Smoothing aim to extract pre-defined temporal variations, such as trends and seasonality. While these methods are effective for capturing well-defined seasonal patterns and trends, they often struggle with more complex, non-linear patterns present in real-world time series data. In recent years, deep learning has made significant contributions to time series forecasting. Recurrent Neural Networks (RNNs) and their variants, such as Long short-term memory (LSTMs) and Gated Recurrent Units (GRUs), have been widely adopted for modeling sequential data. However, they often suffer from the locality, making it difficult to capture local trends and rapid fluctuations. Convolutional Neural Networks (CNNs), particularly Temporal Convolutional Networks (TCNs), leverage convolutional layers to capture temporal dependencies by applying convolutional filters along the temporal dimension. Despite their advantages, TCNs struggle with capturing relationships between distant time points due to the locality of one-dimensional convolution kernels. Transformers have revolutionized time series forecasting with their powerful attention mechanisms, effectively capturing long-term dependencies and relationships between distant time points. However, the attention mechanism may struggle to discern dependencies directly from scattered time points due to intricate temporal patterns. Lastly, Multi-Layer Perceptrons (MLPs) have also been employed, with models like N-BEATS and LightTS demonstrating success. Despite this, MLPs often face high volatility and computational complexity challenges in long-horizon forecasting. To address intricate temporal variations in time series data, this study introduces Times2D, a novel framework that parallelly integrates 2D spectrogram and derivative heatmap techniques. The spectrogram focuses on the frequency domain, capturing periodicity, while the derivative patterns emphasize the time domain, highlighting sharp fluctuations and turning points. This 2D transformation enables the utilization of powerful computer vision techniques to capture various intricate temporal variations. To evaluate the performance of Times2D, extensive experiments were conducted on standard time series datasets and compared with various state-of-the-art algorithms, including DLinear (2023), TimesNet (2023), Non-stationary Transformer (2022), PatchTST (2023), N-HiTS (2023), Crossformer (2023), MICN (2023), LightTS (2022), FEDformer (2022), FiLM (2022), SCINet (2022a), Autoformer (2021), and Informer (2021) under the same modeling conditions. The initial results demonstrated that Times2D achieves consistent state-of-the-art performance in both short-term and long-term forecasting tasks. Furthermore, the generality of the Times2D framework allows it to be applied to various tasks such as time series imputation, clustering, classification, and anomaly detection, offering potential benefits in any domain that involves sequential data analysis.

Keywords: derivative patterns, spectrogram, time series forecasting, times2D, 2D representation

Procedia PDF Downloads 23
17957 The Cleaning Equipment to Prevents Dust Diffusion of Bus Air Filters

Authors: Jiraphorn Satechan, Thanaphon Khamthieng, Warunee Phanwong

Abstract:

This action research aimed at designing and developing the cleaning equipment to preventing dust diffusion of bus air filter. Quantitative and qualitative data collection methods were used to conduct data from October 1st, 2018 to September 30th, 2019. All of participants were male (100.0%) with aged 40- 49 years and 57.15%, of them finish bachelor degree. 71.43% of them was a driver and 57.15% of them had the working experience between 10 and 15 years. Research revealed that the participants assessed the quality of the bus air filter cleaning equipment for preventing dust diffusion at a moderate level (σ= 0.29), and 71.43 of them also suggested the development methods in order to improve the quality of bus air filters cleaning equipment as follows: 1) to install the circuit breaker for cutting the electricity and controlling the on-off of the equipment and to change the motor to the DC system, 2) should install the display monitor for wind pressure and electricity system as well as to install the air pressure gauge, 3) should install the tank lid lock for preventing air leakage and dust diffusion by increasing the blowing force and sucking power, 4) to stabilize the holding points for preventing the filter shaking while rotating and blowing for cleaning and to reduce the rotation speed in order to allow the filters to move slowly for the air system to blow for cleaning more thoroughly, 5) the amount of dust should be measured before and after cleaning and should be designed the cleaning equipment to be able to clean with a variety of filters, and sizes. Moreover, the light-weight materials should be used to build the cleaning equipment and the wheels should be installed at the base of the equipment in order to make it easier to move.

Keywords: Cleaning Equipment, Bus Air Filters, Preventing Dust Diffusion, Innovation

Procedia PDF Downloads 92
17956 PaSA: A Dataset for Patent Sentiment Analysis to Highlight Patent Paragraphs

Authors: Renukswamy Chikkamath, Vishvapalsinhji Ramsinh Parmar, Christoph Hewel, Markus Endres

Abstract:

Given a patent document, identifying distinct semantic annotations is an interesting research aspect. Text annotation helps the patent practitioners such as examiners and patent attorneys to quickly identify the key arguments of any invention, successively providing a timely marking of a patent text. In the process of manual patent analysis, to attain better readability, recognising the semantic information by marking paragraphs is in practice. This semantic annotation process is laborious and time-consuming. To alleviate such a problem, we proposed a dataset to train machine learning algorithms to automate the highlighting process. The contributions of this work are: i) we developed a multi-class dataset of size 150k samples by traversing USPTO patents over a decade, ii) articulated statistics and distributions of data using imperative exploratory data analysis, iii) baseline Machine Learning models are developed to utilize the dataset to address patent paragraph highlighting task, and iv) future path to extend this work using Deep Learning and domain-specific pre-trained language models to develop a tool to highlight is provided. This work assists patent practitioners in highlighting semantic information automatically and aids in creating a sustainable and efficient patent analysis using the aptitude of machine learning.

Keywords: machine learning, patents, patent sentiment analysis, patent information retrieval

Procedia PDF Downloads 71
17955 Valorization of Surveillance Data and Assessment of the Sensitivity of a Surveillance System for an Infectious Disease Using a Capture-Recapture Model

Authors: Jean-Philippe Amat, Timothée Vergne, Aymeric Hans, Bénédicte Ferry, Pascal Hendrikx, Jackie Tapprest, Barbara Dufour, Agnès Leblond

Abstract:

The surveillance of infectious diseases is necessary to describe their occurrence and help the planning, implementation and evaluation of risk mitigation activities. However, the exact number of detected cases may remain unknown whether surveillance is based on serological tests because identifying seroconversion may be difficult. Moreover, incomplete detection of cases or outbreaks is a recurrent issue in the field of disease surveillance. This study addresses these two issues. Using a viral animal disease as an example (equine viral arteritis), the goals were to establish suitable rules for identifying seroconversion in order to estimate the number of cases and outbreaks detected by a surveillance system in France between 2006 and 2013, and to assess the sensitivity of this system by estimating the total number of outbreaks that occurred during this period (including unreported outbreaks) using a capture-recapture model. Data from horses which exhibited at least one positive result in serology using viral neutralization test between 2006 and 2013 were used for analysis (n=1,645). Data consisted of the annual antibody titers and the location of the subjects (towns). A consensus among multidisciplinary experts (specialists in the disease and its laboratory diagnosis, epidemiologists) was reached to consider seroconversion as a change in antibody titer from negative to at least 32 or as a three-fold or greater increase. The number of seroconversions was counted for each town and modeled using a unilist zero-truncated binomial (ZTB) capture-recapture model with R software. The binomial denominator was the number of horses tested in each infected town. Using the defined rules, 239 cases located in 177 towns (outbreaks) were identified from 2006 to 2013. Subsequently, the sensitivity of the surveillance system was estimated as the ratio of the number of detected outbreaks to the total number of outbreaks that occurred (including unreported outbreaks) estimated using the ZTB model. The total number of outbreaks was estimated at 215 (95% credible interval CrI95%: 195-249) and the surveillance sensitivity at 82% (CrI95%: 71-91). The rules proposed for identifying seroconversion may serve future research. Such rules, adjusted to the local environment, could conceivably be applied in other countries with surveillance programs dedicated to this disease. More generally, defining ad hoc algorithms for interpreting the antibody titer could be useful regarding other human and animal diseases and zoonosis when there is a lack of accurate information in the literature about the serological response in naturally infected subjects. This study shows how capture-recapture methods may help to estimate the sensitivity of an imperfect surveillance system and to valorize surveillance data. The sensitivity of the surveillance system of equine viral arteritis is relatively high and supports its relevance to prevent the disease spreading.

Keywords: Bayesian inference, capture-recapture, epidemiology, equine viral arteritis, infectious disease, seroconversion, surveillance

Procedia PDF Downloads 279
17954 Analysis of Ionosphere Anomaly Before Great Earthquake in Java on 2009 Using GPS Tec Data

Authors: Aldilla Damayanti Purnama Ratri, Hendri Subakti, Buldan Muslim

Abstract:

Ionosphere’s anomalies as an effect of earthquake activity is a phenomenon that is now being studied in seismo-ionospheric coupling. Generally, variation in the ionosphere caused by earthquake activity is weaker than the interference generated by different source, such as geomagnetic storms. However, disturbances of geomagnetic storms show a more global behavior, while the seismo-ionospheric anomalies occur only locally in the area which is largely determined by magnitude of the earthquake. It show that the earthquake activity is unique and because of its uniqueness it has been much research done thus expected to give clues as early warning before earthquake. One of the research that has been developed at this time is the approach of seismo-ionospheric-coupling. This study related the state in the lithosphere-atmosphere and ionosphere before and when earthquake occur. This paper choose the total electron content in a vertical (VTEC) in the ionosphere as a parameter. Total Electron Content (TEC) is defined as the amount of electron in vertical column (cylinder) with cross-section of 1m2 along GPS signal trajectory in ionosphere at around 350 km of height. Based on the analysis of data obtained from the LAPAN agency to identify abnormal signals by statistical methods, obtained that there are an anomaly in the ionosphere is characterized by decreasing of electron content of the ionosphere at 1 TECU before the earthquake occurred. Decreasing of VTEC is not associated with magnetic storm that is indicated as an earthquake precursor. This is supported by the Dst index showed no magnetic interference.

Keywords: earthquake, DST Index, ionosphere, seismoionospheric coupling, VTEC

Procedia PDF Downloads 569
17953 Prediction of Marijuana Use among Iranian Early Youth: an Application of Integrative Model of Behavioral Prediction

Authors: Mehdi Mirzaei Alavijeh, Farzad Jalilian

Abstract:

Background: Marijuana is the most widely used illicit drug worldwide, especially among adolescents and young adults, which can cause numerous complications. The aim of this study was to determine the pattern, motivation use, and factors related to marijuana use among Iranian youths based on the integrative model of behavioral prediction Methods: A cross-sectional study was conducted among 174 youths marijuana user in Kermanshah County and Isfahan County, during summer 2014 which was selected with the convenience sampling for participation in this study. A self-reporting questionnaire was applied for collecting data. Data were analyzed by SPSS version 21 using bivariate correlations and linear regression statistical tests. Results: The mean marijuana use of respondents was 4.60 times at during week [95% CI: 4.06, 5.15]. Linear regression statistical showed, the structures of integrative model of behavioral prediction accounted for 36% of the variation in the outcome measure of the marijuana use at during week (R2 = 36% & P < 0.001); and among them attitude, marijuana refuse, and subjective norms were a stronger predictors. Conclusion: Comprehensive health education and prevention programs need to emphasize on cognitive factors that predict youth’s health-related behaviors. Based on our findings it seems, designing educational and behavioral intervention for reducing positive belief about marijuana, marijuana self-efficacy refuse promotion and reduce subjective norms encourage marijuana use has an effective potential to protect youths marijuana use.

Keywords: marijuana, youth, integrative model of behavioral prediction, Iran

Procedia PDF Downloads 541
17952 Implications of Measuring the Progress towards Financial Risk Protection Using Varied Survey Instruments: A Case Study of Ghana

Authors: Jemima C. A. Sumboh

Abstract:

Given the urgency and consensus for countries to move towards Universal Health Coverage (UHC), health financing systems need to be accurately and consistently monitored to provide valuable data to inform policy and practice. Most of the indicators for monitoring UHC, particularly catastrophe and impoverishment, are established based on the impact of out-of-pocket health payments (OOPHP) on households’ living standards, collected through varied household surveys. These surveys, however, vary substantially in survey methods such as the length of the recall period or the number of items included in the survey questionnaire or the farming of questions, potentially influencing the level of OOPHP. Using different survey instruments can provide inaccurate, inconsistent, erroneous and misleading estimates of UHC, subsequently influencing wrong policy decisions. Using data from a household budget survey conducted by the Navrongo Health Research Center in Ghana from May 2017 to December 2018, this study intends to explore the potential implications of using surveys with varied levels of disaggregation of OOPHP data on estimates of financial risk protection. The household budget survey, structured around food and non-food expenditure, compared three OOPHP measuring instruments: Version I (existing questions used to measure OOPHP in household budget surveys), Version II (new questions developed through benchmarking the existing Classification of the Individual Consumption by Purpose (COICOP) OOPHP questions in household surveys) and Version III (existing questions used to measure OOPHP in health surveys integrated into household budget surveys- for this, the demographic and health surveillance (DHS) health survey was used). Version I, II and III contained 11, 44, and 56 health items, respectively. However, the choice of recall periods was held constant across versions. The sample size for Version I, II and III were 930, 1032 and 1068 households, respectively. Financial risk protection will be measured based on the catastrophic and impoverishment methodologies using STATA 15 and Adept Software for each version. It is expected that findings from this study will present valuable contributions to the repository of knowledge on standardizing survey instruments to obtain estimates of financial risk protection that are valid and consistent.

Keywords: Ghana, household budget surveys, measuring financial risk protection, out-of-pocket health payments, survey instruments, universal health coverage

Procedia PDF Downloads 114
17951 Detection of Internal Mold Infection of Intact For Tomatoes by Non-Destructive, Transmittance VIS-NIR Spectroscopy

Authors: K. Petcharaporn, N. Prathengjit

Abstract:

The external characteristics of tomatoes, such as freshness, color and size are typically used in quality control processes for tomatoes sorting. However, the internal mold infection of intact tomato cannot be sorted based on a visible assessment and destructive method alone. In this study, a non-destructive technique was used to predict the internal mold infection of intact tomatoes by using transmittance visible and near infrared (VIS-NIR) spectroscopy. Spectra for 200 samples contained 100 samples for normal tomatoes and 100 samples for mold infected tomatoes were acquired in the wavelength range between 665-955 nm. This data was used in conjunction with partial least squares-discriminant analysis (PLS-DA) method to generate a classification model for tomato quality between groups of internal mold infection of intact tomato samples. For this task, the data was split into two groups, 140 samples were used for a training set and 60 samples were used for a test set. The spectra of both normal and internally mold infected tomatoes showed different features in the visible wavelength range. Combined spectral pretreatments of standard normal variate transformation (SNV) and smoothing (Savitzky-Golay) gave the optimal calibration model in training set, 85.0% (63 out of 71 for the normal samples and 56 out of 69 for the internal mold samples). The classification accuracy of the best model on the test set was 91.7% (29 out of 29 for the normal samples and 26 out of 31 for the internal mold tomato samples). The results from this experiment showed that transmittance VIS-NIR spectroscopy can be used as a non-destructive technique to predict the internal mold infection of intact tomatoes.

Keywords: tomato, mold, quality, prediction, transmittance

Procedia PDF Downloads 500
17950 Teamwork of Teachers in Kindergarten and School Heads Implementing Focused Leadership

Authors: Vilma Zydziunaite, Simona Kersiene

Abstract:

The concept of focused leadership means that the leader gathers the entire community in various ways to communicate and cooperate with each other, to share their knowledge and responsibility, to get involved in problem-solving, to create a safe and trusting environment and to satisfy the needs and interests of each community member. The study's aim is to analyze the teamwork of teachers working in kindergartens and schools and its CEOs by implementing confused leadership. A mixed research design was used for the research study. Quantitative research used the teamwork test "Team-Puls" (2003). Data is processed by the IBM SPSS version 29.0 software package. Semi-structured interviews were used for data collection, and qualitative content analysis was applied for data analysis. The results of quantitative research show that there is no statistically significant difference between the evaluation averages of kindergarten and school teachers. Likewise, the effectiveness and evaluation of teacher teamwork in educational institutions depend on different characteristics and processes, such as the number of participating teachers, the involvement of the institution's administration or the stages of team formation. In the qualitative research, the components of the focused leadership categories applied by the kindergarten and school CEOs emerged. The categories reflect the components of shared leadership. In the study, the sharing of responsibilities and cooperation among teachers and the sharing of knowledge among themselves is distinguished. This shows that the action takes place between the teachers when they participate in the processes voluntarily, according to their wishes or for certain reasons. Distributed leadership components occurs when leadership responsibility is extended beyond the school CEO. The components of servant leadership are expressed when the CEO achieves organizational goals in the service of others. Servant leadership is helping and striving for others, creating a safe environment. The level of the educational institution does not affect working teachers in the evaluation of working in a team. Giving freedom to teachers, the role of the CEO is dividing responsibilities and creating cooperation between teachers as well as ensuring teachers' interests, needs, emotional well-being and professional development.

Keywords: teamwork, school, teacher, school CEO, school environment, mixed research, Team-Puls test, semi-structured interview, questioning survey, qualitative content analysis, focused leadership, teacher leadership

Procedia PDF Downloads 43
17949 Public Participation Best Practices in Environmental Decision-making in Newfoundland and Labrador: Analyzing the Forestry Management Planning Process

Authors: Kimberley K. Whyte-Jones

Abstract:

Public participation may improve the quality of environmental management decisions. However, the quality of such a decision is strongly dependent on the quality of the process that leads to it. In order to ensure an effective and efficient process, key features of best practice in participation should be carefully observed; this would also combat disillusionment of citizens, decision-makers and practitioners. The overarching aim of this study is to determine what constitutes an effective public participation process relevant to the Newfoundland and Labrador, Canada context, and to discover whether the public participation process that led to the 2014-2024 Provincial Sustainable Forest Management Strategy (PSFMS) met best practices criteria. The research design uses an exploratory case study strategy to consider a specific participatory process in environmental decision-making in Newfoundland and Labrador. Data collection methods include formal semi-structured interviews and the review of secondary data sources. The results of this study will determine the validity of a specific public participation best practice framework. The findings will be useful for informing citizen participation processes in general and will deduce best practices in public participation in environmental management in the province. The study is, therefore, meaningful for guiding future policies and practices in the management of forest resources in the province of Newfoundland and Labrador, and will help in filling a noticeable gap in research compiling best practices for environmentally related public participation processes.

Keywords: best practices, environmental decision-making, forest management, public participation

Procedia PDF Downloads 294
17948 Evaluating and Reducing Aircraft Technical Delays and Cancellations Impact on Reliability Operational: Case Study of Airline Operator

Authors: Adel A. Ghobbar, Ahmad Bakkar

Abstract:

Although special care is given to maintenance, aircraft systems fail, and these failures cause delays and cancellations. The occurrence of Delays and Cancellations affects operators and manufacturers negatively. To reduce technical delays and cancellations, one should be able to determine the important systems causing them. The goal of this research is to find a method to define the most expensive delays and cancellations systems for Airline operators. A predictive model was introduced to forecast the failure and their impact after carrying out research that identifies relevant information to tackle the problems faced while answering the questions of this paper. Data were obtained from the manufacturers’ services reliability team database. Subsequently, delays and cancellations evaluation methods were identified. No cost estimation methods were used due to their complexity. The model was developed, and it takes into account the frequency of delays and cancellations and uses weighting factors to give an indication of the severity of their duration. The weighting factors are based on customer experience. The data Analysis approach has shown that delays and cancellations events are not seasonal and do not follow any specific trends. The use of weighting factor does have an influence on the shortlist over short periods (Monthly) but not the analyzed period of three years. Landing gear and the navigation system are among the top 3 factors causing delays and cancellations for all three aircraft types. The results did confirm that the cooperation between certain operators and manufacture reduce the impact of delays and cancellations.

Keywords: reliability, availability, delays & cancellations, aircraft maintenance

Procedia PDF Downloads 115
17947 IoT Continuous Monitoring Biochemical Oxygen Demand Wastewater Effluent Quality: Machine Learning Algorithms

Authors: Sergio Celaschi, Henrique Canavarro de Alencar, Claaudecir Biazoli

Abstract:

Effluent quality is of the highest priority for compliance with the permit limits of environmental protection agencies and ensures the protection of their local water system. Of the pollutants monitored, the biochemical oxygen demand (BOD) posed one of the greatest challenges. This work presents a solution for wastewater treatment plants - WWTP’s ability to react to different situations and meet treatment goals. Delayed BOD5 results from the lab take 7 to 8 analysis days, hindered the WWTP’s ability to react to different situations and meet treatment goals. Reducing BOD turnaround time from days to hours is our quest. Such a solution is based on a system of two BOD bioreactors associated with Digital Twin (DT) and Machine Learning (ML) methodologies via an Internet of Things (IoT) platform to monitor and control a WWTP to support decision making. DT is a virtual and dynamic replica of a production process. DT requires the ability to collect and store real-time sensor data related to the operating environment. Furthermore, it integrates and organizes the data on a digital platform and applies analytical models allowing a deeper understanding of the real process to catch sooner anomalies. In our system of continuous time monitoring of the BOD suppressed by the effluent treatment process, the DT algorithm for analyzing the data uses ML on a chemical kinetic parameterized model. The continuous BOD monitoring system, capable of providing results in a fraction of the time required by BOD5 analysis, is composed of two thermally isolated batch bioreactors. Each bioreactor contains input/output access to wastewater sample (influent and effluent), hydraulic conduction tubes, pumps, and valves for batch sample and dilution water, air supply for dissolved oxygen (DO) saturation, cooler/heater for sample thermal stability, optical ODO sensor based on fluorescence quenching, pH, ORP, temperature, and atmospheric pressure sensors, local PLC/CPU for TCP/IP data transmission interface. The dynamic BOD system monitoring range covers 2 mg/L < BOD < 2,000 mg/L. In addition to the BOD monitoring system, there are many other operational WWTP sensors. The CPU data is transmitted/received to/from the digital platform, which in turn performs analyses at periodic intervals, aiming to feed the learning process. BOD bulletins and their credibility intervals are made available in 12-hour intervals to web users. The chemical kinetics ML algorithm is composed of a coupled system of four first-order ordinary differential equations for the molar masses of DO, organic material present in the sample, biomass, and products (CO₂ and H₂O) of the reaction. This system is solved numerically linked to its initial conditions: DO (saturated) and initial products of the kinetic oxidation process; CO₂ = H₂0 = 0. The initial values for organic matter and biomass are estimated by the method of minimization of the mean square deviations. A real case of continuous monitoring of BOD wastewater effluent quality is being conducted by deploying an IoT application on a large wastewater purification system located in S. Paulo, Brazil.

Keywords: effluent treatment, biochemical oxygen demand, continuous monitoring, IoT, machine learning

Procedia PDF Downloads 59
17946 Impact of Integrated Signals for Doing Human Activity Recognition Using Deep Learning Models

Authors: Milagros Jaén-Vargas, Javier García Martínez, Karla Miriam Reyes Leiva, María Fernanda Trujillo-Guerrero, Francisco Fernandes, Sérgio Barroso Gonçalves, Miguel Tavares Silva, Daniel Simões Lopes, José Javier Serrano Olmedo

Abstract:

Human Activity Recognition (HAR) is having a growing impact in creating new applications and is responsible for emerging new technologies. Also, the use of wearable sensors is an important key to exploring the human body's behavior when performing activities. Hence, the use of these dispositive is less invasive and the person is more comfortable. In this study, a database that includes three activities is used. The activities were acquired from inertial measurement unit sensors (IMU) and motion capture systems (MOCAP). The main objective is differentiating the performance from four Deep Learning (DL) models: Deep Neural Network (DNN), Convolutional Neural Network (CNN), Recurrent Neural Network (RNN) and hybrid model Convolutional Neural Network-Long Short-Term Memory (CNN-LSTM), when considering acceleration, velocity and position and evaluate if integrating the IMU acceleration to obtain velocity and position represent an increment in performance when it works as input to the DL models. Moreover, compared with the same type of data provided by the MOCAP system. Despite the acceleration data is cleaned when integrating, results show a minimal increase in accuracy for the integrated signals.

Keywords: HAR, IMU, MOCAP, acceleration, velocity, position, feature maps

Procedia PDF Downloads 78
17945 Frontier Dynamic Tracking in the Field of Urban Plant and Habitat Research: Data Visualization and Analysis Based on Journal Literature

Authors: Shao Qi

Abstract:

The article uses the CiteSpace knowledge graph analysis tool to sort and visualize the journal literature on urban plants and habitats in the Web of Science and China National Knowledge Infrastructure databases. Based on a comprehensive interpretation of the visualization results of various data sources and the description of the intrinsic relationship between high-frequency keywords using knowledge mapping, the research hotspots, processes and evolution trends in this field are analyzed. Relevant case studies are also conducted for the hotspot contents to explore the means of landscape intervention and synthesize the understanding of research theories. The results show that (1) from 1999 to 2022, the research direction of urban plants and habitats gradually changed from focusing on plant and animal extinction and biological invasion to the field of human urban habitat creation, ecological restoration, and ecosystem services. (2) The results of keyword emergence and keyword growth trend analysis show that habitat creation research has shown a rapid and stable growth trend since 2017, and ecological restoration has gained long-term sustained attention since 2004. The hotspots of future research on urban plants and habitats in China may focus on habitat creation and ecological restoration.

Keywords: research trends, visual analysis, habitat creation, ecological restoration

Procedia PDF Downloads 47
17944 A Fresh Look at Tense System of Qashqaie Dialect of Turkish Language

Authors: Mohammad Sharifi Bohlouli

Abstract:

Turkish language with many dialects is native or official language of great number of people all around the world. The Qashqaie dialect of Turkish language is spoken by the Qashqaie tribe mostly scattered in the southern part of Iran. This paper aims at analyzing the tense system of this dialect to detect the type and number of tense and aspects available to its speakers. To collect a reliable data, a group of 50 old native speakers were randomly chosen as the informants and different techniques such as; Shuy et al interviews, selective listening ,and eavesdropping were used. The results of data analysis showed that the tense system in the Qashqaie dialect of Turkish language includes 3 absolute tenses , 6 aspectual , and 2 subjunctive ones. The interesting part of the study is that Qashqaie dialect enables its speakers to make a kind of aspectual opposition through verb structure which seems to be almost impossible through verb forms in any other nonturkish languages. For example in the following examples sentences 1 &2 and 3&4 have the same translation In English although they are different in both meaning and structure. 1. Ali ensha yazirdi. 2. Ali ensha yazirmush. (Ali was writing a composition.) 3. Ali yadmishdi. 4. Ali yadmishimish. ( Ali had slept.) The changes in the verb structure in Qashqaie dialect enables its speakers to say that whether the doer of the action remembers the process of doing the action or not. So, it presents a new aspectual opposition as Observed /nonobserved. The research findings reveal many other regularities and linguistic features that can be useful for linguists interested in Turkish in general and for those interested in tense and aspect and also they can be helpful for different pedagogical purposes including teaching and translating.

Keywords: qashqaie dialect, tense, aspect, linguistics, Turkish Language

Procedia PDF Downloads 344
17943 In-Flight Aircraft Performance Model Enhancement Using Adaptive Lookup Tables

Authors: Georges Ghazi, Magali Gelhaye, Ruxandra Botez

Abstract:

Over the years, the Flight Management System (FMS) has experienced a continuous improvement of its many features, to the point of becoming the pilot’s primary interface for flight planning operation on the airplane. With the assistance of the FMS, the concept of distance and time has been completely revolutionized, providing the crew members with the determination of the optimized route (or flight plan) from the departure airport to the arrival airport. To accomplish this function, the FMS needs an accurate Aircraft Performance Model (APM) of the aircraft. In general, APMs that equipped most modern FMSs are established before the entry into service of an individual aircraft, and results from the combination of a set of ordinary differential equations and a set of performance databases. Unfortunately, an aircraft in service is constantly exposed to dynamic loads that degrade its flight characteristics. These degradations endow two main origins: airframe deterioration (control surfaces rigging, seals missing or damaged, etc.) and engine performance degradation (fuel consumption increase for a given thrust). Thus, after several years of service, the performance databases and the APM associated to a specific aircraft are no longer representative enough of the actual aircraft performance. It is important to monitor the trend of the performance deterioration and correct the uncertainties of the aircraft model in order to improve the accuracy the flight management system predictions. The basis of this research lies in the new ability to continuously update an Aircraft Performance Model (APM) during flight using an adaptive lookup table technique. This methodology was developed and applied to the well-known Cessna Citation X business aircraft. For the purpose of this study, a level D Research Aircraft Flight Simulator (RAFS) was used as a test aircraft. According to Federal Aviation Administration the level D is the highest certification level for the flight dynamics modeling. Basically, using data available in the Flight Crew Operating Manual (FCOM), a first APM describing the variation of the engine fan speed and aircraft fuel flow w.r.t flight conditions was derived. This model was next improved using the proposed methodology. To do that, several cruise flights were performed using the RAFS. An algorithm was developed to frequently sample the aircraft sensors measurements during the flight and compare the model prediction with the actual measurements. Based on these comparisons, a correction was performed on the actual APM in order to minimize the error between the predicted data and the measured data. In this way, as the aircraft flies, the APM will be continuously enhanced, making the FMS more and more precise and the prediction of trajectories more realistic and more reliable. The results obtained are very encouraging. Indeed, using the tables initialized with the FCOM data, only a few iterations were needed to reduce the fuel flow prediction error from an average relative error of 12% to 0.3%. Similarly, the FCOM prediction regarding the engine fan speed was reduced from a maximum error deviation of 5.0% to 0.2% after only ten flights.

Keywords: aircraft performance, cruise, trajectory optimization, adaptive lookup tables, Cessna Citation X

Procedia PDF Downloads 247