Search results for: spatial frequency measurement
992 Predicting Long-Term Performance of Concrete under Sulfate Attack
Authors: Elakneswaran Yogarajah, Toyoharu Nawa, Eiji Owaki
Abstract:
Cement-based materials have been using in various reinforced concrete structural components as well as in nuclear waste repositories. The sulfate attack has been an environmental issue for cement-based materials exposed to sulfate bearing groundwater or soils, and it plays an important role in the durability of concrete structures. The reaction between penetrating sulfate ions and cement hydrates can result in swelling, spalling and cracking of cement matrix in concrete. These processes induce a reduction of mechanical properties and a decrease of service life of an affected structure. It has been identified that the precipitation of secondary sulfate bearing phases such as ettringite, gypsum, and thaumasite can cause the damage. Furthermore, crystallization of soluble salts such as sodium sulfate crystals induces degradation due to formation and phase changes. Crystallization of mirabilite (Na₂SO₄:10H₂O) and thenardite (Na₂SO₄) or their phase changes (mirabilite to thenardite or vice versa) due to temperature or sodium sulfate concentration do not involve any chemical interaction with cement hydrates. Over the past couple of decades, an intensive work has been carried out on sulfate attack in cement-based materials. However, there are several uncertainties still exist regarding the mechanism for the damage of concrete in sulfate environments. In this study, modelling work has been conducted to investigate the chemical degradation of cementitious materials in various sulfate environments. Both internal and external sulfate attack are considered for the simulation. In the internal sulfate attack, hydrate assemblage and pore solution chemistry of co-hydrating Portland cement (PC) and slag mixing with sodium sulfate solution are calculated to determine the degradation of the PC and slag-blended cementitious materials. Pitzer interactions coefficients were used to calculate the activity coefficients of solution chemistry at high ionic strength. The deterioration mechanism of co-hydrating cementitious materials with 25% of Na₂SO₄ by weight is the formation of mirabilite crystals and ettringite. Their formation strongly depends on sodium sulfate concentration and temperature. For the external sulfate attack, the deterioration of various types of cementitious materials under external sulfate ingress is simulated through reactive transport model. The reactive transport model is verified with experimental data in terms of phase assemblage of various cementitious materials with spatial distribution for different sulfate solution. Finally, the reactive transport model is used to predict the long-term performance of cementitious materials exposed to 10% of Na₂SO₄ for 1000 years. The dissolution of cement hydrates and secondary formation of sulfate-bearing products mainly ettringite are the dominant degradation mechanisms, but not the sodium sulfate crystallization.Keywords: thermodynamic calculations, reactive transport, radioactive waste disposal, PHREEQC
Procedia PDF Downloads 163991 Seismic Data Analysis of Intensity, Orientation and Distribution of Fractures in Basement Rocks for Reservoir Characterization
Authors: Mohit Kumar
Abstract:
Natural fractures are classified in two broad categories of joints and faults on the basis of shear movement in the deposited strata. Natural fracture always has high structural relationship with extensional or non-extensional tectonics and sometimes the result is seen in the form of micro cracks. Geological evidences suggest that both large and small-scale fractures help in to analyze the seismic anisotropy which essentially contribute into characterization of petro physical properties behavior associated with directional migration of fluid. We generally question why basement study is much needed as historically it is being treated as non-productive and geoscientist had no interest in exploration of these basement rocks. Basement rock goes under high pressure and temperature, and seems to be highly fractured because of the tectonic stresses that are applied to the formation along with the other geological factors such as depositional trend, internal stress of the rock body, rock rheology, pore fluid and capillary pressure. Sometimes carbonate rocks also plays the role of basement and igneous body e.g basalt deposited over the carbonate rocks and fluid migrate from carbonate to igneous rock due to buoyancy force and adequate permeability generated by fracturing. So in order to analyze the complete petroleum system, FMC (Fluid Migration Characterization) is necessary through fractured media including fracture intensity, orientation and distribution both in basement rock and county rock. Thus good understanding of fractures can lead to project the correct wellbore trajectory or path which passes through potential permeable zone generated through intensified P-T and tectonic stress condition. This paper deals with the analysis of these fracture property such as intensity, orientation and distribution in basement rock as large scale fracture can be interpreted on seismic section, however, small scale fractures show ambiguity in interpretation because fracture in basement rock lies below the seismic wavelength and hence shows erroneous result in identification. Seismic attribute technique also helps us to delineate the seismic fracture and subtle changes in fracture zone and these can be inferred from azimuthal anisotropy in velocity and amplitude and spectral decomposition. Seismic azimuthal anisotropy derives fracture intensity and orientation from compressional wave and converted wave data and based on variation of amplitude or velocity with azimuth. Still detailed analysis of fractured basement required full isotropic and anisotropic analysis of fracture matrix and surrounding rock matrix in order to characterize the spatial variability of basement fracture which support the migration of fluid from basement to overlying rock.Keywords: basement rock, natural fracture, reservoir characterization, seismic attribute
Procedia PDF Downloads 198990 Transcriptomic Analysis of Fragrant Rice Reveals the Involvement of Post-transcriptional Regulation in Response to Zn Foliar Application
Authors: Muhammad Imran, Sarfraz Shafiq, Xiangru Tang
Abstract:
Alternative splicing (AS) is an important post-transcriptional regulatory mechanism to generate transcripts variability and proteome diversity in plants. Fragrant rice (Oryza sativa L.) has a high economic and nutritional value, and the application of micronutrients regulate 2-acetyl-1-pyrroline (2-AP) production, which is responsible for aroma in fragrant rice. However, no systematic investigation of AS events in response to micronutrients (Zn) has been performed in fragrant rice. Furthermore, the post-transcriptional regulation of genes involved in 2-AP biosynthesis is also not known. In this study, a comprehensive analysis of AS events under two gradients of Zn treatment in two different fragrant rice cultivars (Meixiangzhan-2 and Xiangyaxiangzhan) was performed. A total of 386 and 598 significant AS events were found in Meixiangzhan-2 treated with low and high doses of Zn, respectively. In Xiangyaxiangzhan, a total of 449 and 598 significant AS events were found in low and high doses of Zn, respectively. Go analysis indicated that these genes were highly enriched in physiological processes, metabolism, and cellular process in both cultivars. However, genotype and dose-dependent AS events were also detected in both cultivars. By comparing differential AS (DAS) events with differentially expressed genes (DEGs), we found a weak overlap among DAS and DEGs in both fragrant rice cultivars, indicating that only a few genes are post-transcriptionally regulated in response to Zn treatment. We further report that Zn differentially regulates the expression of 2-AP biosynthesis-related genes in both cultivars, and Zn treatment altered the editing frequency of SNPs in the genes involved in 2-AP biosynthesis. Finally, we showed that epigenetic modifications associated with active gene transcription are generally enriched over 2-AP biosynthesis-related genes. Taken together, our results provide evidence of the post-transcriptional gene regulation in fragrant rice in response to Zn treatment and highlight that the 2-AP biosynthesis pathway may also be post-transcriptionally regulated through epigenetic modifications. These findings will serve as a cornerstone for further investigation to understand the molecular mechanisms of 2-AP biosynthesis in fragrant rice.Keywords: fragrant rice, 2-acetyl-1-pyrroline, gene expression, zinc, alternative splicing, SNPs
Procedia PDF Downloads 112989 Characterization of the Music Admission Requirements and Evaluation of the Relationship among Motivation and Performance Achievement
Authors: Antonio M. Oliveira, Patricia Oliveira-Silva, Jose Matias Alves, Gary McPherson
Abstract:
The music teaching is oriented towards offering formal music training. Due to its specificities, this vocational program starts at a very young age. Although provided by the State, the offer is limited to 6 schools throughout the country, which means that the vacancies for prospective students are very limited every year. It is therefore crucial that these vacancies be taken by especially motivated children grown within households that offer the ideal setting for success. Some of the instruments used to evaluate musical performance are highly sensitive to specific previous training, what represents a severe validity problem for testing children who have had restricted opportunities for formal training. Moreover, these practices may be unfair because, for instance, they may not reflect the candidates’ music aptitudes. Based on what constitutes a prerequisite for making an excellent music student, researchers in this field have long argued that motivation, task commitment, and parents’ support are as important as ability. Thus, the aim of this study is: (1) to prepare an inventory of admission requirements in Australia, Portugal and Ireland; (2) to examine whether the candidates to music conservatories and parents’ level of motivation, assessed at three evaluation points (i.e., admission, at the end of the first year, and at the end of the second year), correlates positively with the candidates’ progress in learning a musical instrument (i.e., whether motivation at the admission may predict student musicianship); (3) an adaptation of an existing instrument to assess the motivation (i.e., to adapt the items to the music setting, focusing on the motivation for playing a musical instrument). The inclusion criteria are: only children registered in the administrative services to be evaluated for entrance to the conservatory will be accepted for this study. The expected number of participants is fifty (5-6 years old) in all the three frequency schemes: integrated, articulated and supplementary. Revisiting musical admission procedures is of particular importance and relevance to musical education because this debate may bring guidance and assistance about the needed improvement to make the process of admission fairer and more transparent.Keywords: music learning, music admission requirements, student’s motivation, parent’s motivation
Procedia PDF Downloads 168988 Recurrent Wheezing and Associated Factors among 6-Year-Old Children in Adama Comprehensive Specialized Hospital Medical College
Authors: Samrawit Tamrat Gebretsadik
Abstract:
Recurrent wheezing is a common respiratory symptom among children, often indicative of underlying airway inflammation and hyperreactivity. Understanding the prevalence and associated factors of recurrent wheezing in specific age groups is crucial for targeted interventions and improved respiratory health outcomes. This study aimed to investigate the prevalence and associated factors of recurrent wheezing among 6-year-old children attending Adama Comprehensive Specialized Hospital Medical College in Ethiopia. A cross-sectional study design was employed, involving structured interviews with parents/guardians, medical records review, and clinical examination of children. Data on demographic characteristics, environmental exposures, family history of respiratory diseases, and socioeconomic status were collected. Logistic regression analysis was used to identify factors associated with recurrent wheezing. The study included X 6-year-old children, with a prevalence of recurrent wheezing found to be Y%. Environmental exposures, including tobacco smoke exposure (OR = Z, 95% CI: X-Y), indoor air pollution (OR = Z, 95% CI: X-Y), and presence of pets at home (OR = Z, 95% CI: X-Y), were identified as significant risk factors for recurrent wheezing. Additionally, a family history of asthma or allergies (OR = Z, 95% CI: X-Y) and low socioeconomic status (OR = Z, 95% CI: X-Y) were associated with an increased likelihood of recurrent wheezing. The impact of recurrent wheezing on the quality of life of affected children and their families was also assessed. Children with recurrent wheezing experienced a higher frequency of respiratory symptoms, increased healthcare utilization, and decreased physical activity compared to their non-wheezing counterparts. In conclusion, recurrent wheezing among 6-year-old children attending Adama Comprehensive Specialized Hospital Medical College is associated with various environmental, genetic, and socioeconomic factors. These findings underscore the importance of targeted interventions aimed at reducing exposure to known triggers and improving respiratory health outcomes in this population. Future research should focus on longitudinal studies to further elucidate the causal relationships between risk factors and recurrent wheezing and evaluate the effectiveness of preventive strategies.Keywords: wheezing, inflammation, respiratory, crucial
Procedia PDF Downloads 54987 Electromagnetic-Mechanical Stimulation on PC12 for Enhancement of Nerve Axonal Extension
Authors: E. Nakamachi, K. Matsumoto, K. Yamamoto, Y. Morita, H. Sakamoto
Abstract:
In recently, electromagnetic and mechanical stimulations have been recognized as the effective extracellular environment stimulation technique to enhance the defected peripheral nerve tissue regeneration. In this study, we developed a new hybrid bioreactor by adopting 50 Hz uniform alternative current (AC) magnetic stimulation and 4% strain mechanical stimulation. The guide tube for nerve regeneration is mesh structured tube made of biodegradable polymer, such as polylatic acid (PLA). However, when neural damage is large, there is a possibility that peripheral nerve undergoes necrosis. So it is quite important to accelerate the nerve tissue regeneration by achieving enhancement of nerve axonal extension rate. Therefore, we try to design and fabricate the system that can simultaneously load the uniform AC magnetic field stimulation and the stretch stimulation to cells for enhancement of nerve axonal extension. Next, we evaluated systems performance and the effectiveness of each stimulation for rat adrenal pheochromocytoma cells (PC12). First, we designed and fabricated the uniform AC magnetic field system and the stretch stimulation system. For the AC magnetic stimulation system, we focused on the use of pole piece structure to carry out in-situ microscopic observation. We designed an optimum pole piece structure using the magnetic field finite element analyses and the response surface methodology. We fabricated the uniform AC magnetic field stimulation system as a bio-reactor by adopting analytically determined design specifications. We measured magnetic flux density that is generated by the uniform AC magnetic field stimulation system. We confirmed that measurement values show good agreement with analytical results, where the uniform magnetic field was observed. Second, we fabricated the cyclic stretch stimulation device under the conditions of particular strains, where the chamber was made of polyoxymethylene (POM). We measured strains in the PC12 cell culture region to confirm the uniform strain. We found slightly different values from the target strain. Finally, we concluded that these differences were allowable in this mechanical stimulation system. We evaluated the effectiveness of each stimulation to enhance the nerve axonal extension using PC12. We confirmed that the average axonal extension length of PC12 under the uniform AC magnetic stimulation was increased by 16 % at 96 h in our bio-reactor. We could not confirm that the axonal extension enhancement under the stretch stimulation condition, where we found the exfoliating of cells. Further, the hybrid stimulation enhanced the axonal extension. Because the magnetic stimulation inhibits the exfoliating of cells. Finally, we concluded that the enhancement of PC12 axonal extension is due to the magnetic stimulation rather than the mechanical stimulation. Finally, we confirmed that the effectiveness of the uniform AC magnetic field stimulation for the nerve axonal extension using PC12 cells.Keywords: nerve cell PC12, axonal extension, nerve regeneration, electromagnetic-mechanical stimulation, bioreactor
Procedia PDF Downloads 265986 Unsupervised Part-of-Speech Tagging for Amharic Using K-Means Clustering
Authors: Zelalem Fantahun
Abstract:
Part-of-speech tagging is the process of assigning a part-of-speech or other lexical class marker to each word into naturally occurring text. Part-of-speech tagging is the most fundamental and basic task almost in all natural language processing. In natural language processing, the problem of providing large amount of manually annotated data is a knowledge acquisition bottleneck. Since, Amharic is one of under-resourced language, the availability of tagged corpus is the bottleneck problem for natural language processing especially for POS tagging. A promising direction to tackle this problem is to provide a system that does not require manually tagged data. In unsupervised learning, the learner is not provided with classifications. Unsupervised algorithms seek out similarity between pieces of data in order to determine whether they can be characterized as forming a group. This paper explicates the development of unsupervised part-of-speech tagger using K-Means clustering for Amharic language since large amount of data is produced in day-to-day activities. In the development of the tagger, the following procedures are followed. First, the unlabeled data (raw text) is divided into 10 folds and tokenization phase takes place; at this level, the raw text is chunked at sentence level and then into words. The second phase is feature extraction which includes word frequency, syntactic and morphological features of a word. The third phase is clustering. Among different clustering algorithms, K-means is selected and implemented in this study that brings group of similar words together. The fourth phase is mapping, which deals with looking at each cluster carefully and the most common tag is assigned to a group. This study finds out two features that are capable of distinguishing one part-of-speech from others these are morphological feature and positional information and show that it is possible to use unsupervised learning for Amharic POS tagging. In order to increase performance of the unsupervised part-of-speech tagger, there is a need to incorporate other features that are not included in this study, such as semantic related information. Finally, based on experimental result, the performance of the system achieves a maximum of 81% accuracy.Keywords: POS tagging, Amharic, unsupervised learning, k-means
Procedia PDF Downloads 452985 Evaluating Daylight Performance in an Office Environment in Malaysia, Using Venetian Blind System: Case Study
Authors: Fatemeh Deldarabdolmaleki, Mohamad Fakri Zaky Bin Ja'afar
Abstract:
Having a daylit space together with view results in a pleasant and productive environment for office employees. A daylit space is a space which utilizes daylight as a basic source of illumination to fulfill user’s visual demands and minimizes the electric energy consumption. Malaysian weather is hot and humid all over the year because of its location in the equatorial belt. however, because most of the commercial buildings in Malaysia are air-conditioned, huge glass windows are normally installed in order to keep the physical and visual relation between inside and outside. As a result of climatic situation and mentioned new trend, an ordinary office has huge heat gain, glare, and discomfort for occupants. Balancing occupant’s comfort and energy conservation in a tropical climate is a real challenge. This study concentrates on evaluating a venetian blind system using per pixel analyzing tools based on the suggested cut-out metrics by the literature. Workplace area in a private office room has been selected as a case study. Eight-day measurement experiment was conducted to investigate the effect of different venetian blind angles in an office area under daylight conditions in Serdang, Malaysia. The study goal was to explore daylight comfort of a commercially available venetian blind system, its’ daylight sufficiency and excess (8:00 AM to 5 PM) as well as Glare examination. Recently developed software, analyzing High Dynamic Range Images (HDRI captured by CCD camera), such as radiance based Evalglare and hdrscope help to investigate luminance-based metrics. The main key factors are illuminance and luminance levels, mean and maximum luminance, daylight glare probability (DGP) and luminance ratio of the selected mask regions. The findings show that in most cases, morning session needs artificial lighting in order to achieve daylight comfort. However, in some conditions (e.g. 10° and 40° slat angles) in the second half of day the workplane illuminance level exceeds the maximum of 2000 lx. Generally, a rising trend is discovered toward mean window luminance and the most unpleasant cases occur after 2 P.M. Considering the luminance criteria rating, the uncomfortable conditions occur in the afternoon session. Surprisingly in no blind condition, extreme case of window/task ratio is not common. Studying the daylight glare probability, there is not any DGP value higher than 0.35 in this experiment.Keywords: daylighting, energy simulation, office environment, Venetian blind
Procedia PDF Downloads 260984 The Challenge of Assessing Social AI Threats
Authors: Kitty Kioskli, Theofanis Fotis, Nineta Polemi
Abstract:
The European Union (EU) directive Artificial Intelligence (AI) Act in Article 9 requires that risk management of AI systems includes both technical and human oversight, while according to NIST_AI_RFM (Appendix C) and ENISA AI Framework recommendations, claim that further research is needed to understand the current limitations of social threats and human-AI interaction. AI threats within social contexts significantly affect the security and trustworthiness of the AI systems; they are interrelated and trigger technical threats as well. For example, lack of explainability (e.g. the complexity of models can be challenging for stakeholders to grasp) leads to misunderstandings, biases, and erroneous decisions. Which in turn impact the privacy, security, accountability of the AI systems. Based on the NIST four fundamental criteria for explainability it can also classify the explainability threats into four (4) sub-categories: a) Lack of supporting evidence: AI systems must provide supporting evidence or reasons for all their outputs. b) Lack of Understandability: Explanations offered by systems should be comprehensible to individual users. c) Lack of Accuracy: The provided explanation should accurately represent the system's process of generating outputs. d) Out of scope: The system should only function within its designated conditions or when it possesses sufficient confidence in its outputs. Biases may also stem from historical data reflecting undesired behaviors. When present in the data, biases can permeate the models trained on them, thereby influencing the security and trustworthiness of the of AI systems. Social related AI threats are recognized by various initiatives (e.g., EU Ethics Guidelines for Trustworthy AI), standards (e.g. ISO/IEC TR 24368:2022 on AI ethical concerns, ISO/IEC AWI 42105 on guidance for human oversight of AI systems) and EU legislation (e.g. the General Data Protection Regulation 2016/679, the NIS 2 Directive 2022/2555, the Directive on the Resilience of Critical Entities 2022/2557, the EU AI Act, the Cyber Resilience Act). Measuring social threats, estimating the risks to AI systems associated to these threats and mitigating them is a research challenge. In this paper it will present the efforts of two European Commission Projects (FAITH and THEMIS) from the HorizonEurope programme that analyse the social threats by building cyber-social exercises in order to study human behaviour, traits, cognitive ability, personality, attitudes, interests, and other socio-technical profile characteristics. The research in these projects also include the development of measurements and scales (psychometrics) for human-related vulnerabilities that can be used in estimating more realistically the vulnerability severity, enhancing the CVSS4.0 measurement.Keywords: social threats, artificial Intelligence, mitigation, social experiment
Procedia PDF Downloads 66983 Tailoring Quantum Oscillations of Excitonic Schrodinger’s Cats as Qubits
Authors: Amit Bhunia, Mohit Kumar Singh, Maryam Al Huwayz, Mohamed Henini, Shouvik Datta
Abstract:
We report [https://arxiv.org/abs/2107.13518] experimental detection and control of Schrodinger’s Cat like macroscopically large, quantum coherent state of a two-component Bose-Einstein condensate of spatially indirect electron-hole pairs or excitons using a resonant tunneling diode of III-V Semiconductors. This provides access to millions of excitons as qubits to allow efficient, fault-tolerant quantum computation. In this work, we measure phase-coherent periodic oscillations in photo-generated capacitance as a function of an applied voltage bias and light intensity over a macroscopically large area. Periodic presence and absence of splitting of excitonic peaks in the optical spectra measured by photocapacitance point towards tunneling induced variations in capacitive coupling between the quantum well and quantum dots. Observation of negative ‘quantum capacitance’ due to a screening of charge carriers by the quantum well indicates Coulomb correlations of interacting excitons in the plane of the sample. We also establish that coherent resonant tunneling in this well-dot heterostructure restricts the available momentum space of the charge carriers within this quantum well. Consequently, the electric polarization vector of the associated indirect excitons collective orients along the direction of applied bias and these excitons undergo Bose-Einstein condensation below ~100 K. Generation of interference beats in photocapacitance oscillation even with incoherent white light further confirm the presence of stable, long-range spatial correlation among these indirect excitons. We finally demonstrate collective Rabi oscillations of these macroscopically large, ‘multipartite’, two-level, coupled and uncoupled quantum states of excitonic condensate as qubits. Therefore, our study not only brings the physics and technology of Bose-Einstein condensation within the reaches of semiconductor chips but also opens up experimental investigations of the fundamentals of quantum physics using similar techniques. Operational temperatures of such two-component excitonic BEC can be raised further with a more densely packed, ordered array of QDs and/or using materials having larger excitonic binding energies. However, fabrications of single crystals of 0D-2D heterostructures using 2D materials (e.g. transition metal di-chalcogenides, oxides, perovskites etc.) having higher excitonic binding energies are still an open challenge for semiconductor optoelectronics. As of now, these 0D-2D heterostructures can already be scaled up for mass production of miniaturized, portable quantum optoelectronic devices using the existing III-V and/or Nitride based semiconductor fabrication technologies.Keywords: exciton, Bose-Einstein condensation, quantum computation, heterostructures, semiconductor Physics, quantum fluids, Schrodinger's Cat
Procedia PDF Downloads 180982 X-Ray Detector Technology Optimization in Computed Tomography
Authors: Aziz Ikhlef
Abstract:
Most of multi-slices Computed Tomography (CT) scanners are built with detectors composed of scintillator - photodiodes arrays. The photodiodes arrays are mainly based on front-illuminated technology for detectors under 64 slices and on back-illuminated photodiode for systems of 64 slices or more. The designs based on back-illuminated photodiodes were being investigated for CT machines to overcome the challenge of the higher number of runs and connection required in front-illuminated diodes. In backlit diodes, the electronic noise has already been improved because of the reduction of the load capacitance due to the routing reduction. This is translated by a better image quality in low signal application, improving low dose imaging in large patient population. With the fast development of multi-detector-rows CT (MDCT) scanners and the increasing number of examinations, the clinical community has raised significant concerns on radiation dose received by the patient in both medical and regulatory community. In order to reduce individual exposure and in response to the recommendations of the International Commission on Radiological Protection (ICRP) which suggests that all exposures should be kept as low as reasonably achievable (ALARA), every manufacturer is trying to implement strategies and solutions to optimize dose efficiency and image quality based on x-ray emission and scanning parameters. The added demands on the CT detector performance also comes from the increased utilization of spectral CT or dual-energy CT in which projection data of two different tube potentials are collected. One of the approaches utilizes a technology called fast-kVp switching in which the tube voltage is switched between 80 kVp and 140 kVp in fraction of a millisecond. To reduce the cross-contamination of signals, the scintillator based detector temporal response has to be extremely fast to minimize the residual signal from previous samples. In addition, this paper will present an overview of detector technologies and image chain improvement which have been investigated in the last few years to improve the signal-noise ratio and the dose efficiency CT scanners in regular examinations and in energy discrimination techniques. Several parameters of the image chain in general and in the detector technology contribute in the optimization of the final image quality. We will go through the properties of the post-patient collimation to improve the scatter-to-primary ratio, the scintillator material properties such as light output, afterglow, primary speed, crosstalk to improve the spectral imaging, the photodiode design characteristics and the data acquisition system (DAS) to optimize for crosstalk, noise and temporal/spatial resolution.Keywords: computed tomography, X-ray detector, medical imaging, image quality, artifacts
Procedia PDF Downloads 195981 Detection of Phoneme [S] Mispronounciation for Sigmatism Diagnosis in Adults
Authors: Michal Krecichwost, Zauzanna Miodonska, Pawel Badura
Abstract:
The diagnosis of sigmatism is mostly based on the observation of articulatory organs. It is, however, not always possible to precisely observe the vocal apparatus, in particular in the oral cavity of the patient. Speech processing can allow to objectify the therapy and simplify the verification of its progress. In the described study the methodology for classification of incorrectly pronounced phoneme [s] is proposed. The recordings come from adults. They were registered with the speech recorder at the sampling rate of 44.1 kHz and the resolution of 16 bit. The database of pathological and normative speech has been collected for the study including reference assessments provided by the speech therapy experts. Ten adult subjects were asked to simulate a certain type of stigmatism under the speech therapy expert supervision. In the recordings, the analyzed phone [s] was surrounded by vowels, viz: ASA, ESE, ISI, SPA, USU, YSY. Thirteen MFCC (mel-frequency cepstral coefficients) and RMS (root mean square) values are calculated within each frame being a part of the analyzed phoneme. Additionally, 3 fricative formants along with corresponding amplitudes are determined for the entire segment. In order to aggregate the information within the segment, the average value of each MFCC coefficient is calculated. All features of other types are aggregated by means of their 75th percentile. The proposed method of features aggregation reduces the size of the feature vector used in the classification. Binary SVM (support vector machine) classifier is employed at the phoneme recognition stage. The first group consists of pathological phones, while the other of the normative ones. The proposed feature vector yields classification sensitivity and specificity measures above 90% level in case of individual logo phones. The employment of a fricative formants-based information improves the sole-MFCC classification results average of 5 percentage points. The study shows that the employment of specific parameters for the selected phones improves the efficiency of pathology detection referred to the traditional methods of speech signal parameterization.Keywords: computer-aided pronunciation evaluation, sibilants, sigmatism diagnosis, speech processing
Procedia PDF Downloads 284980 Price Prediction Line, Investment Signals and Limit Conditions Applied for the German Financial Market
Authors: Cristian Păuna
Abstract:
In the first decades of the 21st century, in the electronic trading environment, algorithmic capital investments became the primary tool to make a profit by speculations in financial markets. A significant number of traders, private or institutional investors are participating in the capital markets every day using automated algorithms. The autonomous trading software is today a considerable part in the business intelligence system of any modern financial activity. The trading decisions and orders are made automatically by computers using different mathematical models. This paper will present one of these models called Price Prediction Line. A mathematical algorithm will be revealed to build a reliable trend line, which is the base for limit conditions and automated investment signals, the core for a computerized investment system. The paper will guide how to apply these tools to generate entry and exit investment signals, limit conditions to build a mathematical filter for the investment opportunities, and the methodology to integrate all of these in automated investment software. The paper will also present trading results obtained for the leading German financial market index with the presented methods to analyze and to compare different automated investment algorithms. It was found that a specific mathematical algorithm can be optimized and integrated into an automated trading system with good and sustained results for the leading German Market. Investment results will be compared in order to qualify the presented model. In conclusion, a 1:6.12 risk was obtained to reward ratio applying the trigonometric method to the DAX Deutscher Aktienindex on 24 months investment. These results are superior to those obtained with other similar models as this paper reveal. The general idea sustained by this paper is that the Price Prediction Line model presented is a reliable capital investment methodology that can be successfully applied to build an automated investment system with excellent results.Keywords: algorithmic trading, automated trading systems, high-frequency trading, DAX Deutscher Aktienindex
Procedia PDF Downloads 131979 Land, History and Housing: Colonial Legacies and Land Tenure in Kuala Lumpur
Authors: Nur Fareza Mustapha
Abstract:
Solutions to policy problems need to be curated to the local context, taking into account the trajectory of the local development path to ensure its efficacy. For Kuala Lumpur, rapid urbanization and migration into the city for the past few decades have increased the demand for housing to accommodate a growing urban population. As a critical factor affecting housing affordability, land supply constraints have been attributed to intensifying market pressures, which grew in tandem with the demands of urban development, along with existing institutional constraints in the governance of land. While demand-side pressures are inevitable given the fixed supply of land, supply-side constraints in regulations distort markets and if addressed inappropriately, may lead to mistargeted policy interventions. Given Malaysia’s historical development, regulatory barriers for land may originate from the British colonial period, when many aspects of the current laws governing tenure were introduced and formalized, and henceforth, became engrained in the system. This research undertakes a postcolonial institutional analysis approach to uncover the causal mechanism driving the evolution of land tenure systems in post-colonial Kuala Lumpur. It seeks to determine the sources of these shifts, focusing on the incentives and bargaining positions of actors during periods of institutional flux/change. It aims to construct a conceptual framework to further this understanding and to elucidate how this historical trajectory affects current access to urban land markets for housing. Archival analysis is used to outline and analyse the evolution of land tenure systems in Kuala Lumpur while stakeholder interviews are used to analyse its impact on the current urban land market, with a particular focus on the provision of and access to affordable housing in the city. Preliminary findings indicate that many aspects of the laws governing tenure that were introduced and formalized during the British colonial period have endured until the present day. Customary rules of tenure were displaced by rules following a European tradition, which found legitimacy through a misguided interpretation of local laws regarding the ownership of land. Colonial notions of race and its binary view of native vs. non-natives have also persisted in the construction and implementation of current legislation regarding land tenure. More concrete findings from this study will generate a more nuanced understanding of the regulatory land supply constraints in Kuala Lumpur, taking into account both the long and short term spatial and temporal processes that affect how these rules are created, implemented and enforced.Keywords: colonial discourse, historical institutionalism, housing, land policy, post-colonial city
Procedia PDF Downloads 128978 Proposed Design Principles for Low-Income Housing in South Africa
Authors: Gerald Steyn
Abstract:
Despite the huge number of identical, tiny, boxy, freestanding houses built by the South African government after the advent of democracy in 1994, squatter camps continue to mushroom, and there is no evidence that the backlog is being reduced. Not only is the wasteful low-density detached-unit approach of the past being perpetuated, but the social, spatial, and economic marginalization is worse than before 1994. The situation is precarious since squatters are vulnerable to fires and flooding. At the same time, the occupants of the housing schemes are trapped far from employment opportunities or any public amenities. Despite these insecurities, the architectural, urban design, and city planning professions are puzzlingly quiet. Design projects address these issues only at the universities, albeit inevitably with somewhat Utopian notions. Geoffrey Payne, the renowned urban housing and urban development consultant and researcher focusing on issues in the Global South, once proclaimed that “we do not have a housing problem – we have a settlement problem.” This dictum was used as the guiding philosophy to conceptualize urban design and architectural principles that foreground the needs of low-income households and allow them to be fully integrated into the larger conurbation. Information was derived from intensive research over two decades, involving frequent visits to informal settlements, historic Black townships, and rural villages. Observations, measured site surveys, and interviews resulted in several scholarly articles from which a set of desirable urban and architectural criteria could be extracted. To formulate culturally appropriate design principles, existing vernacular and informal patterns were analyzed, reconciled with contemporary designs that align with the requirements for the envisaged settlement attributes, and reimagined as residential design principles. Five interrelated design principles are proposed, ranging in scale from (1) Integrating informal settlements into the city, (2) linear neighborhoods, (3) market streets as wards, (4) linear neighborhoods, and (5) typologies and densities for clustered and aggregated patios and courtyards. Each design principle is described, first in terms of its context and associated issues of concern, followed by a discussion of the patterns available to inform a possible solution, and finally, an explanation and graphic illustration of the proposed design. The approach is predominantly bottom-up since each of the five principles is unfolded from existing informal and vernacular practices studied in situ. They are, however, articulated and represented in terms of contemporary design language. Contrary to an idealized vision of housing for South Africa’s low-income urban households, this study proposes actual principles for critical assessment by peers in the tradition of architectural research in design.Keywords: culturally appropriate design principles, informal settlements, South Africa’s housing backlog, squatter camps
Procedia PDF Downloads 51977 Modeling of Geotechnical Data Using GIS and Matlab for Eastern Ahmedabad City, Gujarat
Authors: Rahul Patel, S. P. Dave, M. V Shah
Abstract:
Ahmedabad is a rapidly growing city in western India that is experiencing significant urbanization and industrialization. With projections indicating that it will become a metropolitan city in the near future, various construction activities are taking place, making soil testing a crucial requirement before construction can commence. To achieve this, construction companies and contractors need to periodically conduct soil testing. This study focuses on the process of creating a spatial database that is digitally formatted and integrated with geotechnical data and a Geographic Information System (GIS). Building a comprehensive geotechnical Geo-database involves three essential steps. Firstly, borehole data is collected from reputable sources. Secondly, the accuracy and redundancy of the data are verified. Finally, the geotechnical information is standardized and organized for integration into the database. Once the Geo-database is complete, it is integrated with GIS. This integration allows users to visualize, analyze, and interpret geotechnical information spatially. Using a Topographic to Raster interpolation process in GIS, estimated values are assigned to all locations based on sampled geotechnical data values. The study area was contoured for SPT N-Values, Soil Classification, Φ-Values, and Bearing Capacity (T/m2). Various interpolation techniques were cross-validated to ensure information accuracy. The GIS map generated by this study enables the calculation of SPT N-Values, Φ-Values, and bearing capacities for different footing widths and various depths. This approach highlights the potential of GIS in providing an efficient solution to complex phenomena that would otherwise be tedious to achieve through other means. Not only does GIS offer greater accuracy, but it also generates valuable information that can be used as input for correlation analysis. Furthermore, this system serves as a decision support tool for geotechnical engineers. The information generated by this study can be utilized by engineers to make informed decisions during construction activities. For instance, they can use the data to optimize foundation designs and improve site selection. In conclusion, the rapid growth experienced by Ahmedabad requires extensive construction activities, necessitating soil testing. This study focused on the process of creating a comprehensive geotechnical database integrated with GIS. The database was developed by collecting borehole data from reputable sources, verifying its accuracy and redundancy, and organizing the information for integration. The GIS map generated by this study is an efficient solution that offers greater accuracy and generates valuable information that can be used as input for correlation analysis. It also serves as a decision support tool for geotechnical engineers, allowing them to make informed decisions during construction activities.Keywords: arcGIS, borehole data, geographic information system (GIS), geo-database, interpolation, SPT N-value, soil classification, φ-value, bearing capacity
Procedia PDF Downloads 70976 Railway Ballast Volumes Automated Estimation Based on LiDAR Data
Authors: Bahar Salavati Vie Le Sage, Ismaïl Ben Hariz, Flavien Viguier, Sirine Noura Kahil, Audrey Jacquin, Maxime Convert
Abstract:
The ballast layer plays a key role in railroad maintenance and the geometry of the track structure. Ballast also holds the track in place as the trains roll over it. Track ballast is packed between the sleepers and on the sides of railway tracks. An imbalance in ballast volume on the tracks can lead to safety issues as well as a quick degradation of the overall quality of the railway segment. If there is a lack of ballast in the track bed during the summer, there is a risk that the rails will expand and buckle slightly due to the high temperatures. Furthermore, the knowledge of the ballast quantities that will be excavated during renewal works is important for efficient ballast management. The volume of excavated ballast per meter of track can be calculated based on excavation depth, excavation width, volume of track skeleton (sleeper and rail) and sleeper spacing. Since 2012, SNCF has been collecting 3D points cloud data covering its entire railway network by using 3D laser scanning technology (LiDAR). This vast amount of data represents a modelization of the entire railway infrastructure, allowing to conduct various simulations for maintenance purposes. This paper aims to present an automated method for ballast volume estimation based on the processing of LiDAR data. The estimation of abnormal volumes in ballast on the tracks is performed by analyzing the cross-section of the track. Further, since the amount of ballast required varies depending on the track configuration, the knowledge of the ballast profile is required. Prior to track rehabilitation, excess ballast is often present in the ballast shoulders. Based on 3D laser scans, a Digital Terrain Model (DTM) was generated and automatic extraction of the ballast profiles from this data is carried out. The surplus in ballast is then estimated by performing a comparison between this ballast profile obtained empirically, and a geometric modelization of the theoretical ballast profile thresholds as dictated by maintenance standards. Ideally, this excess should be removed prior to renewal works and recycled to optimize the output of the ballast renewal machine. Based on these parameters, an application has been developed to allow the automatic measurement of ballast profiles. We evaluated the method on a 108 kilometers segment of railroad LiDAR scans, and the results show that the proposed algorithm detects ballast surplus that amounts to values close to the total quantities of spoil ballast excavated.Keywords: ballast, railroad, LiDAR , cloud point, track ballast, 3D point
Procedia PDF Downloads 112975 Current Situation and Need in Learning Management for Developing the Analytical Thinking of Teachers in Basic Education of Thailand
Authors: S. Art-in
Abstract:
This research was a survey research. The objective of this study was to study current situation and need in learning management for developing the analytical thinking of teachers in basic education of Thailand. The target group consisted of 400 teachers teaching in basic education level. They were selected by multi-stage random sampling. The instrument used in this study was the questionnaire asking current situation and need in learning management for developing the analytical thinking, 5 level rating scale. Data were analyzed by calculating the frequency, mean, standard deviation, percentage and content analysis. The research found that: 1) For current situation, the teachers provided learning management for developing analytical thinking, in overall, in “high” level. The issue with lowest level of practice: the teachers had competency in designing and establishing the learning management plan for developing the students’ analytical thinking. Considering each aspect it was found that: 1.1) the teacher aspect; the issue with lowest level of practice was: the teachers had competency in designing and establishing the learning management plan for developing the students’ analytical thinking, and 1.2) the learning management aspect for developing the students’ analytical thinking, the issue with lowest level of practice was: the learning activities provided opportunity for students to evaluate their analytical thinking process in each learning session. 2) The teachers showed their need in learning management for developing the analytical thinking, in overall, in “the highest” level. The issue with highest level of the need was: to obtain knowledge and competency in model, technique, and method for learning management or steps of learning management for developing the students’ analytical thinking. Considering each aspect it was found that: 2.1) teacher aspect; the issue with highest level of the need was: to obtain knowledge and comprehension in model, technique, and method for learning management or steps of learning management for developing the students’ analytical thinking, and 2.2) learning management aspect for developing the analytical thinking, the issue with highest level of need consisted of the determination of learning activities as problem situation, and the opportunity for students to comprehend the problem situation as well as practice their analytical thinking in order to find the answer.Keywords: current situation and need, learning management, analytical thinking, teachers in basic education level, Thailand
Procedia PDF Downloads 352974 Analysis of Thermal Comfort in Educational Buildings Using Computer Simulation: A Case Study in Federal University of Parana, Brazil
Authors: Ana Julia C. Kfouri
Abstract:
A prerequisite of any building design is to provide security to the users, taking the climate and its physical and physical-geometrical variables into account. It is also important to highlight the relevance of the right material elements, which arise between the person and the agent, and must provide improved thermal comfort conditions and low environmental impact. Furthermore, technology is constantly advancing, as well as computational simulations for projects, and they should be used to develop sustainable building and to provide higher quality of life for its users. In relation to comfort, the more satisfied the building users are, the better their intellectual performance will be. Based on that, the study of thermal comfort in educational buildings is of relative relevance, since the thermal characteristics in these environments are of vital importance to all users. Moreover, educational buildings are large constructions and when they are poorly planned and executed they have negative impacts to the surrounding environment, as well as to the user satisfaction, throughout its whole life cycle. In this line of thought, to evaluate university classroom conditions, it was accomplished a detailed case study on the thermal comfort situation at Federal University of Parana (UFPR). The main goal of the study is to perform a thermal analysis in three classrooms at UFPR, in order to address the subjective and physical variables that influence thermal comfort inside the classroom. For the assessment of the subjective components, a questionnaire was applied in order to evaluate the reference for the local thermal conditions. Regarding the physical variables, it was carried out on-site measurements, which consist of performing measurements of air temperature and air humidity, both inside and outside the building, as well as meteorological variables, such as wind speed and direction, solar radiation and rainfall, collected from a weather station. Then, a computer simulation based on results from the EnergyPlus software to reproduce air temperature and air humidity values of the three classrooms studied was conducted. The EnergyPlus outputs were analyzed and compared with the on-site measurement results to be possible to come out with a conclusion related to the local thermal conditions. The methodological approach included in the study allowed a distinct perspective in an educational building to better understand the classroom thermal performance, as well as the reason of such behavior. Finally, the study induces a reflection about the importance of thermal comfort for educational buildings and propose thermal alternatives for future projects, as well as a discussion about the significant impact of using computer simulation on engineering solutions, in order to improve the thermal performance of UFPR’s buildings.Keywords: computer simulation, educational buildings, EnergyPlus, humidity, temperature, thermal comfort
Procedia PDF Downloads 388973 Comparative Morphometric Analysis of Yelganga-Shivbhadra and Kohilla River Sub-Basins in Aurangabad District Maharashtra India
Authors: Chandrakant Gurav, Md Babar, Ajaykumar Asode
Abstract:
Morphometric analysis is the first stage of any basin analysis. By using these morphometric parameters we give indirect information about the nature and relations of stream with other streams, Geology of the area, groundwater condition and tectonic history of the basin. In the present study, Yelganga, Shivbhadra and Kohilla rivers, tributaries of the Godavari River in Aurangabad district, Maharashtra, India are considered to compare and study their morphometric characters. The linear, areal and relief morphometric aspects of the sub-basins have been assessed and evaluated in GIS environment. For this study, ArcGIS 10.1 software has been used for delineating, digitizing and generating different thematic maps. The Survey of India (SOI) toposheets maps and Shuttle Radar Topography Mission (SRTM) Digital Elevation Model (DEM) on resolution 30 m downloaded from United States Geological Survey (USGS) have been used for preparation of map and data generation. Geologically, the study area is covered by Central Deccan Volcanic Province (CDVP). It mainly consists of ‘aa’ type of basaltic lava flows of Late (upper) Cretaceous to Early (lower) Eocene age. The total geographical area of Yelganga, Shivbhadra and Kohilla river sub-basins are 185.5 sq. km., 142.6 sq. km and 122.3 sq. km. respectively The stream ordering method as suggested by the Strahler has been employed for present study and found that all the sub-basins are of 5th order streams. The average bifurcation ratio value of the sub-basins is below 5, indicates that there appears to be no strong structural control on drainage development, homogeneous nature of lithology and drainage network is in well-developed stage of erosion. The drainage density of Yelganga, Shivbhadra and Kohilla Sub-basins is 1.79 km/km2, 1.48 km/km2 and 1.89 km/km2 respectively and stream frequency is 1.94 streams/km2, 1.19 streams/km2 and 1.68 streams/km2 respectively, indicating semi-permeable sub-surface. Based on textural ratio values it indicates that the sub-basins have coarse texture. Shape parameters such as form factor ratio, circularity ratio and elongation ratio values shows that all three sub- basins are elongated in shape.Keywords: GIS, Kohilla, morphometry, Shivbhadra, Yelganga
Procedia PDF Downloads 157972 Finite Element Analysis of the Anaconda Device: Efficiently Predicting the Location and Shape of a Deployed Stent
Authors: Faidon Kyriakou, William Dempster, David Nash
Abstract:
Abdominal Aortic Aneurysm (AAA) is a major life-threatening pathology for which modern approaches reduce the need for open surgery through the use of stenting. The success of stenting though is sometimes jeopardized by the final position of the stent graft inside the human artery which may result in migration, endoleaks or blood flow occlusion. Herein, a finite element (FE) model of the commercial medical device AnacondaTM (Vascutek, Terumo) has been developed and validated in order to create a numerical tool able to provide useful clinical insight before the surgical procedure takes place. The AnacondaTM device consists of a series of NiTi rings sewn onto woven polyester fabric, a structure that despite its column stiffness is flexible enough to be used in very tortuous geometries. For the purposes of this study, a FE model of the device was built in Abaqus® (version 6.13-2) with the combination of beam, shell and surface elements; the choice of these building blocks was made to keep the computational cost to a minimum. The validation of the numerical model was performed by comparing the deployed position of a full stent graft device inside a constructed AAA with a duplicate set-up in Abaqus®. Specifically, an AAA geometry was built in CAD software and included regions of both high and low tortuosity. Subsequently, the CAD model was 3D printed into a transparent aneurysm, and a stent was deployed in the lab following the steps of the clinical procedure. Images on the frontal and sagittal planes of the experiment allowed the comparison with the results of the numerical model. By overlapping the experimental and computational images, the mean and maximum distances between the rings of the two models were measured in the longitudinal, and the transverse direction and, a 5mm upper bound was set as a limit commonly used by clinicians when working with simulations. The two models showed very good agreement of their spatial positioning, especially in the less tortuous regions. As a result, and despite the inherent uncertainties of a surgical procedure, the FE model allows confidence that the final position of the stent graft, when deployed in vivo, can also be predicted with significant accuracy. Moreover, the numerical model run in just a few hours, an encouraging result for applications in the clinical routine. In conclusion, the efficient modelling of a complicated structure which combines thin scaffolding and fabric has been demonstrated to be feasible. Furthermore, the prediction capabilities of the location of each stent ring, as well as the global shape of the graft, has been shown. This can allow surgeons to better plan their procedures and medical device manufacturers to optimize their designs. The current model can further be used as a starting point for patient specific CFD analysis.Keywords: AAA, efficiency, finite element analysis, stent deployment
Procedia PDF Downloads 193971 Efficacy of Transcranial Magnetic Therapy on Balance in Patients with Vestibular Dysfunction
Authors: Ibrahim M. I. Hamoda, Ahmed R. Z. Baghdadi, Mohammed K. Mohamed, Nawal A. Abu-Shady
Abstract:
Background: Most of patients with vestibular dysfunction suffering from balance disorders, Abnormality in balance increase effort and exertion which affect the independency, so this study might be a guide in managing balance problem and consequently improve walking with less exertion and maximum function. Purpose: to analyze and discuss the effect of transcranial magnetic therapy on balance in patients with vestibular dysfunction. Methods: forty subjects from both sexes were classified to divided randomly into two equal groups; Group I study group: this group received transcranial magnetic therapy, with a selected physical therapy program for improving balance and vestibular disorders (Balance training, Cawthorne-Cooksey Exercises) and group II (control group): this group received a selected physical therapy program as group I without transcranial magnetic therapy. This treatment procedure will be applied three times weekly for three months. The mean age was 54.53±3.44 and 55.33±2.32 years and BMI 35.7±3.03 and 35.73±1.03 kg/m2 for group I and II respectively. The Biodex Balance System, Berge balances scale (BBS) and brain MRI were used for assessment. Assessments were conducted before and after treatment. The treatment program for group I included balance training, Cawthorne-Cooksey Exercises and pulsed magnetic therapy (Parameters used in the program of 20 minutes, Intensity 2 gausses, Frequency 1 Hz). This selected program was done in approximately one hour every other day for three month. The treatment program group II Patients received the same program as group A without transcranial magnetic therapy. Results: The One-way ANOVA revealed that there were no significant differences in BBS scores, overall balance index, Anterior / posterior balance index, Medial / lateral balance index and dynamic limits of stability between both groups. Moreover, the BBS scores increased and overall balance index, Anterior / posterior balance index, Medial / lateral balance index and dynamic limits of stability decreased significantly after treatment in group I and II compared with before treatment. Interpretation/Conclusion: Adding pulsed magnetic therapy to balance training, Cawthorne-Cooksey Exercises has no effect on static and dynamic balance in patients with balance problems due to benign positional paroxysmal vertigo.Keywords: balance, transcranial magnetic therapy, vestibular dysfunction, biomechanic
Procedia PDF Downloads 482970 A Comparative Study of Motion Events Encoding in English and Italian
Authors: Alfonsina Buoniconto
Abstract:
The aim of this study is to investigate the degree of cross-linguistic and intra-linguistic variation in the encoding of motion events (MEs) in English and Italian, these being typologically different languages both showing signs of disobedience to their respective types. As a matter of fact, the traditional typological classification of MEs encoding distributes languages into two macro-types, based on the preferred locus for the expression of Path, the main ME component (other components being Figure, Ground and Manner) characterized by conceptual and structural prominence. According to this model, Satellite-framed (SF) languages typically express Path information in verb-dependent items called satellites (e.g. preverbs and verb particles) with main verbs encoding Manner of motion; whereas Verb-framed languages (VF) tend to include Path information within the verbal locus, leaving Manner to adjuncts. Although this dichotomy is valid altogether, languages do not always behave according to their typical classification patterns. English, for example, is usually ascribed to the SF type due to the rich inventory of postverbal particles and phrasal verbs used to express spatial relations (i.e. the cat climbed down the tree); nevertheless, it is not uncommon to find constructions such as the fog descended slowly, which is typical of the VF type. Conversely, Italian is usually described as being VF (cf. Paolo uscì di corsa ‘Paolo went out running’), yet SF constructions like corse via in lacrime ‘She ran away in tears’ are also frequent. This paper will try to demonstrate that such a typological overlapping is due to the fact that the semantic units making up MEs are distributed within several loci of the sentence –not only verbs and satellites– thus determining a number of different constructions stemming from convergent factors. Indeed, the linguistic expression of motion events depends not only on the typological nature of languages in a traditional sense, but also on a series morphological, lexical, and syntactic resources, as well as on inferential, discursive, usage-related, and cultural factors that make semantic information more or less accessible, frequent, and easy to process. Hence, rather than describe English and Italian in dichotomic terms, this study focuses on the investigation of cross-linguistic and intra-linguistic variation in the use of all the strategies made available by each linguistic system to express motion. Evidence for these assumptions is provided by parallel corpora analysis. The sample texts are taken from two contemporary Italian novels and their respective English translations. The 400 motion occurrences selected (200 in English and 200 in Italian) were scanned according to the MODEG (an acronym for Motion Decoding Grid) methodology, which grants data comparability through the indexation and retrieval of combined morphosyntactic and semantic information at different levels of detail.Keywords: construction typology, motion event encoding, parallel corpora, satellite-framed vs. verb-framed type
Procedia PDF Downloads 261969 A New Perspective in Cervical Dystonia: Neurocognitive Impairment
Authors: Yesim Sucullu Karadag, Pinar Kurt, Sule Bilen, Nese Subutay Oztekin, Fikri Ak
Abstract:
Background: Primary cervical dystonia is thought to be a purely motor disorder. But recent studies revealed that patients with dystonia had additional non-motor features. Sensory and psychiatric disturbances could be included into the non-motor spectrum of dystonia. The Basal Ganglia receive inputs from all cortical areas and throughout the thalamus project to several cortical areas, thus participating to circuits that have been linked to motor as well as sensory, emotional and cognitive functions. However, there are limited studies indicating cognitive impairment in patients with cervical dystonia. More evidence is required regarding neurocognitive functioning in these patients. Objective: This study is aimed to investigate neurocognitive profile of cervical dystonia patients in comparison to healthy controls (HC) by employing a detailed set of neuropsychological tests in addition to self-reported instruments. Methods: Totally 29 (M/F: 7/22) cervical dystonia patients and 30 HC (M/F: 10/20) were included into the study. Exclusion criteria were depression and not given informed consent. Standard demographic, educational data and clinical reports (disease duration, disability index) were recorded for all patients. After a careful neurological evaluation, all subjects were given a comprehensive battery of neuropsychological tests: Self report of neuropsychological condition (by visual analogue scale-VAS, 0-100), RAVLT, STROOP, PASAT, TMT, SDMT, JLOT, DST, COWAT, ACTT, and FST. Patients and HC were compared regarding demographic, clinical features and neurocognitive tests. Also correlation between disease duration, disability index and self report -VAS were assessed. Results: There was no difference between patients and HCs regarding socio-demographic variables such as age, gender and years of education (p levels were 0.36, 0.436, 0.869; respectively). All of the patients were assessed at the peak of botulinum toxine effect and they were not taking an anticholinergic agent or benzodiazepine. Dystonia patients had significantly impaired verbal learning and memory (RAVLT, p<0.001), divided attention and working memory (ACTT, p<0.001), attention speed (TMT-A and B, p=0.008, 0.050), executive functions (PASAT, p<0.001; SDMT, p= 0.001; FST, p<0.001), verbal attention (DST, p=0.001), verbal fluency (COWAT, p<0.001), visio-spatial processing (JLOT, p<0.001) in comparison to healthy controls. But focused attention (STROOP-spontaneous correction) was not different between two groups (p>0.05). No relationship was found regarding disease duration and disability index with any neurocognitive tests. Conclusions: Our study showed that neurocognitive functions of dystonia patients were worse than control group with the similar age, sex, and education independently clinical expression like disease duration and disability index. This situation may be the result of possible cortical and subcortical changes in dystonia patients. Advanced neuroimaging techniques might be helpful to explain these changes in cervical dystonia patients.Keywords: cervical dystonia, neurocognitive impairment, neuropsychological test, dystonia disability index
Procedia PDF Downloads 420968 The Effects of Billboard Content and Visible Distance on Driver Behavior
Authors: Arsalan Hassan Pour, Mansoureh Jeihani, Samira Ahangari
Abstract:
Distracted driving has been one of the most integral concerns surrounding our daily use of vehicles since the invention of the automobile. While much attention has been recently given to cell phones related distraction, commercial billboards along roads are also candidates for drivers' visual and cognitive distractions, as they may take drivers’ eyes from the road and their minds off the driving task to see, perceive and think about the billboard’s content. Using a driving simulator and a head-mounted eye-tracking system, speed change, acceleration, deceleration, throttle response, collision, lane changing, and offset from the center of the lane data along with gaze fixation duration and frequency data were collected in this study. Some 92 participants from a fairly diverse sociodemographic background drove on a simulated freeway in Baltimore, Maryland area and were exposed to three different billboards to investigate the effects of billboards on drivers’ behavior. Participants glanced at the billboards several times with different frequencies, the maximum of which occurred on the billboard with the highest cognitive load. About 74% of the participants didn’t look at billboards for more than two seconds at each glance except for the billboard with a short visible area. Analysis of variance (ANOVA) was performed to find the variations in driving behavior when they are invisible, readable, and post billboards area. The results show a slight difference in speed, throttle, brake, steering velocity, and lane changing, among different areas. Brake force and deviation from the center of the lane increased in the readable area in comparison with the visible area, and speed increased right after each billboard. The results indicated that billboards have a significant effect on driving performance and visual attention based on their content and visibility status. Generalized linear model (GLM) analysis showed no connection between participants’ age and driving experience with gaze duration. However, the visible distance of the billboard, gender, and billboard content had a significant effect on gaze duration.Keywords: ANOVA, billboards, distracted driving, drivers' behavior, driving simulator, eye-Tracking system, GLM
Procedia PDF Downloads 130967 Diurnal Circle of Rainfall and Convective Properties over West and Central Africa
Authors: Balogun R. Ayodeji, Adefisan E. Adesanya, Adeyewa Z. Debo, E. C. Okogbue
Abstract:
The need to investigate diurnal weather circles in West Africa is coined in the fact that complex interactions often results from diurnal weather patterns. This study investigates diurnal circles of wind, rainfall and convective properties using six (6) hour interval data from the ERA-Interim and the Tropical Rainfall Measurement Mission (TRMM). The seven distinct zones, used in this work and classified as rainforest (west-coast, dry, Nigeria-Cameroon), Savannah (Nigeria, and Central Africa and South Sudan (CASS)), Sudano-Sahel, and Sahel, were clearly indicated by the rainfall pattern in each zones. Results showed that the land‐ocean warming contrast was more strongly sensitive to seasonal cycle and has been very weak during March-May (MAM) but clearly spelt out during June-September (JJAS). Dipoles of wind convergence/divergence and wet/dry precipitation, between CASS and Nigeria Savannah zones, were identified in morning and evening hours of MAM, whereas distinct night and day anomaly, in the same location of CASS, were found to be consistent during the JJAS season. Diurnal variation of convective properties showed that stratiform precipitation, due to the extremely low occurrence of flashcount climatology, was dominant during morning hours for both MAM and JJAS than other periods of the day. On the other hand, diurnal variation of the system sizes showed that small system sizes were most dominant during the day time periods for both MAM and JJAS, whereas larger system sizes were frequent during the evening, night, and morning hours. The locations of flashcount and system sizes agreed with earlier results that morning and day-time hours were dominated by stratiform precipitation and small system sizes respectively. Most results clearly showed that the eastern locations of Sudano and Sahel were consistently dry because rainfall and precipitation features were predominantly few. System sizes greater than or equal to 800 km² were found in the western axis of the Sudano and Sahel zones, whereas the eastern axis, particularly in the Sahel zone, had minimal occurrences of small/large system sizes. From the results of locations of extreme systems, flashcount greater than 275 in one single system was never observed during the morning (6Z) diurnal, whereas, the evening (18Z) diurnal had the most frequent cases (at least 8) of flashcount exceeding 275 in one single system. Results presented had shown the importance of diurnal variation in understanding precipitation, flashcount, system sizes patterns at diurnal scales, and understanding land-ocean contrast, precipitation, and wind field anomaly at diurnal scales.Keywords: convective properties, diurnal circle, flashcount, system sizes
Procedia PDF Downloads 133966 Potential Impacts of Climate Change on Hydrological Droughts in the Limpopo River Basin
Authors: Nokwethaba Makhanya, Babatunde J. Abiodun, Piotr Wolski
Abstract:
Climate change possibly intensifies hydrological droughts and reduces water availability in river basins. Despite this, most research on climate change effects in southern Africa has focused exclusively on meteorological droughts. This thesis projects the potential impact of climate change on the future characteristics of hydrological droughts in the Limpopo River Basin (LRB). The study uses regional climate model (RCM) measurements (from the Coordinated Regional Climate Downscaling Experiment, CORDEX) and a combination of hydrological simulations (using the Soil and Water Assessment Tool Plus model, SWAT+) to predict the impacts at four global warming levels (GWLs: 1.5℃, 2.0℃, 2.5℃, and 3.0℃) under the RCP8.5 future climate scenario. The SWAT+ model was calibrated and validated with a streamflow dataset observed over the basin, and the sensitivity of model parameters was investigated. The performance of the SWAT+LRB model was verified using the Nash-Sutcliffe efficiency (NSE), Percent Bias (PBIAS), Root Mean Square Error (RMSE), and coefficient of determination (R²). The Standardized Precipitation Evapotranspiration Index (SPEI) and the Standardized Precipitation Index (SPI) have been used to detect meteorological droughts. The Soil Water Index (SSI) has been used to define agricultural drought, while the Water Yield Drought Index (WYLDI), the Surface Run-off Index (SRI), and the Streamflow Index (SFI) have been used to characterise hydrological drought. The performance of the SWAT+ model simulations over LRB is sensitive to the parameters CN2 (initial SCS runoff curve number for moisture condition II) and ESCO (soil evaporation compensation factor). The best simulation generally performed better during the calibration period than the validation period. In calibration and validation periods, NSE is ≤ 0.8, while PBIAS is ≥ ﹣80.3%, RMSE ≥ 11.2 m³/s, and R² ≤ 0.9. The simulations project a future increase in temperature and potential evapotranspiration over the basin, but they do not project a significant future trend in precipitation and hydrological variables. However, the spatial distribution of precipitation reveals a projected increase in precipitation in the southern part of the basin and a decline in the northern part of the basin, with the region of reduced precipitation projected to increase with GWLs. A decrease in all hydrological variables is projected over most parts of the basin, especially over the eastern part of the basin. The simulations predict meteorological droughts (i.e., SPEI and SPI), agricultural droughts (i.e., SSI), and hydrological droughts (i.e., WYLDI, SRI) would become more intense and severe across the basin. SPEI-drought has a greater magnitude of increase than SPI-drought, and agricultural and hydrological droughts have a magnitude of increase between the two. As a result, this research suggests that future hydrological droughts over the LRB could be more severe than the SPI-drought projection predicts but less severe than the SPEI-drought projection. This research can be used to mitigate the effects of potential climate change on basin hydrological drought.Keywords: climate change, CORDEX, drought, hydrological modelling, Limpopo River Basin
Procedia PDF Downloads 129965 Comparison of Yb and Tm-Fiber Laser Cutting Processes of Fiber Reinforced Plastics
Authors: Oktay Celenk, Ugur Karanfil, Iskender Demir, Samir Lamrini, Jorg Neumann, Arif Demir
Abstract:
Due to its favourable material characteristics, fiber reinforced plastics are amongst the main topics of all actual lightweight construction megatrends. Especially in transportation trends ranging from aeronautics over the automotive industry to naval transportation (yachts, cruise liners) the expected economic and environmental impact is huge. In naval transportation components like yacht bodies, antenna masts, decorative structures like deck lamps, light houses and pool areas represent cheap and robust solutions. Commercially available laser tools like carbon dioxide gas lasers (CO₂), frequency tripled solid state UV lasers, and Neodymium-YAG (Nd:YAG) lasers can be used. These tools have emission wavelengths of 10 µm, 0.355 µm, and 1.064 µm, respectively. The scientific goal is first of all the generation of a parameter matrix for laser processing of each used material for a Tm-fiber laser system (wavelength 2 µm). These parameters are the heat affected zone, process gas pressure, work piece feed velocity, intensity, irradiation time etc. The results are compared with results obtained with well-known material processing lasers, such as a Yb-fiber lasers (wavelength 1 µm). Compared to the CO₂-laser, the Tm-laser offers essential advantages for future laser processes like cutting, welding, ablating for repair and drilling in composite part manufacturing (components of cruise liners, marine pipelines). Some of these are the possibility of beam delivery in a standard fused silica fiber which enables hand guided processing, eye safety which results from the wavelength, excellent beam quality and brilliance due to the fiber nature. There is one more feature that is economically absolutely important for boat, automotive and military projects manufacturing that the wavelength of 2 µm is highly absorbed by the plastic matrix and thus enables selective removal of it for repair procedures.Keywords: Thulium (Tm) fiber laser, laser processing of fiber-reinforced plastics (FRP), composite, heat affected zone
Procedia PDF Downloads 193964 The Impact of Information and Communications Technology (ICT)-Enabled Service Adaptation on Quality of Life: Insights from Taiwan
Authors: Chiahsu Yang, Peiling Wu, Ted Ho
Abstract:
From emphasizing economic development to stressing public happiness, the international community mainly hopes to be able to understand whether the quality of life for the public is becoming better. The Better Life Index (BLI) constructed by OECD uses living conditions and quality of life as starting points to cover 11 areas of life and to convey the state of the general public’s well-being. In light of the BLI framework, the Directorate General of Budget, Accounting and Statistics (DGBAS) of the Executive Yuan instituted the Gross National Happiness Index to understand the needs of the general public and to measure the progress of the aforementioned conditions in residents across the island. Whereas living conditions consist of income and wealth, jobs and earnings, and housing conditions, health status, work and life balance, education and skills, social connections, civic engagement and governance, environmental quality, personal security. The ICT area consists of health care, living environment, ICT-enabled communication, transportation, government, education, pleasure, purchasing, job & employment. In the wake of further science and technology development, rapid formation of information societies, and closer integration between lifestyles and information societies, the public’s well-being within information societies has indeed become a noteworthy topic. the Board of Science and Technology of the Executive Yuan use the OECD’s BLI as a reference in the establishment of the Taiwan-specific ICT-Enabled Better Life Index. Using this index, the government plans to examine whether the public’s quality of life is improving as well as measure the public’s satisfaction with current digital quality of life. This understanding will enable the government to gauge the degree of influence and impact that each dimension of digital services has on digital life happiness while also serving as an important reference for promoting digital service development. The content of the ICT Enabled Better Life Index. Information and communications technology (ICT) has been affecting people’s living styles, and further impact people’s quality of life (QoL). Even studies have shown that ICT access and usage have both positive and negative impact on life satisfaction and well-beings, many governments continue to invest in e-government programs to initiate their path to information society. This research is the few attempts to link the e-government benchmark to the subjective well-being perception, and further address the gap between user’s perception and existing hard data assessment, then propose a model to trace measurement results back to the original public policy in order for policy makers to justify their future proposals.Keywords: information and communications technology, quality of life, satisfaction, well-being
Procedia PDF Downloads 356963 An Investigation of the Operation and Performance of London Cycle Hire Scheme
Authors: Amer Ali, Jessica Cecchinelli, Antonis Charalambous
Abstract:
Cycling is one of the most environmentally friendly, economic and healthy modes of transport but it needs more efficient cycle infrastructure and more effective safety measures. This paper represents an investigation into the performance and operation of the London Cycle Hire Scheme which started to operate in July 2010 using 5,000 cycles and 315 docking stations and currently has more than 10,000 cycles and over 700 docking stations across London which are available 24/7, 365 days a year. The study, which was conducted during the second half of 2014, consists of two parts; namely, the longitudinal review of the hire scheme between its introduction in 2010 and November 2014, and a field survey in November 2014 in the form of face-face interviews of the users of the cycle scheme to ascertain the existing limitations and difficulties experienced by those users and how it could be improved in terms of capability and safety. The study also includes a correlation between the usage of the cycle scheme and the corresponding weather conditions. The main findings are that on average the number of users (hiring frequency) had increased from just over two millions hires in 2010 to just less than ten millions in 2014. The field survey showed that 80% of the users are satisfied with the performance of the scheme whilst 50% of the users raised concern about the safety level of using the available cycle routes and infrastructure. The study also revealed that a high percentage of the cycle trips were relatively short (less than 30 minutes). Although the weather condition had some effect on cycling, the cost of using the cycle scheme and the main events in London had more effect on the number of cycle hires. The key conclusions are that despite the safety concern and the lack of infrastructure for continuous routes there was an encouraging number of people who opted for cycling as a clean, affordable, and healthy mode of transport. There is a need to expand the scheme by providing more cycles and docking stations and to support that by more well-designed and maintained cycle routes. More details about the development of London Cycle Hire Scheme during the last five years, its performance and the key issues revealed by the surveyed users will be reported in the full version of the paper.Keywords: cycling mode of transport, london cycle hire scheme, safety, environmental and health benefits, user satisfaction
Procedia PDF Downloads 387