Search results for: GPS signal acquisition
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2607

Search results for: GPS signal acquisition

447 Precoding-Assisted Frequency Division Multiple Access Transmission Scheme: A Cyclic Prefixes- Available Modulation-Based Filter Bank Multi-Carrier Technique

Authors: Ying Wang, Jianhong Xiang, Yu Zhong

Abstract:

The offset Quadrature Amplitude Modulation-based Filter Bank Multi-Carrier (FBMC) system provides superior spectral properties over Orthogonal Frequency Division Multiplexing. However, seriously affected by imaginary interference, its performances are hampered in many areas. In this paper, we propose a Precoding-Assisted Frequency Division Multiple Access (PA-FDMA) modulation scheme. By spreading FBMC symbols into the frequency domain and transmitting them with a precoding matrix, the impact of imaginary interference can be eliminated. Specifically, we first generate the coding pre-solution matrix with a nonuniform Fast Fourier Transform and pick the best columns by introducing auxiliary factors. Secondly, according to the column indexes, we obtain the precoding matrix for one symbol and impose scaling factors to ensure that the power is approximately constant throughout the transmission time. Finally, we map the precoding matrix of one symbol to multiple symbols and transmit multiple data frames, thus achieving frequency-division multiple access. Additionally, observing the interference between adjacent frames, we mitigate them by adding frequency Cyclic Prefixes (CP) and evaluating them with a signal-to-interference ratio. Note that PA-FDMA can be considered a CP-available FBMC technique because the underlying strategy is FBMC. Simulation results show that the proposed scheme has better performance compared to Single Carrier Frequency Division Multiple Access (SC-FDMA), etc.

Keywords: PA-FDMA, SC-FDMA, FBMC, non-uniform fast fourier transform

Procedia PDF Downloads 39
446 A Methodology Based on Image Processing and Deep Learning for Automatic Characterization of Graphene Oxide

Authors: Rafael do Amaral Teodoro, Leandro Augusto da Silva

Abstract:

Originated from graphite, graphene is a two-dimensional (2D) material that promises to revolutionize technology in many different areas, such as energy, telecommunications, civil construction, aviation, textile, and medicine. This is possible because its structure, formed by carbon bonds, provides desirable optical, thermal, and mechanical characteristics that are interesting to multiple areas of the market. Thus, several research and development centers are studying different manufacturing methods and material applications of graphene, which are often compromised by the scarcity of more agile and accurate methodologies to characterize the material – that is to determine its composition, shape, size, and the number of layers and crystals. To engage in this search, this study proposes a computational methodology that applies deep learning to identify graphene oxide crystals in order to characterize samples by crystal sizes. To achieve this, a fully convolutional neural network called U-net has been trained to segment SEM graphene oxide images. The segmentation generated by the U-net is fine-tuned with a standard deviation technique by classes, which allows crystals to be distinguished with different labels through an object delimitation algorithm. As a next step, the characteristics of the position, area, perimeter, and lateral measures of each detected crystal are extracted from the images. This information generates a database with the dimensions of the crystals that compose the samples. Finally, graphs are automatically created showing the frequency distributions by area size and perimeter of the crystals. This methodological process resulted in a high capacity of segmentation of graphene oxide crystals, presenting accuracy and F-score equal to 95% and 94%, respectively, over the test set. Such performance demonstrates a high generalization capacity of the method in crystal segmentation, since its performance considers significant changes in image extraction quality. The measurement of non-overlapping crystals presented an average error of 6% for the different measurement metrics, thus suggesting that the model provides a high-performance measurement for non-overlapping segmentations. For overlapping crystals, however, a limitation of the model was identified. To overcome this limitation, it is important to ensure that the samples to be analyzed are properly prepared. This will minimize crystal overlap in the SEM image acquisition and guarantee a lower error in the measurements without greater efforts for data handling. All in all, the method developed is a time optimizer with a high measurement value, considering that it is capable of measuring hundreds of graphene oxide crystals in seconds, saving weeks of manual work.

Keywords: characterization, graphene oxide, nanomaterials, U-net, deep learning

Procedia PDF Downloads 147
445 1-D Convolutional Neural Network Approach for Wheel Flat Detection for Freight Wagons

Authors: Dachuan Shi, M. Hecht, Y. Ye

Abstract:

With the trend of digitalization in railway freight transport, a large number of freight wagons in Germany have been equipped with telematics devices, commonly placed on the wagon body. A telematics device contains a GPS module for tracking and a 3-axis accelerometer for shock detection. Besides these basic functions, it is desired to use the integrated accelerometer for condition monitoring without any additional sensors. Wheel flats as a common type of failure on wheel tread cause large impacts on wagons and infrastructure as well as impulsive noise. A large wheel flat may even cause safety issues such as derailments. In this sense, this paper proposes a machine learning approach for wheel flat detection by using car body accelerations. Due to suspension systems, impulsive signals caused by wheel flats are damped significantly and thus could be buried in signal noise and disturbances. Therefore, it is very challenging to detect wheel flats using car body accelerations. The proposed algorithm considers the envelope spectrum of car body accelerations to eliminate the effect of noise and disturbances. Subsequently, a 1-D convolutional neural network (CNN), which is well known as a deep learning method, is constructed to automatically extract features in the envelope-frequency domain and conduct classification. The constructed CNN is trained and tested on field test data, which are measured on the underframe of a tank wagon with a wheel flat of 20 mm length in the operational condition. The test results demonstrate the good performance of the proposed algorithm for real-time fault detection.

Keywords: fault detection, wheel flat, convolutional neural network, machine learning

Procedia PDF Downloads 111
444 ePA-Coach: Design of the Intelligent Virtual Learning Coach for Senior Learners in Support of Digital Literacy in the Context of Electronic Patient Record

Authors: Ilona Buchem, Carolin Gellner

Abstract:

Over the last few years, the call for the support of senior learners in the development of their digital literacy has become prevalent, mainly due to the progression towards ageing societies paired with advances in digitalisation in all spheres of life, including e-health and electronic patient record (EPA). While major research efforts in supporting senior learners in developing digital literacy have been invested so far in e-learning focusing on knowledge acquisition and cognitive tasks, little research exists in learning models which target virtual mentoring and coaching with the help of pedagogical agents and address the social dimensions of learning. Research from studies with students in the context of formal education has already provided methods for designing intelligent virtual agents in support of personalised learning. However, this research has mostly focused on cognitive skills and has not yet been applied to the context of mentoring/coaching of senior learners, who have different characteristics and learn in different contexts. In this paper, we describe how insights from previous research can be used to develop an intelligent virtual learning coach (agent) for senior learners with a focus on building the social relationship between the agent and the learner and the key task of the agent to socialize learners to the larger context of digital literacy with a focus on electronic health records. Following current approaches to mentoring and coaching, the agent is designed not to enhance and monitor the cognitive performance of the learner but to serve as a trusted friend and advisor, whose role is to provide one-to-one guidance and support sharing of experiences among learners (peers). Based on literature review and synopsis of research on virtual agents and current coaching/mentoring models under consideration of the specific characteristics and requirements of senior learners, we describe the design framework which was applied to design an intelligent virtual learning coach as part of the e-learning system for digital literacy of senior learners in the ePA-Coach project founded by the German Ministry of Education and Research. This paper also presents the results from the evaluation study, which compared the use of the first prototype of the virtual learning coach designed according to the design framework with a voice narration in a multimedia learning environment with senior learners. The focus of the study was to validate the agent design in the context of the persona effect (Lester et al., 1997). Since the persona effect is related to the hypothesis that animated agents are perceived as more socially engaging, the study evaluated possible impacts of agent coaching in comparison with voice coaching on motivation, engagement, experience, and digital literacy.

Keywords: virtual learning coach, virtual mentor, pedagogical agent, senior learners, digital literacy, electronic health records

Procedia PDF Downloads 100
443 The Value of Computerized Corpora in EFL Textbook Design: The Case of Modal Verbs

Authors: Lexi Li

Abstract:

This study aims to contribute to the field of how computer technology can be exploited to enhance EFL textbook design. Specifically, the study demonstrates how computerized native and learner corpora can be used to enhance modal verb treatment in EFL textbooks. The linguistic focus is will, would, can, could, may, might, shall, should, must. The native corpus is the spoken component of BNC2014 (hereafter BNCS2014). The spoken part is chosen because the pedagogical purpose of the textbooks is communication-oriented. Using the standard query option of CQPweb, 5% of each of the nine modals was sampled from BNCS2014. The learner corpus is the POS-tagged Ten-thousand English Compositions of Chinese Learners (TECCL). All the essays under the “secondary school” section were selected. A series of five secondary coursebooks comprise the textbook corpus. All the data in both the learner and the textbook corpora are retrieved through the concordance functions of WordSmith Tools (version, 5.0). Data analysis was divided into two parts. The first part compared the patterns of modal verbs in the textbook corpus and BNC2014 with respect to distributional features, semantic functions, and co-occurring constructions to examine whether the textbooks reflect the authentic use of English. Secondly, the learner corpus was compared with the textbook corpus in terms of the use (distributional features, semantic functions, and co-occurring constructions) in order to examine the degree of influence of the textbook on learners’ use of modal verbs. Moreover, the learner corpus was analyzed for the misuse (syntactic errors, e.g., she can sings*.) of the nine modal verbs to uncover potential difficulties that confront learners. The results indicate discrepancies between the textbook presentation of modal verbs and authentic modal use in natural discourse in terms of distributions of frequencies, semantic functions, and co-occurring structures. Furthermore, there are consistent patterns of use between the learner corpus and the textbook corpus with respect to the three above-mentioned aspects, except could, will and must, partially confirming the correlation between the frequency effects and L2 grammar acquisition. Further analysis reveals that the exceptions are caused by both positive and negative L1 transfer, indicating that the frequency effects can be intercepted by L1 interference. Besides, error analysis revealed that could, would, should and must are the most difficult for Chinese learners due to both inter-linguistic and intra-linguistic interference. The discrepancies between the textbook corpus and the native corpus point to a need to adjust the presentation of modal verbs in the textbooks in terms of frequencies, different meanings, and verb-phrase structures. Along with the adjustment of modal verb treatment based on authentic use, it is important for textbook writers to take into consideration the L1 interference as well as learners’ difficulties in their use of modal verbs. The present study is a methodological showcase of the combination both native and learner corpora in the enhancement of EFL textbook language authenticity and appropriateness for learners.

Keywords: EFL textbooks, learner corpus, modal verbs, native corpus

Procedia PDF Downloads 104
442 Microfluidic Impedimetric Biochip and Related Methods for Measurement Chip Manufacture and Counting Cells

Authors: Amina Farooq, Nauman Zafar Butt

Abstract:

This paper is about methods and tools for counting particles of interest, such as cells. A microfluidic system with interconnected electronics on a flexible substrate, inlet-outlet ports and interface schemes, sensitive and selective detection of cells specificity, and processing of cell counting at polymer interfaces in a microscale biosensor for use in the detection of target biological and non-biological cells. The development of fluidic channels, planar fluidic contact ports, integrated metal electrodes on a flexible substrate for impedance measurements, and a surface modification plasma treatment as an intermediate bonding layer are all part of the fabrication process. Magnetron DC sputtering is used to deposit a double metal layer (Ti/Pt) over the polypropylene film. Using a photoresist layer, specified and etched zones are established. Small fluid volumes, a reduced detection region, and electrical impedance measurements over a range of frequencies for cell counts improve detection sensitivity and specificity. The procedure involves continuous flow of fluid samples that contain particles of interest through the microfluidic channels, counting all types of particles in a portion of the sample using the electrical differential counter to generate a bipolar pulse for each passing cell—calculating the total number of particles of interest originally in the fluid sample by using MATLAB program and signal processing. It's indeed potential to develop a robust and economical kit for cell counting in whole-blood samples using these methods and similar devices.

Keywords: impedance, biochip, cell counting, microfluidics

Procedia PDF Downloads 148
441 Weak Electric Fields Enhance Growth and Nutritional Quality of Kale

Authors: So-Ra Lee, Myung-Min Oh

Abstract:

Generally, plants growing on the earth are under the influence of natural electric fields and may even require exposure of the electric field to survive. Electric signals have been observed within plants and seem to play an important role on various metabolic processes, but their role is not fully understood. In this study, we attempted to explore the response of plants under external electric fields in kale (Brassica oleracea var. acephala). The plants were hydroponically grown for 28 days in a plant factory. Electric currents at 10, 50 and 100 mA were supplied to nutrient solution for 3 weeks. Additionally, some of the plants were cultivated in a Faraday cage to remove the natural electric field. Kale plants exposed to electric fields had higher fresh weight than the control and plants in Faraday cage. Absence of electric field caused a significant decrease in shoot dry weight and root growth. Leaf area also showed a similar response with shoot fresh weight. Supplying weak electric stimulation enhanced nutritional quality including total phenolic content and antioxidant capacity. This work provides basic information on the effects of electric fields on plants and is a meaningful attempt for developing a new economical technology to increase crop productivity and quality by applying an electric field. This work was supported by Korea Institute of Planning and Evaluation for Technology in Food, Agriculture, Forestry and Fisheries (IPET) through Agriculture, Food and Rural Affairs Research Center Support Program, funded by Ministry of Agriculture, Food and Rural Affairs (MAFRA) (717001-07-02-HD240).

Keywords: electroculture, electric signal, faraday cage, electric field

Procedia PDF Downloads 277
440 Smart Irrigation System for Applied Irrigation Management in Tomato Seedling Production

Authors: Catariny C. Aleman, Flavio B. Campos, Matheus A. Caliman, Everardo C. Mantovani

Abstract:

The seedling production stage is a critical point in the vegetable production system. Obtaining high-quality seedlings is a prerequisite for subsequent cropping to occur well and productivity optimization is required. The water management is an important step in agriculture production. The adequate water requirement in horticulture seedlings can provide higher quality and increase field production. The practice of irrigation is indispensable and requires a duly adjusted quality irrigation system, together with a specific water management plan to meet the water demand of the crop. Irrigation management in seedling management requires a great deal of specific information, especially when it involves the use of inputs such as hydrorentering polymers and automation technologies of the data acquisition and irrigation system. The experiment was conducted in a greenhouse at the Federal University of Viçosa, Viçosa - MG. Tomato seedlings (Lycopersicon esculentum Mill) were produced in plastic trays of 128 cells, suspended at 1.25 m from the ground. The seedlings were irrigated by 4 micro sprinklers of fixed jet 360º per tray, duly isolated by sideboards, following the methodology developed for this work. During Phase 1, in January / February 2017 (duration of 24 days), the cultivation coefficient (Kc) of seedlings cultured in the presence and absence of hydrogel was evaluated by weighing lysimeter. In Phase 2, September 2017 (duration of 25 days), the seedlings were submitted to 4 irrigation managements (Kc, timer, 0.50 ETo, and 1.00 ETo), in the presence and absence of hydrogel and then evaluated in relation to quality parameters. The microclimate inside the greenhouse was monitored with the use of air temperature, relative humidity and global radiation sensors connected to a microcontroller that performed hourly calculations of reference evapotranspiration by Penman-Monteith standard method FAO56 modified for the balance of long waves according to Walker, Aldrich, Short (1983), and conducted water balance and irrigation decision making for each experimental treatment. Kc of seedlings cultured on a substrate with hydrogel (1.55) was higher than Kc on a pure substrate (1.39). The use of the hydrogel was a differential for the production of earlier tomato seedlings, with higher final height, the larger diameter of the colon, greater accumulation of a dry mass of shoot, a larger area of crown projection and greater the rate of relative growth. The handling 1.00 ETo promoted higher relative growth rate.

Keywords: automatic system; efficiency of water use; precision irrigation, micro sprinkler.

Procedia PDF Downloads 97
439 Simultaneous Detection of Dopamine and Uric Acid in the Presence of Ascorbic Acid at Physiological Level Using Anodized Multiwalled Carbon Nanotube–Poldimethylsiloxane Paste Electrode

Authors: Angelo Gabriel Buenaventura, Allan Christopher Yago

Abstract:

A carbon paste electrode (CPE) composed of Multiwalled Carbon Nanotube (MWCNT) conducting particle and Polydimethylsiloxane (PDMS) binder was used for simultaneous detection of Dopamine (DA) and Uric Acid (UA) in the presence of Ascorbic Acid (AA) at physiological level. The MWCNT-PDMS CPE was initially activated via potentiodynamic cycling in a basic (NaOH) solution, which resulted in enhanced electrochemical properties. Electrochemical Impedance Spectroscopy measurements revealed a significantly lower charge transfer resistance (Rct) for the OH--activated MWCNT-PDMS CPE (Rct = 5.08kΩ) as compared to buffer (pH 7)-activated MWCNT-PDMS CPE (Rct = 25.9kΩ). Reversibility analysis of Fe(CN)63-/4- redox couple of both Buffer-Activated CPE and OH--Activated CPE showed that the OH—Activated CPE have peak current ratio (Ia/Ic) of 1.11 at 100mV/s while 2.12 for the Buffer-Activated CPE; this showed an electrochemically reversible behavior for Fe(CN)63-/4- redox couple even at relatively fast scan rate using the OH--activated CPE. Enhanced voltammetric signal for DA and significant peak separation between DA and UA was obtained using the OH--activated MWCNT-PDMS CPE in the presence of 50 μM AA via Differential Pulse Voltammetry technique. The anodic peak currents which appeared at 0.263V and 0.414 V were linearly increasing with increasing concentrations of DA and UA, respectively. The linear ranges were obtained at 25 μM – 100 μM for both DA and UA. The detection limit was determined to be 3.86 μM for DA and 5.61 μM for UA. These results indicate a practical approach in the simultaneous detection of important bio-organic molecules using a simple CPE composed of MWCNT and PDMS with base anodization as activation technique.

Keywords: anodization, ascorbic acid, carbon paste electrodes, dopamine, uric acid

Procedia PDF Downloads 268
438 Application of Electro-Optical Hybrid Cables in Horizontal Well Production Logging

Authors: Daofan Guo, Dong Yang

Abstract:

For decades, well logging with coiled tubing has relied solely on surface data such as pump pressure, wellhead pressure, depth counter, and weight indicator readings. While this data serves the oil industry well, modern smart logging utilizes real-time downhole information, which automatically increases operational efficiency and optimizes intervention qualities. For example, downhole pressure, temperature, and depth measurement data can be transmitted through the electro-optical hybrid cable in the coiled tubing to surface operators on a real-time base. This paper mainly introduces the unique structural features and various applications of the electro-optical hybrid cables which were deployed into downhole with the help of coiled tubing technology. Fiber optic elements in the cable enable optical communications and distributed measurements, such as distributed temperature and acoustic sensing. The electrical elements provide continuous surface power for downhole tools, eliminating the limitations of traditional batteries, such as temperature, operating time, and safety concerns. The electrical elements also enable cable telemetry operation of cable tools. Both power supply and signal transmission were integrated into an electro-optical hybrid cable, and the downhole information can be captured by downhole electrical sensors and distributed optical sensing technologies, then travels up through an optical fiber to the surface, which greatly improves the accuracy of measurement data transmission.

Keywords: electro-optical hybrid cable, underground photoelectric composite cable, seismic cable, coiled tubing, real-time monitoring

Procedia PDF Downloads 121
437 Development of a System for Measuring the Three-axis Pedal Force in Cycling and Its Applications

Authors: Joo-Hack Lee, Jin-Seung Choi, Dong-Won Kang, Jeong-Woo Seo, Ju-Young Kim, Dae-Hyeok Kim, Seung-Tae Yang, Gye-Rae Tack

Abstract:

For cycling, the analysis of the pedal force is one of the important factors in the study of exercise ability assessment and overuse injuries. In past studies, a two-axis measurement sensor was used at the sagittal plane to measure the force only in the anterior, posterior, and vertical directions and to analyze the loss of force and the injury on the frontal plane due to the forces in the right and left directions. In this study, which is a basic study on diverse analyses of the pedal force that consider the forces on the sagittal plane and the frontal plane, a three-axis pedal force measurement sensor was developed to measure the anterior-posterior (Fx), medio-lateral (Fz), and vertical (Fy) forces. The sensor was fabricated with a size and shape similar to those of the general flat pedal, and had a 550g weight that allowed smooth pedaling. Its measurement range was ±1000 N for Fx and Fz and ±2000 N for Fy, and its non-linearity, hysteresis, and repeatability were approximately 0.5%. The data were sampled at 1000 Hz using a signal collector. To use the developed sensor, the pedaling efficiency (index of efficiency, IE) and the range of left and right (medio-lateral, ML) forces were measured with two seat heights (low and high). The results of the measurement showed that the IE was higher and the force range in the ML direction was lower with the high position than with the low position. The developed measurement sensor and its application results will be useful in understanding and explaining the complicated pedaling technique, and will enable diverse kinematic analyses of the pedal force on the sagittal plane and the frontal plane.

Keywords: cycling, pedal force, index of effectiveness, measuring

Procedia PDF Downloads 644
436 Non-Uniform Filter Banks-based Minimum Distance to Riemannian Mean Classifition in Motor Imagery Brain-Computer Interface

Authors: Ping Tan, Xiaomeng Su, Yi Shen

Abstract:

The motion intention in the motor imagery braincomputer interface is identified by classifying the event-related desynchronization (ERD) and event-related synchronization ERS characteristics of sensorimotor rhythm (SMR) in EEG signals. When the subject imagines different limbs or different parts moving, the rhythm components and bandwidth will change, which varies from person to person. How to find the effective sensorimotor frequency band of subjects is directly related to the classification accuracy of brain-computer interface. To solve this problem, this paper proposes a Minimum Distance to Riemannian Mean Classification method based on Non-Uniform Filter Banks. During the training phase, the EEG signals are decomposed into multiple different bandwidt signals by using multiple band-pass filters firstly; Then the spatial covariance characteristics of each frequency band signal are computered to be as the feature vectors. these feature vectors will be classified by the MDRM (Minimum Distance to Riemannian Mean) method, and cross validation is employed to obtain the effective sensorimotor frequency bands. During the test phase, the test signals are filtered by the bandpass filter of the effective sensorimotor frequency bands, and the extracted spatial covariance feature vectors will be classified by using the MDRM. Experiments on the BCI competition IV 2a dataset show that the proposed method is superior to other classification methods.

Keywords: non-uniform filter banks, motor imagery, brain-computer interface, minimum distance to Riemannian mean

Procedia PDF Downloads 97
435 Long Wavelength Coherent Pulse of Sound Propagating in Granular Media

Authors: Rohit Kumar Shrivastava, Amalia Thomas, Nathalie Vriend, Stefan Luding

Abstract:

A mechanical wave or vibration propagating through granular media exhibits a specific signature in time. A coherent pulse or wavefront arrives first with multiply scattered waves (coda) arriving later. The coherent pulse is micro-structure independent i.e. it depends only on the bulk properties of the disordered granular sample, the sound wave velocity of the granular sample and hence bulk and shear moduli. The coherent wavefront attenuates (decreases in amplitude) and broadens with distance from its source. The pulse attenuation and broadening effects are affected by disorder (polydispersity; contrast in size of the granules) and have often been attributed to dispersion and scattering. To study the effect of disorder and initial amplitude (non-linearity) of the pulse imparted to the system on the coherent wavefront, numerical simulations have been carried out on one-dimensional sets of particles (granular chains). The interaction force between the particles is given by a Hertzian contact model. The sizes of particles have been selected randomly from a Gaussian distribution, where the standard deviation of this distribution is the relevant parameter that quantifies the effect of disorder on the coherent wavefront. Since, the coherent wavefront is system configuration independent, ensemble averaging has been used for improving the signal quality of the coherent pulse and removing the multiply scattered waves. The results concerning the width of the coherent wavefront have been formulated in terms of scaling laws. An experimental set-up of photoelastic particles constituting a granular chain is proposed to validate the numerical results.

Keywords: discrete elements, Hertzian contact, polydispersity, weakly nonlinear, wave propagation

Procedia PDF Downloads 185
434 Aggregation of Electric Vehicles for Emergency Frequency Regulation of Two-Area Interconnected Grid

Authors: S. Agheb, G. Ledwich, G.Walker, Z.Tong

Abstract:

Frequency control has become more of concern for reliable operation of interconnected power systems due to the integration of low inertia renewable energy sources to the grid and their volatility. Also, in case of a sudden fault, the system has less time to recover before widespread blackouts. Electric Vehicles (EV)s have the potential to cooperate in the Emergency Frequency Regulation (EFR) by a nonlinear control of the power system in case of large disturbances. The time is not adequate to communicate with each individual EV on emergency cases, and thus, an aggregate model is necessary for a quick response to prevent from much frequency deviation and the occurrence of any blackout. In this work, an aggregate of EVs is modelled as a big virtual battery in each area considering various aspects of uncertainty such as the number of connected EVs and their initial State of Charge (SOC) as stochastic variables. A control law was proposed and applied to the aggregate model using Lyapunov energy function to maximize the rate of reduction of total kinetic energy in a two-area network after the occurrence of a fault. The control methods are primarily based on the charging/ discharging control of available EVs as shunt capacity in the distribution system. Three different cases were studied considering the locational aspect of the model with the virtual EV either in the center of the two areas or in the corners. The simulation results showed that EVs could help the generator lose its kinetic energy in a short time after a contingency. Earlier estimation of possible contributions of EVs can help the supervisory control level to transmit a prompt control signal to the subsystems such as the aggregator agents and the grid. Thus, the percentage of EVs contribution for EFR will be characterized in the future as the goal of this study.

Keywords: emergency frequency regulation, electric vehicle, EV, aggregation, Lyapunov energy function

Procedia PDF Downloads 87
433 High-Frequency Modulation of Light-Emitting Diodes for New Ultraviolet Communications

Authors: Meng-Chyi Wu, Bonn Lin, Jyun-Hao Liao, Chein-Ju Chen, Yu-Cheng Jhuang, Mau-Phon Houng, Fang-Hsing Wang, Min-Chu Liu, Cheng-Fu Yang, Cheng-Shong Hong

Abstract:

Since the use of wireless communications has become critical nowadays, the available RF spectrum has become limited. Ultraviolet (UV) communication system can alleviate the spectrum constraint making UV communication system a potential alternative to future communication demands. Also, UV links can provide faster communication rate and can be used in combination with existing RF communication links, providing new communications diversity with higher user capacity. The UV region of electromagnetic spectrum has been of interest to detector, imaging and communication technologies because the stratospheric ozone layer effectively absorbs some solar UV radiation from reaching the earth surface. The wavebands where most of UV radiation is absorbed by the ozone are commonly known as the solar blind region. By operating in UV-C band (200-280 nm) the communication system can minimize the transmission power consumption since it will have less radiation noise. UV communication uses the UV ray as the medium. Electric signal is carried on this band after being modulated and then be transmitted within the atmosphere as channel. Though the background noise of UV-C communication is very low owing to the solar-blind feature, it leads to a large propagation loss. The 370 nm UV provides a much lower propagation loss than that the UV-C does and the recent device technology for UV source on this band is more mature. The fabricated 370 nm AlGaN light-emitting diodes (LEDs) with an aperture size of 45 m exhibit a modulation bandwidth of 165 MHz at 30 mA and a high power of 7 W/cm2 at 230 A/cm2. In order to solve the problem of low power in single UV LED, a UV LED array is presented in.

Keywords: ultraviolet (UV) communication, light-emitting diodes (LEDs), modulation bandwidth, LED array, 370 nm

Procedia PDF Downloads 394
432 Tertiary Level Teachers' Beliefs about Codeswitching

Authors: Hoa Pham

Abstract:

Code switching, which can be described as the use of students’ first language in second language classrooms, has long been a controversial topic in the area of language teaching and second language acquisition. While this has been widely investigated across different contexts, little empirical research has been undertaken in Vietnam. The findings of this study contribute to our understanding of bilingual discourse and code switching practices in content and language integrated classrooms, which has significant implications for language teaching and learning in general and in particular for language pedagogy at tertiary level in Vietnam. This study examines the accounts the teachers articulated for their code switching practices in content-based Business English in Vietnam. Data were collected from five teachers through the use of stimulated recall interviews facilitated by the video data to garner the teachers' cognitive reflection, and allowed them to vocalise the motivations behind their code switching behaviour in particular contexts. The literature has recommended that when participants are provided with a large amount of stimuli or cues, they will experience an original situation again in their imagination with great accuracy. This technique can also provide a valuable "insider" perspective on the phenomenon under investigation which complements the researcher’s "outsider" observation. This can create a relaxed atmosphere during the interview process, which in turn promotes the collection of rich and diverse data. Also, participants can be empowered by this technique as they can raise their own concerns and discuss instances which they find important or interesting. The data generated through this study were analysed using a constant comparative approach. The study found that the teachers indicated their support for the use of code switching in their pedagogical practices. Particularly, as a pedagogical resource, the teachers saw code switching to the L1 playing a key role in facilitating the students' comprehension of both content knowledge and the target language. They believed the use of the L1 accommodates the students' current language competence and content knowledge. They also expressed positive opinions about the role that code switching plays in stimulating students' schematic language and content knowledge, encouraging retention and interest in learning and promoting a positive affective environment in the classroom. The teachers perceived that their use of code switching to the L1 helps them meet the students' language needs and prepares them for their study in subsequent courses and addresses functional needs so that students can cope with English language use outside the classroom. Several factors shaped the teachers' perceptions of their code switching practices, including their accumulated teaching experience, their previous experience as language learners, their theoretical understanding of language teaching and learning, and their knowledge of the teaching context. Code switching was a typical phenomenon in the observed classes and was supported by the teachers in certain contexts. This study reinforces the call in the literature to recognise this practice as a useful instructional resource.

Keywords: codeswitching, language teaching, teacher beliefs, tertiary level

Procedia PDF Downloads 420
431 Lamb Waves Wireless Communication in Healthy Plates Using Coherent Demodulation

Authors: Rudy Bahouth, Farouk Benmeddour, Emmanuel Moulin, Jamal Assaad

Abstract:

Guided ultrasonic waves are used in Non-Destructive Testing (NDT) and Structural Health Monitoring (SHM) for inspection and damage detection. Recently, wireless data transmission using ultrasonic waves in solid metallic channels has gained popularity in some industrial applications such as nuclear, aerospace and smart vehicles. The idea is to find a good substitute for electromagnetic waves since they are highly attenuated near metallic components due to Faraday shielding. The proposed solution is to use ultrasonic guided waves such as Lamb waves as an information carrier due to their capability of propagation for long distances. In addition to this, valuable information about the health of the structure could be extracted simultaneously. In this work, the reliable frequency bandwidth for communication is extracted experimentally from dispersion curves at first. Then, an experimental platform for wireless communication using Lamb waves is described and built. After this, coherent demodulation algorithm used in telecommunications is tested for Amplitude Shift Keying, On-Off Keying and Binary Phase Shift Keying modulation techniques. Signal processing parameters such as threshold choice, number of cycles per bit and Bit Rate are optimized. Experimental results are compared based on the average Bit Error Rate. Results have shown high sensitivity to threshold selection for Amplitude Shift Keying and On-Off Keying techniques resulting a Bit Rate decrease. Binary Phase Shift Keying technique shows the highest stability and data rate between all tested modulation techniques.

Keywords: lamb waves communication, wireless communication, coherent demodulation, bit error rate

Procedia PDF Downloads 229
430 Chaotic Electronic System with Lambda Diode

Authors: George Mahalu

Abstract:

The Chua diode has been configured over time in various ways, using electronic structures like as operational amplifiers (OAs) or devices with gas or semiconductors. When discussing the use of semiconductor devices, tunnel diodes (Esaki diodes) are most often considered, and more recently, transistorized configurations such as lambda diodes. The paper-work proposed here uses in the modeling a lambda diode type configuration consisting of two Junction Field Effect Transistors (JFET). The original scheme is created in the MULTISIM electronic simulation environment and is analyzed in order to identify the conditions for the appearance of evolutionary unpredictability specific to nonlinear dynamic systems with chaos-induced behavior. The chaotic deterministic oscillator is one autonomous type, a fact that places it in the class of Chua’s type oscillators, the only significant and most important difference being the presence of a nonlinear device like the one mentioned structure above. The chaotic behavior is identified both by means of strange attractor-type trajectories and visible during the simulation and by highlighting the hypersensitivity of the system to small variations of one of the input parameters. The results obtained through simulation and the conclusions drawn are useful in the further research of ways to implement such constructive electronic solutions in theoretical and practical applications related to modern small signal amplification structures, to systems for encoding and decoding messages through various modern ways of communication, as well as new structures that can be imagined both in modern neural networks and in those for the physical implementation of some requirements imposed by current research with the aim of obtaining practically usable solutions in quantum computing and quantum computers.

Keywords: chaos, lambda diode, strange attractor, nonlinear system

Procedia PDF Downloads 64
429 The Effect of General Corrosion on the Guided Wave Inspection of the Pipeline

Authors: Shiuh-Kuang Yang, Sheam-Chyun Lin, Jyin-Wen Cheng, Deng-Guei Hsu

Abstract:

The torsional mode of guided wave, T(0,1), has been applied to detect characteristics and defects in pipelines, especially in the cases of coated, elevated and buried pipes. The signals of minor corrosions would be covered by the noise, unfortunately, because the coated material and buried medium always induce a strong attenuation of the guided wave. Furthermore, the guided wave would be attenuated more seriously and make the signals hard to be identified when setting the array ring of the transducers on a general corrosion area of the pipe. The objective of this study is then to discuss the effects of the above-mentioned general corrosion on guided wave tests by experiments and signal processing techniques, based on the use of the finite element method, the two-dimensional Fourier transform and the continuous wavelet transform. Results show that the excitation energy would be reduced when the array ring set on the pipe surface having general corrosion. The non-uniformed contact surface also produces the unwanted asymmetric modes of the propagating guided wave. Some of them are even mixing together with T(0,1) mode and increase the difficulty of measurements, especially when a defect or local corrosion merged in the general corrosion area. It is also showed that the guided waves attenuation are increasing with the increasing corrosion depth or the rising inspection frequency. However, the coherent signals caused by the general corrosion would be decayed with increasing frequency. The results obtained from this research should be able to provide detectors to understand the impact when the array ring set on the area of general corrosion and the way to distinguish the localized corrosion which is inside the area of general corrosion.

Keywords: guided wave, finite element method, two-dimensional fourier transform, wavelet transform, general corrosion, localized corrosion

Procedia PDF Downloads 386
428 Roadway Infrastructure and Bus Safety

Authors: Richard J. Hanowski, Rebecca L. Hammond

Abstract:

Very few studies have been conducted to investigate safety issues associated with motorcoach/bus operations. The current study investigates the impact that roadway infrastructure, including locality, roadway grade, traffic flow and traffic density, have on bus safety. A naturalistic driving study was conducted in the U.S.A that involved 43 motorcoaches. Two fleets participated in the study and over 600,000 miles of naturalistic driving data were collected. Sixty-five bus drivers participated in this study; 48 male and 17 female. The average age of the drivers was 49 years. A sophisticated data acquisition system (DAS) was installed on each of the 43 motorcoaches and a variety of kinematic and video data were continuously recorded. The data were analyzed by identifying safety critical events (SCEs), which included crashes, near-crashes, crash-relevant conflicts, and unintentional lane deviations. Additionally, baseline (normative driving) segments were also identified and analyzed for comparison to the SCEs. This presentation highlights the need for bus safety research and the methods used in this data collection effort. With respect to elements of roadway infrastructure, this study highlights the methods used to assess locality, roadway grade, traffic flow, and traffic density. Locality was determined by manual review of the recorded video for each event and baseline and was characterized in terms of open country, residential, business/industrial, church, playground, school, urban, airport, interstate, and other. Roadway grade was similarly determined through video review and characterized in terms of level, grade up, grade down, hillcrest, and dip. The video was also used to make a determination of the traffic flow and traffic density at the time of the event or baseline segment. For traffic flow, video was used to assess which of the following best characterized the event or baseline: not divided (2-way traffic), not divided (center 2-way left turn lane), divided (median or barrier), one-way traffic, or no lanes. In terms of traffic density, level-of-service categories were used: A1, A2, B, C, D, E, and F. Highlighted in this abstract are only a few of the many roadway elements that were coded in this study. Other elements included lighting levels, weather conditions, roadway surface conditions, relation to junction, and roadway alignment. Note that a key component of this study was to assess the impact that driver distraction and fatigue have on bus operations. In this regard, once the roadway elements had been coded, the primary research questions that were addressed were (i) “What environmental condition are associated with driver choice of engagement in tasks?”, and (ii) “what are the odds of being in a SCE while engaging in tasks while encountering these conditions?”. The study may be of interest to researchers and traffic engineers that are interested in the relationship between roadway infrastructure elements and safety events in motorcoach bus operations.

Keywords: bus safety, motorcoach, naturalistic driving, roadway infrastructure

Procedia PDF Downloads 167
427 On the Qarat Kibrit Salt Dome Faulting System South of Adam, Oman: In Search of Uranium Anomalies

Authors: Alaeddin Ebrahimi, Narasimman Sundararajan, Bernhard Pracejus

Abstract:

Development of salt domes, often a rising from depths of some 10 km or more, causes an intense faulting of the surrounding host rocks (salt tectonics). The fractured rocks then present ideal space for oil that can migrate and get trapped. If such moving of hydrocarbons passes uranium-carrying rock units (e.g., shales), uranium is collected and enriched by organic carbon compounds. Brines from the salt body are also ideal carriers for oxidized uranium species and will further dislocate uranium when in contact with uranium-enriched oils. Uranium then has the potential to mineralize in the vicinity of the dome (blue halite is evidence for radiation having affected salt deposits elsewhere in the world). Based on this knowledge, the Qarat Kibrit salt dome was investigated by a well-established geophysical method like very low frequency electromagnetic (VLF-EM) along five traverses approximately 250 m in length (10 m intervals) in order to identify subsurface fault systems. In-phase and quadrature components of the VLF-EM signal were recorded at two different transmitter frequencies (24.0 and 24.9 kHz). The images of Fraser filtered response of the in-phase components indicate a conductive zone (fault) in the southeast and southwest of the study area. The Karous-Hjelt current density pseudo section delineates subsurface faults at depths between 10 and 40 m. The stacked profiles of the Fraser filtered responses brought out two plausible trends/directions of faults. However, there seems to be no evidence for uranium enrichment has been recorded in this area.

Keywords: salt dome, uranium, fault, in-phase component, quadrature component, Fraser filter, Karous-Hjelt current density

Procedia PDF Downloads 226
426 Biodistribution of Fluorescence-Labelled Epidermal Growth Factor Protein from Slow Release Nanozolid Depots in Mouse

Authors: Stefan Gruden, Charlott Brunmark, Bo Holmqvist, Erwin D. Brenndorfer, Martin Johansson, Jian Liu, Ying Zhao, Niklas Axen, Moustapha Hassan

Abstract:

Aim: The study was designed to evaluate the ability of the calcium sulfate-based NanoZolid® drug delivery technology to locally release the epidermal growth factor (EGF) protein while maintaining its biological activity. Methods: NanoZolid-formulated EGF protein labelled with a near-infrared dye (EGF-NIR) depots or EGF-NIR dissolved in PBS were injected subcutaneously into mice bearing EGF receptor (EGFR) positive human A549 lung cancer tumors inoculated subcutaneously. The release and biodistribution of the EGF-NIR were investigated in vivo longitudinally up to 96 hours post-administration, utilizing whole-body fluorescence imaging. In order to confirm the in vivo findings, histological analysis of tumor cryosections was performed to investigate EGF-NIR fluorescent signal and EGFR expression level by immunofluorescence labelling. Results: The in vivo fluorescence imaging showed a controlled release profile of the EGF-NIR loaded in the NanoZolid depots compared to free EGF-NIR. Histological analysis of the tumors further demonstrated a prevailing distribution of EGF-NIR in regions with high levels of EGFR expression. Conclusion: Calcium sulfate based depots can be used to formulate EGF while maintaining its biological activity, e.g., receptor binding capability. This may have good clinical potential for local delivery of biomolecules to enhance treatment efficacy and minimize systemic adverse effects.

Keywords: bioresorbable, calcium sulfate, controlled release, NanoZolid

Procedia PDF Downloads 153
425 Synthesis of Human Factors Theories and Industry 4.0

Authors: Andrew Couch, Nicholas Loyd, Nathan Tenhundfeld

Abstract:

The rapid emergence of technology observably induces disruptive effects that carry implications for internal organizational dynamics as well as external market opportunities, strategic pressures, and threats. An examination of the historical tendencies of technology innovation shows that the body of managerial knowledge for addressing such disruption is underdeveloped. Fundamentally speaking, the impacts of innovation are unique and situationally oriented. Hence, the appropriate managerial response becomes a complex function that depends on the nature of the emerging technology, the posturing of internal organizational dynamics, the rate of technological growth, and much more. This research considers a particular case of mismanagement, the BP Texas City Refinery explosion of 2005, that carries notable discrepancies on the basis of human factors principles. Moreover, this research considers the modern technological climate (shaped by Industry 4.0 technologies) and seeks to arrive at an appropriate conceptual lens by which human factors principles and Industry 4.0 may be favorably integrated. In this manner, the careful examination of these phenomena helps to better support the sustainment of human factors principles despite the disruptive impacts that are imparted by technological innovation. In essence, human factors considerations are assessed through the application of principles that stem from usability engineering, the Swiss Cheese Model of accident causation, human-automation interaction, signal detection theory, alarm design, and other factors. Notably, this stream of research supports a broader framework in seeking to guide organizations amid the uncertainties of Industry 4.0 to capture higher levels of adoption, implementation, and transparency.

Keywords: Industry 4.0, human factors engineering, management, case study

Procedia PDF Downloads 45
424 Linguistic Competence Analysis and the Development of Speaking Instructional Material

Authors: Felipa M. Rico

Abstract:

Linguistic oral competence plays a vital role in attaining effective communication. Since the English language is considered as universally used language and has a high demand skill needed in the work-place, mastery is the expected output from learners. To achieve this, learners should be given integrated differentiated tasks which help them develop and strengthen the expected skills. This study aimed to develop speaking instructional supplementary material to enhance the English linguistic competence of Grade 9 students in areas of pronunciation, intonation and stress, voice projection, diction and fluency. A descriptive analysis was utilized to analyze the speaking level of performance of the students in order to employ appropriate strategies. There were two sets of respondents: 178 Grade 9 students selected through a stratified sampling and chosen at random. The other set comprised English teachers who evaluated the usefulness of the devised teaching materials. A teacher conducted a speaking test and activities were employed to analyze the speaking needs of students. Observation and recordings were also used to evaluate the students’ performance. The findings revealed that the English pronunciation of the students was slightly unclear at times, but generally fair. There were lapses but generally they rated moderate in intonation and stress, because of other language interference. In terms of voice projection, students have erratic high volume pitch. For diction, the students’ ability to produce comprehensible language is limited, and as to fluency, the choice of vocabulary and use of structure were severely limited. Based on the students’ speaking needs analyses, the supplementary material devised was based on Nunan’s IM model, incorporating context of daily life and global work settings, considering the principle that language is best learned in the actual meaningful situation. To widen the mastery of skill, a rich learning environment, filled with a variety instructional material tends to foster faster acquisition of the requisite skills for sustained learning and development. The role of IM is to encourage information to stick in the learners’ mind, as what is seen is understood more than what is heard. Teachers say they found the IM “very useful.” This implied that English teachers could adopt the materials to improve the speaking skills of students. Further, teachers should provide varied opportunities for students to get involved in real life situations where they could take turns in asking and answering questions and share information related to the activities. This would minimize anxiety among students in the use of the English language.

Keywords: diction, fluency, intonation, instructional materials, linguistic competence

Procedia PDF Downloads 221
423 Structural Damage Detection in a Steel Column-Beam Joint Using Piezoelectric Sensors

Authors: Carlos H. Cuadra, Nobuhiro Shimoi

Abstract:

Application of piezoelectric sensors to detect structural damage due to seismic action on building structures is investigated. Plate-type piezoelectric sensor was developed and proposed for this task. A film-type piezoelectric sheet was attached on a steel plate and covered by a layer of glass. A special glue is used to fix the glass. This glue is a silicone that requires the application of ultraviolet rays for its hardening. Then, the steel plate was set up at a steel column-beam joint of a test specimen that was subjected to bending moment when test specimen is subjected to monotonic load and cyclic load. The structural behavior of test specimen during cyclic loading was verified using a finite element model, and it was found good agreement between both results on load-displacement characteristics. The cross section of steel elements (beam and column) is a box section of 100 mm×100 mm with a thin of 6 mm. This steel section is specified by the Japanese Industrial Standards as carbon steel square tube for general structure (STKR400). The column and beam elements are jointed perpendicularly using a fillet welding. The resulting test specimen has a T shape. When large deformation occurs the glass plate of the sensor device cracks and at that instant, the piezoelectric material emits a voltage signal which would be the indicator of a certain level of deformation or damage. Applicability of this piezoelectric sensor to detect structural damages was verified; however, additional analysis and experimental tests are required to establish standard parameters of the sensor system.

Keywords: piezoelectric sensor, static cyclic test, steel structure, seismic damages

Procedia PDF Downloads 111
422 The Shannon Entropy and Multifractional Markets

Authors: Massimiliano Frezza, Sergio Bianchi, Augusto Pianese

Abstract:

Introduced by Shannon in 1948 in the field of information theory as the average rate at which information is produced by a stochastic set of data, the concept of entropy has gained much attention as a measure of uncertainty and unpredictability associated with a dynamical system, eventually depicted by a stochastic process. In particular, the Shannon entropy measures the degree of order/disorder of a given signal and provides useful information about the underlying dynamical process. It has found widespread application in a variety of fields, such as, for example, cryptography, statistical physics and finance. In this regard, many contributions have employed different measures of entropy in an attempt to characterize the financial time series in terms of market efficiency, market crashes and/or financial crises. The Shannon entropy has also been considered as a measure of the risk of a portfolio or as a tool in asset pricing. This work investigates the theoretical link between the Shannon entropy and the multifractional Brownian motion (mBm), stochastic process which recently is the focus of a renewed interest in finance as a driving model of stochastic volatility. In particular, after exploring the current state of research in this area and highlighting some of the key results and open questions that remain, we show a well-defined relationship between the Shannon (log)entropy and the memory function H(t) of the mBm. In details, we allow both the length of time series and time scale to change over analysis to study how the relation modify itself. On the one hand, applications are developed after generating surrogates of mBm trajectories based on different memory functions; on the other hand, an empirical analysis of several international stock indexes, which confirms the previous results, concludes the work.

Keywords: Shannon entropy, multifractional Brownian motion, Hurst–Holder exponent, stock indexes

Procedia PDF Downloads 89
421 Parallelization of Random Accessible Progressive Streaming of Compressed 3D Models over Web

Authors: Aayushi Somani, Siba P. Samal

Abstract:

Three-dimensional (3D) meshes are data structures, which store geometric information of an object or scene, generally in the form of vertices and edges. Current technology in laser scanning and other geometric data acquisition technologies acquire high resolution sampling which leads to high resolution meshes. While high resolution meshes give better quality rendering and hence is used often, the processing, as well as storage of 3D meshes, is currently resource-intensive. At the same time, web applications for data processing have become ubiquitous owing to their accessibility. For 3D meshes, the advancement of 3D web technologies, such as WebGL, WebVR, has enabled high fidelity rendering of huge meshes. However, there exists a gap in ability to stream huge meshes to a native client and browser application due to high network latency. Also, there is an inherent delay of loading WebGL pages due to large and complex models. The focus of our work is to identify the challenges faced when such meshes are streamed into and processed on hand-held devices, owing to its limited resources. One of the solutions that are conventionally used in the graphics community to alleviate resource limitations is mesh compression. Our approach deals with a two-step approach for random accessible progressive compression and its parallel implementation. The first step includes partition of the original mesh to multiple sub-meshes, and then we invoke data parallelism on these sub-meshes for its compression. Subsequent threaded decompression logic is implemented inside the Web Browser Engine with modification of WebGL implementation in Chromium open source engine. This concept can be used to completely revolutionize the way e-commerce and Virtual Reality technology works for consumer electronic devices. These objects can be compressed in the server and can be transmitted over the network. The progressive decompression can be performed on the client device and rendered. Multiple views currently used in e-commerce sites for viewing the same product from different angles can be replaced by a single progressive model for better UX and smoother user experience. Can also be used in WebVR for commonly and most widely used activities like virtual reality shopping, watching movies and playing games. Our experiments and comparison with existing techniques show encouraging results in terms of latency (compressed size is ~10-15% of the original mesh), processing time (20-22% increase over serial implementation) and quality of user experience in web browser.

Keywords: 3D compression, 3D mesh, 3D web, chromium, client-server architecture, e-commerce, level of details, parallelization, progressive compression, WebGL, WebVR

Procedia PDF Downloads 153
420 Investigating the English Speech Processing System of EFL Japanese Older Children

Authors: Hiromi Kawai

Abstract:

This study investigates the nature of EFL older children’s L2 perceptive and productive abilities using classroom data, in order to find a pedagogical solution to the teaching of L2 sounds at an early stage of learning in a formal school setting. It is still inconclusive whether older children with only EFL formal school instruction at the initial stage of L2 learning are able to attain native-like perception and production in English within the very limited amount of exposure to the target language available. Based on the notion of the lack of study of EFL Japanese children’s acquisition of English segments, the researcher uses a model of L1 speech processing which was developed for investigating L1 English children’s speech and literacy difficulties using a psycholinguistic framework. The model is composed of input channel, output channel, and lexical representation, and examines how a child receives information from spoken or written language, remembers and stores it within the lexical representations and how the child selects and produces spoken or written words. Concerning language universality and language specificity in the language acquisitional process, the aim of finding any sound errors in L1 English children seemed to conform to the author’s intention to find abilities of English sounds in older Japanese children at the novice level of English in an EFL setting. 104 students in Grade 5 (between the ages of 10 and 11 years old) of an elementary school in Tokyo participated in this study. Four tests to measure their perceptive ability and three oral repetition tests to measure their productive ability were conducted with/without reference to lexical representation. All the test items were analyzed to calculate item facility (IF) indices, and correlational analyses and Structural Equation Modeling (SEM) were conducted to examine the relationship between the receptive ability and the productive ability. IF analysis showed that (1) the participants were better at perceiving a segment than producing a segment, (2) they had difficulty in auditory discrimination of paired consonants when one of them does not exist in the Japanese inventory, (3) they had difficulty in both perceiving and producing English vowels, and (4) their L1 loan word knowledge had an influence on their ability to perceive and produce L2 sounds. The result of the Multiple Regression Modeling showed that the two production tests could predict the participants’ auditory ability of real words in English. The result of SEM showed that the hypothesis that perceptive ability affects productive ability was supported. Based on these findings, the author discusses the possible explicit method of teaching English segments to EFL older children in a formal school setting.

Keywords: EFL older children, english segments, perception, production, speech processing system

Procedia PDF Downloads 227
419 Potassium-Phosphorus-Nitrogen Detection and Spectral Segmentation Analysis Using Polarized Hyperspectral Imagery and Machine Learning

Authors: Nicholas V. Scott, Jack McCarthy

Abstract:

Military, law enforcement, and counter terrorism organizations are often tasked with target detection and image characterization of scenes containing explosive materials in various types of environments where light scattering intensity is high. Mitigation of this photonic noise using classical digital filtration and signal processing can be difficult. This is partially due to the lack of robust image processing methods for photonic noise removal, which strongly influence high resolution target detection and machine learning-based pattern recognition. Such analysis is crucial to the delivery of reliable intelligence. Polarization filters are a possible method for ambient glare reduction by allowing only certain modes of the electromagnetic field to be captured, providing strong scene contrast. An experiment was carried out utilizing a polarization lens attached to a hyperspectral imagery camera for the purpose of exploring the degree to which an imaged polarized scene of potassium, phosphorus, and nitrogen mixture allows for improved target detection and image segmentation. Preliminary imagery results based on the application of machine learning algorithms, including competitive leaky learning and distance metric analysis, to polarized hyperspectral imagery, suggest that polarization filters provide a slight advantage in image segmentation. The results of this work have implications for understanding the presence of explosive material in dry, desert areas where reflective glare is a significant impediment to scene characterization.

Keywords: explosive material, hyperspectral imagery, image segmentation, machine learning, polarization

Procedia PDF Downloads 124
418 Automated Computer-Vision Analysis Pipeline of Calcium Imaging Neuronal Network Activity Data

Authors: David Oluigbo, Erik Hemberg, Nathan Shwatal, Wenqi Ding, Yin Yuan, Susanna Mierau

Abstract:

Introduction: Calcium imaging is an established technique in neuroscience research for detecting activity in neural networks. Bursts of action potentials in neurons lead to transient increases in intracellular calcium visualized with fluorescent indicators. Manual identification of cell bodies and their contours by experts typically takes 10-20 minutes per calcium imaging recording. Our aim, therefore, was to design an automated pipeline to facilitate and optimize calcium imaging data analysis. Our pipeline aims to accelerate cell body and contour identification and production of graphical representations reflecting changes in neuronal calcium-based fluorescence. Methods: We created a Python-based pipeline that uses OpenCV (a computer vision Python package) to accurately (1) detect neuron contours, (2) extract the mean fluorescence within the contour, and (3) identify transient changes in the fluorescence due to neuronal activity. The pipeline consisted of 3 Python scripts that could both be easily accessed through a Python Jupyter notebook. In total, we tested this pipeline on ten separate calcium imaging datasets from murine dissociate cortical cultures. We next compared our automated pipeline outputs with the outputs of manually labeled data for neuronal cell location and corresponding fluorescent times series generated by an expert neuroscientist. Results: Our results show that our automated pipeline efficiently pinpoints neuronal cell body location and neuronal contours and provides a graphical representation of neural network metrics accurately reflecting changes in neuronal calcium-based fluorescence. The pipeline detected the shape, area, and location of most neuronal cell body contours by using binary thresholding and grayscale image conversion to allow computer vision to better distinguish between cells and non-cells. Its results were also comparable to manually analyzed results but with significantly reduced result acquisition times of 2-5 minutes per recording versus 10-20 minutes per recording. Based on these findings, our next step is to precisely measure the specificity and sensitivity of the automated pipeline’s cell body and contour detection to extract more robust neural network metrics and dynamics. Conclusion: Our Python-based pipeline performed automated computer vision-based analysis of calcium image recordings from neuronal cell bodies in neuronal cell cultures. Our new goal is to improve cell body and contour detection to produce more robust, accurate neural network metrics and dynamic graphs.

Keywords: calcium imaging, computer vision, neural activity, neural networks

Procedia PDF Downloads 69