Search results for: information and communication technologies
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 15771

Search results for: information and communication technologies

6861 The Regulation of Alternative Dispute Resolution Institutions in Consumer Redress and Enforcement: A South African Perspective

Authors: Jacolien Barnard, Corlia Van Heerden

Abstract:

Effective and accessible consensual dispute resolution and in particular alternative dispute resolution, are central to consumer protection legislation. In this regard, the Consumer Protection Act 68 of 2008 (CPA) of South Africa is no exception. Due to the nature of consumer disputes, alternative dispute resolution (in theory) is an effective vehicle for the adjudication of disputes in a timely manner avoiding overburdening of the courts. The CPA sets down as one of its core purposes the provision of ‘an accessible, consistent, harmonized, effective and efficient system of redress for consumers’ (section 3(1)(h) of the CPA). Section 69 of the Act provides for the enforcement of consumer rights and provides for the National Consumer Commission to be the Central Authority which streamlines, adjudicates and channels disputes to the appropriate forums which include Alternative Dispute Resolution Agents (ADR-agents). The purpose of this paper is to analyze the regulation of these enforcement and redress mechanisms with particular focus on the Central Authority as well as the ADR-agents and their crucial role in successful and efficient adjudication of disputes in South Africa. The South African position will be discussed comparatively with the European Union (EU) position. In this regard, the European Union (EU) Directive on Alternative Dispute Resolution for Consumer Disputes (2013/11/EU) will be discussed (The ADR Directive). The aim of the ADR Directive is to solve contractual disputes between consumers and traders (suppliers or businesses) regardless of whether the agreement was concluded offline or online or whether or not the trader is situated in another member state (Recitals 4-6). The ADR Directive provides for a set of quality requirements that an ADR body or entity tasked with resolving consumer disputes should adhere to in member states which include regulatory mechanisms for control. Transparency, effectiveness, fairness, liberty and legality are all requirements for a successful ADR body and discussed within this chapter III of the Directive. Chapters III and IV govern the importance of information and co-operation. This includes information between ADR bodies and the European Commission (EC) but also between ADR bodies or entities and national authorities enforcing legal acts on consumer protection and traders. (In South Africa the National Consumer Tribunal, Provincial Consumer Protectors and Industry ombuds come to mind). All of which have a responsibility to keep consumers informed. Ultimately the papers aims to provide recommendations as to the successfulness of the current South African position in light of the comparative position in Europe and the highlight the importance of proper regulation of these redress and enforcement institutions.

Keywords: alternative dispute resolution, consumer protection law, enforcement, redress

Procedia PDF Downloads 223
6860 Different Data-Driven Bivariate Statistical Approaches to Landslide Susceptibility Mapping (Uzundere, Erzurum, Turkey)

Authors: Azimollah Aleshzadeh, Enver Vural Yavuz

Abstract:

The main goal of this study is to produce landslide susceptibility maps using different data-driven bivariate statistical approaches; namely, entropy weight method (EWM), evidence belief function (EBF), and information content model (ICM), at Uzundere county, Erzurum province, in the north-eastern part of Turkey. Past landslide occurrences were identified and mapped from an interpretation of high-resolution satellite images, and earlier reports as well as by carrying out field surveys. In total, 42 landslide incidence polygons were mapped using ArcGIS 10.4.1 software and randomly split into a construction dataset 70 % (30 landslide incidences) for building the EWM, EBF, and ICM models and the remaining 30 % (12 landslides incidences) were used for verification purposes. Twelve layers of landslide-predisposing parameters were prepared, including total surface radiation, maximum relief, soil groups, standard curvature, distance to stream/river sites, distance to the road network, surface roughness, land use pattern, engineering geological rock group, topographical elevation, the orientation of slope, and terrain slope gradient. The relationships between the landslide-predisposing parameters and the landslide inventory map were determined using different statistical models (EWM, EBF, and ICM). The model results were validated with landslide incidences, which were not used during the model construction. In addition, receiver operating characteristic curves were applied, and the area under the curve (AUC) was determined for the different susceptibility maps using the success (construction data) and prediction (verification data) rate curves. The results revealed that the AUC for success rates are 0.7055, 0.7221, and 0.7368, while the prediction rates are 0.6811, 0.6997, and 0.7105 for EWM, EBF, and ICM models, respectively. Consequently, landslide susceptibility maps were classified into five susceptibility classes, including very low, low, moderate, high, and very high. Additionally, the portion of construction and verification landslides incidences in high and very high landslide susceptibility classes in each map was determined. The results showed that the EWM, EBF, and ICM models produced satisfactory accuracy. The obtained landslide susceptibility maps may be useful for future natural hazard mitigation studies and planning purposes for environmental protection.

Keywords: entropy weight method, evidence belief function, information content model, landslide susceptibility mapping

Procedia PDF Downloads 129
6859 A Proposal for Systematic Mapping Study of Software Security Testing, Verification and Validation

Authors: Adriano Bessa Albuquerque, Francisco Jose Barreto Nunes

Abstract:

Software vulnerabilities are increasing and not only impact services and processes availability as well as information confidentiality, integrity and privacy, but also cause changes that interfere in the development process. Security test could be a solution to reduce vulnerabilities. However, the variety of test techniques with the lack of real case studies of applying tests focusing on software development life cycle compromise its effective use. This paper offers an overview of how a Systematic Mapping Study (MS) about security verification, validation and test (VVT) was performed, besides presenting general results about this study.

Keywords: software test, software security verification validation and test, security test institutionalization, systematic mapping study

Procedia PDF Downloads 400
6858 Spectrogram Pre-Processing to Improve Isotopic Identification to Discriminate Gamma and Neutrons Sources

Authors: Mustafa Alhamdi

Abstract:

Industrial application to classify gamma rays and neutron events is investigated in this study using deep machine learning. The identification using a convolutional neural network and recursive neural network showed a significant improvement in predication accuracy in a variety of applications. The ability to identify the isotope type and activity from spectral information depends on feature extraction methods, followed by classification. The features extracted from the spectrum profiles try to find patterns and relationships to present the actual spectrum energy in low dimensional space. Increasing the level of separation between classes in feature space improves the possibility to enhance classification accuracy. The nonlinear nature to extract features by neural network contains a variety of transformation and mathematical optimization, while principal component analysis depends on linear transformations to extract features and subsequently improve the classification accuracy. In this paper, the isotope spectrum information has been preprocessed by finding the frequencies components relative to time and using them as a training dataset. Fourier transform implementation to extract frequencies component has been optimized by a suitable windowing function. Training and validation samples of different isotope profiles interacted with CdTe crystal have been simulated using Geant4. The readout electronic noise has been simulated by optimizing the mean and variance of normal distribution. Ensemble learning by combing voting of many models managed to improve the classification accuracy of neural networks. The ability to discriminate gamma and neutron events in a single predication approach using deep machine learning has shown high accuracy using deep learning. The paper findings show the ability to improve the classification accuracy by applying the spectrogram preprocessing stage to the gamma and neutron spectrums of different isotopes. Tuning deep machine learning models by hyperparameter optimization of neural network models enhanced the separation in the latent space and provided the ability to extend the number of detected isotopes in the training database. Ensemble learning contributed significantly to improve the final prediction.

Keywords: machine learning, nuclear physics, Monte Carlo simulation, noise estimation, feature extraction, classification

Procedia PDF Downloads 145
6857 Comparative Study of Various Treatment Positioning Technique: A Site Specific Study-CA. Breast

Authors: Kamal Kaushik, Dandpani Epili, Ajay G. V., Ashutosh, S. Pradhaan

Abstract:

Introduction: Radiation therapy has come a long way over a period of decades, from 2-dimensional radiotherapy to intensity-modulated radiation therapy (IMRT) or VMAT. For advanced radiation therapy, we need better patient position reproducibility to deliver precise and quality treatment, which raises the need for better image guidance technologies for precise patient positioning. This study presents a two tattoo simulation with roll correction technique which is comparable to other advanced patient positioning techniques. Objective: This is a site-specific study is aimed to perform a comparison between various treatment positioning techniques used for the treatment of patients of Ca- Breast undergoing radiotherapy. In this study, we are comparing 5 different positioning methods used for the treatment of ca-breast, namely i) Vacloc with 3 tattoos, ii) Breast board with three tattoos, iii) Thermoplastic cast with three fiducials, iv) Breast board with a thermoplastic mask with 3 tattoo, v) Breast board with 2 tattoos – A roll correction method. Methods and material: All in one (AIO) solution immobilization was used in all patient positioning techniques for immobilization. The process of two tattoo simulations includes positioning of the patient with the help of a thoracic-abdomen wedge, armrest & knee rest. After proper patient positioning, we mark two tattoos on the treatment side of the patient. After positioning, place fiducials as per the clinical borders markers (1) sternum notch (lower border of clavicle head) (2) 2 cm below from contralateral breast (3) midline between 1 & 2 markers (4) mid axillary on the same axis of 3 markers (Marker 3 & 4 should be on the same axis). During plan implementation, a roll depth correction is applied as per the anterior and lateral positioning tattoos, followed by the shifts required for the Isocentre position. The shifts are then verified by SSD on the patient surface followed by radiographic verification using Cone Beam Computed Tomography (CBCT). Results: When all the five positioning techniques were compared all together, the produced shifts in Vertical, Longitudinal and lateral directions are as follows. The observations clearly suggest that the Longitudinal average shifts in two tattoo roll correction techniques are less than every other patient positioning technique. Vertical and lateral Shifts are also comparable to other modern positioning techniques. Concluded: The two tattoo simulation with roll correction technique provides us better patient setup with a technique that can be implemented easily in most of the radiotherapy centers across the developing nations where 3D verification techniques are not available along with delivery units as the shifts observed are quite minimal and are comparable to those with Vacloc and modern amenities.

Keywords: Ca. breast, breast board, roll correction technique, CBCT

Procedia PDF Downloads 130
6856 Micro-CT Imaging Of Hard Tissues

Authors: Amir Davood Elmi

Abstract:

From the earliest light microscope to the most innovative X-ray imaging techniques, all of them have refined and improved our knowledge about the organization and composition of living tissues. The old techniques are time consuming and ultimately destructive to the tissues under the examination. In recent few decades, thanks to the boost of technology, non-destructive visualization techniques, such as X-ray computed tomography (CT), magnetic resonance imaging (MRI), selective plane illumination microscopy (SPIM), and optical projection tomography (OPT), have come to the forefront. Among these techniques, CT is excellent for mineralized tissues such as bone or dentine. In addition, CT it is faster than other aforementioned techniques and the sample remains intact. In this article, applications, advantages, and limitations of micro-CT is discussed, in addition to some information about micro-CT of soft tissue.

Keywords: Micro-CT, hard tissue, bone, attenuation coefficient, rapid prototyping

Procedia PDF Downloads 138
6855 Detailed Analysis of Mechanism of Crude Oil and Surfactant Emulsion

Authors: Riddhiman Sherlekar, Umang Paladia, Rachit Desai, Yash Patel

Abstract:

A number of surfactants which exhibit ultra-low interfacial tension and an excellent microemulsion phase behavior with crude oils of low to medium gravity are not sufficiently soluble at optimum salinity to produce stable aqueous solutions. Such solutions often show phase separation after a few days at reservoir temperature, which does not suffice the purpose and the time is short when compared to the residence time in a reservoir for a surfactant flood. The addition of polymer often exacerbates the problem although the poor stability of the surfactant at high salinity remains a pivotal issue. Surfactants such as SDS, Ctab with large hydrophobes produce lowest IFT, but are often not sufficiently water soluble at desired salinity. Hydrophilic co-solvents and/or co-surfactants are needed to make the surfactant-polymer solution stable at the desired salinity. This study focuses on contrasting the effect of addition of a co-solvent in stability of a surfactant –oil emulsion. The idea is to use a co-surfactant to increase stability of an emulsion. Stability of the emulsion is enhanced because of creation of micro-emulsion which is verified both visually and with the help of particle size analyzer at varying concentration of salinity, surfactant and co-surfactant. A lab-experimental method description is provided and the method is described in detail to permit readers to emulate all results. The stability of the oil-water emulsion is visualized with respect to time, temperature, salinity of the brine and concentration of the surfactant. Nonionic surfactant TX-100 when used as a co-surfactant increases the stability of the oil-water emulsion. The stability of the prepared emulsion is checked by observing the particle size distribution. For stable emulsion in volume% vs particle size curve, the peak should be obtained for particle size of 5-50 nm while for the unstable emulsion a bigger sized particles are observed. The UV-Visible spectroscopy is also used to visualize the fraction of oil that plays important role in the formation of micelles in stable emulsion. This is important as the study will help us to decide applicability of the surfactant based EOR method for a reservoir that contains a specific type of crude. The use of nonionic surfactant as a co-surfactant would also increase the efficiency of surfactant EOR. With the decline in oil discoveries during the last decades it is believed that EOR technologies will play a key role to meet the energy demand in years to come. Taking this into consideration, the work focuses on the optimization of the secondary recovery(Water flooding) with the help of surfactant and/or co-surfactants by creating desired conditions in the reservoir.

Keywords: co-surfactant, enhanced oil recovery, micro-emulsion, surfactant flooding

Procedia PDF Downloads 249
6854 The Student's Satisfaction toward Web Based Instruction on Puppet Show

Authors: Piyanut Suchit

Abstract:

The purposes of this study was to investigate students’ satisfaction learning with the web based instruction on the puppet show. The population of this study includes 53 students in the Program of Library and Information Sciences who registered in the subject of Puppet for Assisting Learning Development in semester 2/2011, Suansunandha Rajabhat University, Bangkok, Thailand. The research instruments consist of web based instruction on the puppet show, and questionnaires for students’ satisfaction. The research statistics includes arithmetic mean, and standard deviation. The results revealed that the students reported very high satisfaction with mean = 4.63, SD = 0.52, on the web based instruction.

Keywords: puppet show, web based instruction, satisfaction, Suansunandha Rajabhat University

Procedia PDF Downloads 383
6853 Teachers' Perceptions of Physical Education and Sports Calendar and Conducted in the Light of the Objective of the Lesson Approach Competencies

Authors: Chelali Mohammed

Abstract:

In the context of the application of the competency-based approach in the system educational Algeria, the price of physical education and sport must privilege the acquisition of learning approaches and especially the approach science, which from problem situations, research and develops him information processing and application of knowledge and know-how in new situations in the words of ‘JOHN DEWEY’ ‘learning by practice’. And to achieve these goals and make teaching more EPS motivating, consistent and concrete, it is appropriate to perform a pedagogical approach freed from the constraints and open to creativity and student-centered in the light of the competency approach adopted in the formal curriculum. This approach is not unusual, but we think it is a highly professional nature requires the competence of the teacher.

Keywords: approach competencies, physical, education, teachers

Procedia PDF Downloads 599
6852 Revised Risk Priority Number in Failure Mode and Effects Analysis Model from the Perspective of Healthcare System

Authors: Fatemeh Rezaei, Mohammad H. Yarmohammadian, Masoud Ferdosi, Abbas Haghshnas

Abstract:

Background: Failure Modes and Effect Analysis is now having known as the main methods of risk assessment and the accreditation requirements for many organizations. The Risk Priority Number (RPN) approach is generally preferred, especially for its easiness of use. Indeed it does not require statistical data, but it is based on subjective evaluations given by the experts about the Occurrence (O i), the Severity (Si) and the Detectability (D i) of each cause of failure. Methods: This study is a quantitative – qualitative research. In terms of qualitative dimension, method of focus groups with inductive approach is used. To evaluate the results of the qualitative study, quantitative assessment was conducted to calculate RPN score. Results; We have studied patient’s journey process in surgery ward and the most important phase of the process determined Transport of the patient from the holding area to the operating room. Failures of the phase with the highest priority determined by defining inclusion criteria included severity (clinical effect, claim consequence, waste of time and financial loss), occurrence (time- unit occurrence and degree of exposure to risk) and preventability (degree of preventability and defensive barriers) and quantifying risks priority criteria in the context of RPN index. Ability of improved RPN reassess by root cause (RCA) analysis showed some variations. Conclusions: Finally, It could be concluded that understandable criteria should have been developed according to personnel specialized language and communication field. Therefore, participation of both technical and clinical groups is necessary to modify and apply these models.

Keywords: failure mode, effects analysis, risk priority number(RPN), health system, risk assessment

Procedia PDF Downloads 309
6851 Design Thinking and Requirements Engineering in Application Development: Case Studies in Brazil

Authors: V. Prodocimo, A. Malucelli, S. Reinehr

Abstract:

Organizations, driven by business digitization, have in software the main core of value generation and the main channel of communication with their clients. The software, as well as responding to momentary market needs, spans an extensive product family, ranging from mobile applications to multilateral platforms. Thus, the software specification needs to represent solutions focused on consumer problems and market needs. However, requirements engineering, whose approach is strongly linked to technology, becomes deficient and ineffective when the problem is not well defined or when looking for an innovative solution, thus needing a complementary approach. Research has cited the combination of design thinking and requirements engineering, many correlating design thinking as a support technique for the elicitation step, however, little is known about the real benefits and challenges that this combination can bring. From the point of view of the development process, there is little empirical evidence of how Design Thinking interactions with requirements engineering occur. Given this scenario, this paper aims to understand how design thinking practices are applied in each of the requirements engineering stages in software projects. To elucidate these interactions, a qualitative and exploratory research was carried out through the application of the case study method in IT organizations in Brazil that work in the development of software projects. The results indicate that design thinking has aided requirements engineering, both in projects that adopt agile methods and those that adopt the waterfall process, bringing a complementary thought that seeks to build the best software solution design for business problems. It was also possible to conclude that organizations choose to use design thinking not based on a specific software family (e.g. mobile or desktop applications), but given the characteristics of the software projects, such as: vague nature of the problem, complex problems and/or need for innovative solutions.

Keywords: software engineering, requirements engineering, design thinking, innovative solutions

Procedia PDF Downloads 122
6850 Gaming Mouse Redesign Based on Evaluation of Pragmatic and Hedonic Aspects of User Experience

Authors: Thedy Yogasara, Fredy Agus

Abstract:

In designing a product, it is currently crucial to focus not only on the product’s usability based on performance measures, but also on user experience (UX) that includes pragmatic and hedonic aspects of product use. These aspects play a significant role in fulfillment of user needs, both functionally and psychologically. Pragmatic quality refers to as product’s perceived ability to support the fulfillment of behavioral goals. It is closely linked to functionality and usability of the product. In contrast, hedonic quality is product’s perceived ability to support the fulfillment of psychological needs. Hedonic quality relates to the pleasure of ownership and use of the product, including stimulation for personal development and communication of user’s identity to others through the product. This study evaluates the pragmatic and hedonic aspects of gaming mice G600 and Razer Krait using AttrakDiff tool to create an improved design that is able to generate positive UX. AttrakDiff is a method that measures pragmatic and hedonic scores of a product with a scale between -3 to +3 through four attributes (i.e. Pragmatic Quality, Hedonic Quality-Identification, Hedonic Quality-Stimulation, and Attractiveness), represented by 28 pairs of opposite words. Based on data gathered from 15 participants, it is identified that gaming mouse G600 needs to be redesigned because of its low grades (pragmatic score: -0.838, hedonic score: 1, attractiveness score: 0.771). The redesign process focuses on the attributes with poor scores and takes into account improvement suggestions collected from interview with the participants. The redesigned mouse G600 is evaluated using the previous method. The result shows higher scores in pragmatic quality (1.929), hedonic quality (1.703), and attractiveness (1.667), indicating that the redesigned mouse is more capable of creating pleasurable experience of product use.

Keywords: AttrakDiff, hedonic aspect, pragmatic aspect, product design, user experience

Procedia PDF Downloads 153
6849 Road Traffic Noise Mapping for Riyadh City Using GIS and Lima

Authors: Khalid A. Alsaif, Mosaad A. Foda

Abstract:

The primary objective of this study is to develop the first round of road traffic noise maps for Riyadh City using Geographical Information Systems (GIS) and software LimA 7810 predictor. The road traffic data were measured or estimated as accurate as possible in order to obtain reliable noise maps. Meanwhile, the attributes of the roads and buildings are automatically exported from GIS. The simulation results at some chosen locations are validated by actual field measurements, which are obtained by a system that consists of a sound level meter, a GPS receiver and a database to manage the measured data. The results show that the average error between the predicted and measured noise levels is below 3.0 dB.

Keywords: noise pollution, road traffic noise, LimA predictor, GIS

Procedia PDF Downloads 402
6848 Ground States of Structure of Even ¹⁰⁴-¹⁰⁶ Ru Isotopes

Authors: I. Hossain, Huda H. Kassim, Fadhil I. Sharrad, Said A. Mansour

Abstract:

In this conference, we apply the interacting boson model-1 (IBM-1) formula for U(5) symmetry in order to calculate the energy levels and reduced transition probabilities for a few yrast transitions in Ru with neutron N=60, 62. The neutron rich even-even isotopes of Ru are very interesting to investigate using IBM-1, because even 104,106Ru isotopes are great consequence due to excited near the magic number 50. The calculation of ground state band and B(E2) values using IBM-1 for Z=44 are not calculated to describe the valuable information of nuclear structure by U(5) limit. The parameters in the formula are deduced based on the experimental energy level and value of B(E2, 2+->0+). The yrast states and transition strength B(E2) from 1st 4+ to 1st 2+, 1st 6+ to 1st 4+ and 1st 8+ to 1st 6+ states of Ru for even N= 60, 62 were calculated. The quadrupole moments, deformation parameters and U(5) limit were discussed for those nuclei.

Keywords: B(E2), energy level, ¹⁰⁴Ru, ¹⁰⁶Ru

Procedia PDF Downloads 343
6847 Activity-Based Costing in the Hospitality Industry: A Case Study in a Hotel

Authors: Bita Mashayekhi, Mohammad Ara

Abstract:

The purpose of this study is to provide some empirical evidence about implementing Activity-Based Costing (ABC) in the hospitality industry in Iran. For this purpose, we consider the Tabriz International Hotel as our sample hotel and then gather the relevant data from its cost accounting system in 2012. Then, we use ABC as our costing method and compare the cost of each service unit with that cost which had been extracted for the traditional costing method. The results show a different cost per unit for two methods. Also, because of its more precise and detailed provided information, an ABC system facilitates the decision-making process for managers on decisions related to profitability analysis, budgeting, pricing, and so on.

Keywords: Activity-Based Costing (ABC), activity, cost driver, hospitality industry

Procedia PDF Downloads 296
6846 A Comparative Study between Japan and the European Union on Software Vulnerability Public Policies

Authors: Stefano Fantin

Abstract:

The present analysis outcomes from the research undertaken in the course of the European-funded project EUNITY, which targets the gaps in research and development on cybersecurity and privacy between Europe and Japan. Under these auspices, the research presents a study on the policy approach of Japan, the EU and a number of Member States of the Union with regard to the handling and discovery of software vulnerabilities, with the aim of identifying methodological differences and similarities. This research builds upon a functional comparative analysis of both public policies and legal instruments from the identified jurisdictions. The result of this analysis is based on semi-structured interviews with EUNITY partners, as well as by the participation of the researcher to a recent report from the Center for EU Policy Study on software vulnerability. The European Union presents a rather fragmented legal framework on software vulnerabilities. The presence of a number of different legislations at the EU level (including Network and Information Security Directive, Critical Infrastructure Directive, Directive on the Attacks at Information Systems and the Proposal for a Cybersecurity Act) with no clear focus on such a subject makes it difficult for both national governments and end-users (software owners, researchers and private citizens) to gain a clear understanding of the Union’s approach. Additionally, the current data protection reform package (general data protection regulation), seems to create legal uncertainty around security research. To date, at the member states level, a few efforts towards transparent practices have been made, namely by the Netherlands, France, and Latvia. This research will explain what policy approach such countries have taken. Japan has started implementing a coordinated vulnerability disclosure policy in 2004. To date, two amendments can be registered on the framework (2014 and 2017). The framework is furthermore complemented by a series of instruments allowing researchers to disclose responsibly any new discovery. However, the policy has started to lose its efficiency due to a significant increase in reports made to the authority in charge. To conclude, the research conducted reveals two asymmetric policy approaches, time-wise and content-wise. The analysis therein will, therefore, conclude with a series of policy recommendations based on the lessons learned from both regions, towards a common approach to the security of European and Japanese markets, industries and citizens.

Keywords: cybersecurity, vulnerability, European Union, Japan

Procedia PDF Downloads 152
6845 Effects of Global Validity of Predictive Cues upon L2 Discourse Comprehension: Evidence from Self-paced Reading

Authors: Binger Lu

Abstract:

It remains unclear whether second language (L2) speakers could use discourse context cues to predict upcoming information as native speakers do during online comprehension. Some researchers propose that L2 learners may have a reduced ability to generate predictions during discourse processing. At the same time, there is evidence that discourse-level cues are weighed more heavily in L2 processing than in L1. Previous studies showed that L1 prediction is sensitive to the global validity of predictive cues. The current study aims to explore whether and to what extent L2 learners can dynamically and strategically adjust their prediction in accord with the global validity of predictive cues in L2 discourse comprehension as native speakers do. In a self-paced reading experiment, Chinese native speakers (N=128), C-E bilinguals (N=128), and English native speakers (N=128) read high-predictable (e.g., Jimmy felt thirsty after running. He wanted to get some water from the refrigerator.) and low-predictable (e.g., Jimmy felt sick this morning. He wanted to get some water from the refrigerator.) discourses in two-sentence frames. The global validity of predictive cues was manipulated by varying the ratio of predictable (e.g., Bill stood at the door. He opened it with the key.) and unpredictable fillers (e.g., Bill stood at the door. He opened it with the card.), such that across conditions, the predictability of the final word of the fillers ranged from 100% to 0%. The dependent variable was reading time on the critical region (the target word and the following word), analyzed with linear mixed-effects models in R. C-E bilinguals showed reliable prediction across all validity conditions (β = -35.6 ms, SE = 7.74, t = -4.601, p< .001), and Chinese native speakers showed significant effect (β = -93.5 ms, SE = 7.82, t = -11.956, p< .001) in two of the four validity conditions (namely, the High-validity and MedLow conditions, where fillers ended with predictable words in 100% and 25% cases respectively), whereas English native speakers didn’t predict at all (β = -2.78 ms, SE = 7.60, t = -.365, p = .715). There was neither main effect (χ^²(3) = .256, p = .968) nor interaction (Predictability: Background: Validity, χ^²(3) = 1.229, p = .746; Predictability: Validity, χ^²(3) = 2.520, p = .472; Background: Validity, χ^²(3) = 1.281, p = .734) of Validity with speaker groups. The results suggest that prediction occurs in L2 discourse processing but to a much less extent in L1, witha significant effect in some conditions of L1 Chinese and anull effect in L1 English processing, consistent with the view that L2 speakers are more sensitive to discourse cues compared with L1 speakers. Additionally, the pattern of L1 and L2 predictive processing was not affected by the global validity of predictive cues. C-E bilinguals’ predictive processing could be partly transferred from their L1, as prior research showed that discourse information played a more significant role in L1 Chinese processing.

Keywords: bilingualism, discourse processing, global validity, prediction, self-paced reading

Procedia PDF Downloads 135
6844 Ozonation as an Effective Method to Remove Pharmaceuticals from Biologically Treated Wastewater of Different Origin

Authors: Agne Jucyte Cicine, Vytautas Abromaitis, Zita Rasuole Gasiunaite, Mindaugas Zilius

Abstract:

Pharmaceutical pollution in aquatic environments has become a growing concern. Various active pharmaceutical ingredient (API) residues: hormones, antibiotics, or/and psychiatric drugs, have already been discovered in different environmental compartments. Due to ineffective wastewater treatment technologies, an underestimated amount of APIs can enter the ecosystem by discharged treated wastewater. Especially, psychiatric compounds, such as carbamazepine (CBZ) and venlafaxine (VNX), persist in effluent even post-treatment. Therefore, these pharmaceuticals usually exceed safe environmental levels and pose risks to the aquatic environment, particularly, to sensitive ecosystems such as the Baltic Sea. CBZ, known for its chemical stability and long biodegradation time, accumulates in the environment, threatening aquatic life and human health through the food chain. As the use of medication rises, there is an urgent need for advanced wastewater treatment to reduce pharmaceutical contamination and meet future regulatory requirements. In this study, we tested advanced oxidation technology using ozone to remove two commonly used psychiatric drugs, carbamazepine, and venlafaxine, from biologically treated wastewater effluent. Additionally, general water quality parameters (SPM, DOC, COD) and bacterial contamination were analyzed. Three wastewater treatment plants (WWTPs) were selected to represent varying dominant anthropogenic activities: 1) resort, 2) resort and residential, and 3) residential, industrial, and resort. Wastewater samples for the experiment were collected during the summer season after mechanical and biological treatment and ozonated for 5, 10, and 15 minutes. Pharmaceutical levels in this study exceeded the predicted no-effect concentration (PNEC) of 500 and 90 ng L−1, for CBZ and VNX, respectively in all WWTPs, except CBZ in WWTP 1. Initial CBZ contamination was found to be lower in WWTP 1 (427.4 ng L-1), compared with WWTP 2 (1266.5 ng L-1) and 3 (119.2 ng L-1). VNX followed a similar trend with concentrations of 341.2 ng L-1, 361.4 ng L-1, and 390.0 ng L-1, respectively, for WWTPs 1, 2, and 3. It was determined that CBZ was not detected in the effluent after 5 minutes of ozonation in any of the WWTPs. Contrarily, VNX, was still detected after 5, 10, and 15 minutes of treatment with ozone, however under the limits of quantification (LOD) (<5ng L-1). Additionally, general pollution of SPM, DOC, COD, and bacterial contamination was reduced notably after 5 minutes of treatment with ozone. Although initial pharmaceutical levels exceeded PNECs, indicating ongoing environmental risks, ozonation demonstrated high efficiency in reducing pharmaceutical and general contamination in wastewater with different pollution matrices.

Keywords: Baltic Sea, ozonation, pharmaceuticals, wastewater treatment plants.

Procedia PDF Downloads 13
6843 Semantic Data Schema Recognition

Authors: Aïcha Ben Salem, Faouzi Boufares, Sebastiao Correia

Abstract:

The subject covered in this paper aims at assisting the user in its quality approach. The goal is to better extract, mix, interpret and reuse data. It deals with the semantic schema recognition of a data source. This enables the extraction of data semantics from all the available information, inculding the data and the metadata. Firstly, it consists of categorizing the data by assigning it to a category and possibly a sub-category, and secondly, of establishing relations between columns and possibly discovering the semantics of the manipulated data source. These links detected between columns offer a better understanding of the source and the alternatives for correcting data. This approach allows automatic detection of a large number of syntactic and semantic anomalies.

Keywords: schema recognition, semantic data profiling, meta-categorisation, semantic dependencies inter columns

Procedia PDF Downloads 414
6842 Insight into the Visual Attentional Correlates Underpinning Autistic-Like Traits in Fragile X and Down Syndrome

Authors: Jennifer M. Glennon, Hana D'Souza, Luke Mason, Annette Karmiloff-Smith, Michael S. C. Thomas

Abstract:

Genetic syndrome groups that feature high rates of autism comorbidity, like Down syndrome (DS) and fragile X syndrome (FXS), have been presented as useful models for understanding risk and protective factors involved in the emergence of autistic traits. Yet despite reaching clinical thresholds, these ‘syndromic’ forms of autism appear to differ in important ways from the idiopathic or ‘non-syndromic’ autism phenotype. To uncover the true nature of these comorbidities, it is necessary to extend definitions of autism to include the cognitive characteristics of the disorder and to then apply this broadened conceptualisation to the study of syndromic autism profiles. The current study employs a variety of well-established eye-tracking paradigms to assess visual attentional performance in children with DS and FXS who reach thresholds for autism on the Social Communication Questionnaire. It investigates whether autism profiles in these children are accompanied by visual orienting difficulties (‘sticky attention’), decreased social attention, and enhanced visual search performance, all of which are characteristic of the idiopathic autism phenotype. Data is collected from children with DS and FXS aged between 6 and 10 years, in addition to two control groups matched on age and intellectual ability (i.e., children with idiopathic autism and neurotypical controls). Cross-sectional developmental trajectory analyses are conducted to enable visuo-attentional profile comparisons. Significant differences in the visuo-attentional processes underpinning autism presentations in children with FXS and DS are hypothesised, supporting notions of syndrome specificity. The study provides insight into the complex heterogeneity associated with syndromic autism presentations and autism per se, with clinical implications for the utility of autism intervention programmes in DS and FXS populations.

Keywords: autism, down syndrome, fragile X syndrome, eye tracking

Procedia PDF Downloads 233
6841 Passing-On Cultural Heritage Knowledge: Entrepreneurial Approaches for a Higher Educational Sustainability

Authors: Ioana Simina Frincu

Abstract:

As institutional initiatives often fail to provide good practices when it comes to heritage management or to adapt to the changing environment in which they function and to the audiences they address, private actions represent viable strategies for sustainable knowledge acquisition. Information dissemination to future generations is one of the key aspects in preserving cultural heritage and is successfully feasible even in the absence of original artifacts. Combined with the (re)discovery of natural landscape, open-air exploratory approaches (archeoparks) versus an enclosed monodisciplinary rigid framework (traditional museums) are more likely to 'speak the language' of a larger number of people, belonging to a variety of categories, ages, and professions. Interactive sites are efficient ways of stimulating heritage awareness and increasing the number of visitors of non-interactive/static cultural institutions owning original pieces of history, delivering specialized information, and making continuous efforts to preserve historical evidence (relics, manuscripts, etc.). It is high time entrepreneurs took over the role of promoting cultural heritage, bet it under a more commercial yet more attractive form (business). Inclusive, participatory type of activities conceived by experts from different domains/fields (history, anthropology, tourism, sociology, business management, integrative sustainability, etc.) have better chances to ensure long term cultural benefits for both adults and children, especially when and where the educational discourse fails. These unique self-experience leisure activities, which offer everyone the opportunity to recreate history by him-/her-self, to relive the ancestors’ way of living, surviving and exploring should be regarded not as pseudo-scientific approaches but as important pre-steps to museum experiences. In order to support this theory, focus will be laid on two different examples: one dynamic, in the outdoors (the Boario Terme Archeopark from Italy) and one experimental, held indoor (the reconstruction of the Neolithic sanctuary of Parta, Romania as part of a transdisciplinary academic course) and their impact on young generations. The conclusion of this study shows that the increasingly lower engagement of youth (students) in discovering and understanding history, archaeology, and heritage can be revived by entrepreneurial projects.

Keywords: archeopark, educational tourism, open air museum, Parta sanctuary, prehistory

Procedia PDF Downloads 133
6840 Fair Value Accounting and Evolution of the Ohlson Model

Authors: Mohamed Zaher Bouaziz

Abstract:

Our study examines the Ohlson Model, which links a company's market value to its equity and net earnings, in the context of the evolution of the Canadian accounting model, characterized by more extensive use of fair value and a broader measure of performance after IFRS adoption. Our hypothesis is that if equity is reported at its fair value, this valuation is closely linked to market capitalization, so the weight of earnings weakens or even disappears in the Ohlson Model. Drawing on Canada's adoption of the International Financial Reporting Standards (IFRS), our results support our hypothesis that equity appears to include most of the relevant information for investors, while earnings have become less important. However, the predictive power of earnings does not disappear.

Keywords: fair value accounting, Ohlson model, IFRS adoption, value-relevance of equity and earnings

Procedia PDF Downloads 183
6839 Comparative Dielectric Properties of 1,2-Dichloroethane with n-Methylformamide and n,n-Dimethylformamide Using Time Domain Reflectometry Technique in Microwave Frequency

Authors: Shagufta Tabassum, V. P. Pawar, jr., G. N. Shinde

Abstract:

The study of dielectric relaxation properties of polar liquids in the binary mixture has been carried out at 10, 15, 20 and 25 ºC temperatures for 11 different concentrations using time domain reflectometry technique. The dielectric properties of a solute-solvent mixture of polar liquids in the frequency range of 10 MHz to 30 GHz gives the information regarding formation of monomers and multimers and also an interaction between the molecules of the liquid mixture under study. The dielectric parameters have been obtained by the least squares fit method using the Debye equation characterized by a single relaxation time without relaxation time distribution.

Keywords: excess properties, relaxation time, static dielectric constant, and time domain reflectometry technique

Procedia PDF Downloads 148
6838 Comparing Emotion Recognition from Voice and Facial Data Using Time Invariant Features

Authors: Vesna Kirandziska, Nevena Ackovska, Ana Madevska Bogdanova

Abstract:

The problem of emotion recognition is a challenging problem. It is still an open problem from the aspect of both intelligent systems and psychology. In this paper, both voice features and facial features are used for building an emotion recognition system. A Support Vector Machine classifiers are built by using raw data from video recordings. In this paper, the results obtained for the emotion recognition are given, and a discussion about the validity and the expressiveness of different emotions is presented. A comparison between the classifiers build from facial data only, voice data only and from the combination of both data is made here. The need for a better combination of the information from facial expression and voice data is argued.

Keywords: emotion recognition, facial recognition, signal processing, machine learning

Procedia PDF Downloads 312
6837 Copyright Clearance for Artificial Intelligence Training Data: Challenges and Solutions

Authors: Erva Akin

Abstract:

– The use of copyrighted material for machine learning purposes is a challenging issue in the field of artificial intelligence (AI). While machine learning algorithms require large amounts of data to train and improve their accuracy and creativity, the use of copyrighted material without permission from the authors may infringe on their intellectual property rights. In order to overcome copyright legal hurdle against the data sharing, access and re-use of data, the use of copyrighted material for machine learning purposes may be considered permissible under certain circumstances. For example, if the copyright holder has given permission to use the data through a licensing agreement, then the use for machine learning purposes may be lawful. It is also argued that copying for non-expressive purposes that do not involve conveying expressive elements to the public, such as automated data extraction, should not be seen as infringing. The focus of such ‘copy-reliant technologies’ is on understanding language rules, styles, and syntax and no creative ideas are being used. However, the non-expressive use defense is within the framework of the fair use doctrine, which allows the use of copyrighted material for research or educational purposes. The questions arise because the fair use doctrine is not available in EU law, instead, the InfoSoc Directive provides for a rigid system of exclusive rights with a list of exceptions and limitations. One could only argue that non-expressive uses of copyrighted material for machine learning purposes do not constitute a ‘reproduction’ in the first place. Nevertheless, the use of machine learning with copyrighted material is difficult because EU copyright law applies to the mere use of the works. Two solutions can be proposed to address the problem of copyright clearance for AI training data. The first is to introduce a broad exception for text and data mining, either mandatorily or for commercial and scientific purposes, or to permit the reproduction of works for non-expressive purposes. The second is that copyright laws should permit the reproduction of works for non-expressive purposes, which opens the door to discussions regarding the transposition of the fair use principle from the US into EU law. Both solutions aim to provide more space for AI developers to operate and encourage greater freedom, which could lead to more rapid innovation in the field. The Data Governance Act presents a significant opportunity to advance these debates. Finally, issues concerning the balance of general public interests and legitimate private interests in machine learning training data must be addressed. In my opinion, it is crucial that robot-creation output should fall into the public domain. Machines depend on human creativity, innovation, and expression. To encourage technological advancement and innovation, freedom of expression and business operation must be prioritised.

Keywords: artificial intelligence, copyright, data governance, machine learning

Procedia PDF Downloads 80
6836 Stress and Rhythm in the Educated Nigerian Accent of English

Authors: Nkereke M. Essien

Abstract:

The intention of this paper is to examine stress in the Educated Nigerian Accent of English (ENAE) with the aim of analyzing stress and rhythmic patterns of Nigerian English. Our aim also is to isolate differences and similarities in the stress patterns studied and also know what forms the accent of these Educated Nigerian English (ENE) which marks them off from other groups or English’s of the world, to ascertain and characterize it and to provide documented evidence for its existence. Nigerian stress and rhythmic patterns are significantly different from the British English stress and rhythmic patterns consequently, the educated Nigerian English (ENE) features more stressed syllables than the native speakers’ varieties. The excessive stressed of syllables causes a contiguous “Ss” in the rhythmic flow of ENE, and this brings about a “jerky rhythm’ which distorts communication. To ascertain this claim, ten (10) Nigerian speakers who are educated in the English Language were selected by a stratified Random Sampling technique from two Federal Universities in Nigeria. This classification belongs to the education to the educated class or standard variety. Their performance was compared to that of a Briton (control). The Metrical system of analysis was used. The respondents were made to read some words and utterance which was recorded and analyzed perceptually, statistically and acoustically using the one-way Analysis of Variance (ANOVA). The Turky-Kramer Post Hoc test, the Wilcoxon Matched Pairs Signed Ranks test, and the Praat analysis software were used in the analysis. It was revealed from our findings that the Educated Nigerian English speakers feature more stressed syllables in their productions by spending more time in pronouncing stressed syllables and sometimes lesser time in pronouncing the unstressed syllables. Their overall tempo was faster. The ENE speakers used tone to mark prominence while the native speaker used stress to mark pronounce, typified by the control. We concluded that the stress pattern of the ENE speakers was significantly different from the native speaker’s variety represented by the control’s performance.

Keywords: accent, Nigerian English, rhythm, stress

Procedia PDF Downloads 236
6835 Health Transformation Program and Effects on Health Expenditures

Authors: Zeynep Karacor, Rahime Hulya Ozturk

Abstract:

In recent years, the rise of population density and the problem of aging population took attention to the health expenditures. In Turkey, some regulations and infrastructure changes in health sector have occurred. These changes are called Health Transformation Program. The productivity of health services, patient satisfaction, quality of services are tried to be improved with this program. Some radical changes are applied in Turkish economy in this context. The aim of this paper is to present the effects of Health Transformation Program on health expenditures. In the first part of the paper, some information’s about health system and applications in Turkey are discussed. In the second part, the aims of Health Transformation Program are explained. And in the third part the effects of Health Transformation Program on health expenditures are examined.

Keywords: health transformation program, Turkey, health services, health expenditures

Procedia PDF Downloads 384
6834 RAPD Analysis of Genetic Diversity of Castor Bean

Authors: M. Vivodík, Ž. Balážová, Z. Gálová

Abstract:

The aim of this work was to detect genetic variability among the set of 40 castor genotypes using 8 RAPD markers. Amplification of genomic DNA of 40 genotypes, using RAPD analysis, yielded in 66 fragments, with an average of 8.25 polymorphic fragments per primer. Number of amplified fragments ranged from 3 to 13, with the size of amplicons ranging from 100 to 1200 bp. Values of the polymorphic information content (PIC) value ranged from 0.556 to 0.895 with an average of 0.784 and diversity index (DI) value ranged from 0.621 to 0.896 with an average of 0.798. The dendrogram based on hierarchical cluster analysis using UPGMA algorithm was prepared and analyzed genotypes were grouped into two main clusters and only two genotypes could not be distinguished. Knowledge on the genetic diversity of castor can be used for future breeding programs for increased oil production for industrial uses.

Keywords: dendrogram, polymorphism, RAPD technique, Ricinus communis L.

Procedia PDF Downloads 467
6833 Representations of Childcare Robots as a Controversial Issue

Authors: Raya A. Jones

Abstract:

This paper interrogates online representations of robot companions for children, including promotional material by manufacturers, media articles and technology blogs. The significance of the study lies in its contribution to understanding attitudes to robots. The prospect of childcare robots is particularly controversial ethically, and is associated with emotive arguments. The sampled material is restricted to relatively recent posts (the past three years) though the analysis identifies both continuous and changing themes across the past decade. The method extrapolates social representations theory towards examining the ways in which information about robotic products is provided for the general public. Implications for social acceptance of robot companions for the home and robot ethics are considered.

Keywords: acceptance of robots, childcare robots, ethics, social representations

Procedia PDF Downloads 246
6832 Anomaly Detection Based on System Log Data

Authors: M. Kamel, A. Hoayek, M. Batton-Hubert

Abstract:

With the increase of network virtualization and the disparity of vendors, the continuous monitoring and detection of anomalies cannot rely on static rules. An advanced analytical methodology is needed to discriminate between ordinary events and unusual anomalies. In this paper, we focus on log data (textual data), which is a crucial source of information for network performance. Then, we introduce an algorithm used as a pipeline to help with the pretreatment of such data, group it into patterns, and dynamically label each pattern as an anomaly or not. Such tools will provide users and experts with continuous real-time logs monitoring capability to detect anomalies and failures in the underlying system that can affect performance. An application of real-world data illustrates the algorithm.

Keywords: logs, anomaly detection, ML, scoring, NLP

Procedia PDF Downloads 89