Search results for: computer code
3513 Braille Code Matrix
Authors: Mohammed E. A. Brixi Nigassa, Nassima Labdelli, Ahmed Slami, Arnaud Pothier, Sofiane Soulimane
Abstract:
According to the world health organization (WHO), there are almost 285 million people with visual disability, 39 million of these people are blind. Nevertheless, there is a code for these people that make their life easier and allow them to access information more easily; this code is the Braille code. There are several commercial devices allowing braille reading, unfortunately, most of these devices are not ergonomic and too expensive. Moreover, we know that 90 % of blind people in the world live in low-incomes countries. Our contribution aim is to concept an original microactuator for Braille reading, as well as being ergonomic, inexpensive and lowest possible energy consumption. Nowadays, the piezoelectric device gives the better actuation for low actuation voltage. In this study, we focus on piezoelectric (PZT) material which can bring together all these conditions. Here, we propose to use one matrix composed by six actuators to form the 63 basic combinations of the Braille code that contain letters, numbers, and special characters in compliance with the standards of the braille code. In this work, we use a finite element model with Comsol Multiphysics software for designing and modeling this type of miniature actuator in order to integrate it into a test device. To define the geometry and the design of our actuator, we used physiological limits of perception of human being. Our results demonstrate in our study that piezoelectric actuator could bring a large deflection out-of-plain. Also, we show that microactuators can exhibit non uniform compression. This deformation depends on thin film thickness and the design of membrane arm. The actuator composed of four arms gives the higher deflexion and it always gives a domed deformation at the center of the deviceas in case of the Braille system. The maximal deflection can be estimated around ten micron per Volt (~ 10µm/V). We noticed that the deflection according to the voltage is a linear function, and this deflection not depends only on the voltage the voltage, but also depends on the thickness of the film used and the design of the anchoring arm. Then, we were able to simulate the behavior of the entire matrix and thus display different characters in Braille code. We used these simulations results to achieve our demonstrator. This demonstrator is composed of a layer of PDMS on which we put our piezoelectric material, and then added another layer of PDMS to isolate our actuator. In this contribution, we compare our results to optimize the final demonstrator.Keywords: Braille code, comsol software, microactuators, piezoelectric
Procedia PDF Downloads 3563512 Code Mixing and Code-Switching Patterns in Kannada-English Bilingual Children and Adults Who Stutter
Authors: Vasupradaa Manivannan, Santosh Maruthy
Abstract:
Background/Aims: Preliminary evidence suggests that code-switching and code-mixing may act as one of the voluntary coping behavior to avoid the stuttering characteristics in children and adults; however, less is known about the types and patterns of code-mixing (CM) and code-switching (CS). Further, it is not known how it is different between children to adults who stutter. This study aimed to identify and compare the CM and CS patterns between Kannada-English bilingual children and adults who stutter. Method: A standard group comparison was made between five children who stutter (CWS) in the age range of 9-13 years and five adults who stutter (AWS) in the age range of 20-25 years. The participants who are proficient in Kannada (first language- L1) and English (second language- L2) were considered for the study. There were two tasks given to both the groups, a) General conversation (GC) with 10 random questions, b) Narration task (NAR) (Story / General Topic, for example., A Memorable Life Event) in three different conditions {Mono Kannada (MK), Mono English (ME), and Bilingual (BIL) Condition}. The children and adults were assessed online (via Zoom session) with a high-quality internet connection. The audio and video samples of the full assessment session were auto-recorded and manually transcribed. The recorded samples were analyzed for the percentage of dysfluencies using SSI-4 and CM, and CS exhibited in each participant using Matrix Language Frame (MLF) model parameters. The obtained data were analyzed using the Statistical Package for the Social Sciences (SPSS) software package (Version 20.0). Results: The mean, median, and standard deviation values were obtained for the percentage of dysfluencies (%SS) and frequency of CM and CS in Kannada-English bilingual children and adults who stutter for various parameters obtained through the MLF model. The inferential results indicated that %SS significantly varied between population (AWS vs CWS), languages (L1 vs L2), and tasks (GC vs NAR) but not across free (BIL) and bound (MK, ME) conditions. It was also found that the frequency of CM and CS patterns varies between CWS and AWS. The AWS had a lesser %SS but greater use of CS patterns than CWS, which is due to their excessive coping skills. The language mixing patterns were more observed in L1 than L2, and it was significant in most of the MLF parameters. However, there was a significantly higher (P<0.05) %SS in L2 than L1. The CS and CS patterns were more in conditions 1 and 3 than 2, which may be due to the higher proficiency of L2 than L1. Conclusion: The findings highlight the importance of assessing the CM and CS behaviors, their patterns, and the frequency of CM and CS between CWS and AWS on MLF parameters in two different tasks across three conditions. The results help us to understand CM and CS strategies in bilingual persons who stutter.Keywords: bilinguals, code mixing, code switching, stuttering
Procedia PDF Downloads 783511 A Rapid Prototyping Tool for Suspended Biofilm Growth Media
Authors: Erifyli Tsagkari, Stephanie Connelly, Zhaowei Liu, Andrew McBride, William Sloan
Abstract:
Biofilms play an essential role in treating water in biofiltration systems. The biofilm morphology and function are inextricably linked to the hydrodynamics of flow through a filter, and yet engineers rarely explicitly engineer this interaction. We develop a system that links computer simulation and 3-D printing to optimize and rapidly prototype filter media to optimize biofilm function with the hypothesis that biofilm function is intimately linked to the flow passing through the filter. A computational model that numerically solves the incompressible time-dependent Navier Stokes equations coupled to a model for biofilm growth and function is developed. The model is imbedded in an optimization algorithm that allows the model domain to adapt until criteria on biofilm functioning are met. This is applied to optimize the shape of filter media in a simple flow channel to promote biofilm formation. The computer code links directly to a 3-D printer, and this allows us to prototype the design rapidly. Its validity is tested in flow visualization experiments and by microscopy. As proof of concept, the code was constrained to explore a small range of potential filter media, where the medium acts as an obstacle in the flow that sheds a von Karman vortex street that was found to enhance the deposition of bacteria on surfaces downstream. The flow visualization and microscopy in the 3-D printed realization of the flow channel validated the predictions of the model and hence its potential as a design tool. Overall, it is shown that the combination of our computational model and the 3-D printing can be effectively used as a design tool to prototype filter media to optimize biofilm formation.Keywords: biofilm, biofilter, computational model, von karman vortices, 3-D printing.
Procedia PDF Downloads 1433510 Investigation of Building Loads Effect on the Stability of Slope
Authors: Hadj Brahim Mounia, Belhamel Farid, Souici Messoud
Abstract:
In big cities, construction on sloping land (landslide) is becoming increasingly prevalent due to the unavailability of flat lands. This has created a major challenge for structural engineers with regard to structure design, due to the difficulties encountered during the implementation of projects, both for the structure and the soil. This paper analyses the effect of the number of floors of a building, founded on isolated footing on the stability of the slope using the computer code finite element PLAXIS 2D v. 8.2. The isolated footings of a building in this case were anchored in soil so that the levels of successive isolated footing realize a maximum slope of base of three for two heights, which connects the edges of the nearest footings, according to the Algerian building code DTR-BC 2.331: Shallow foundations. The results show that the embedment of the foundation into the soil reduces the value of the safety factor due to the change of the stress state of the soil by these foundations. The number of floors a building has also influences the safety factor. It has been noticed from this case of study that there is no risk of collapse of slopes for an inclination between 5° and 8°. In the case of slope inclination greater than 10° it has been noticed that the urbanization is prohibited.Keywords: isolated footings, multi-storeys building, PLAXIS 2D, slope
Procedia PDF Downloads 2543509 Comparison of High Speed Railway Bride Foundation Design
Authors: Hussein Yousif Aziz
Abstract:
This paper discussed the design and analysis of bridge foundation subjected to load of train with three codes, namely AASHTO code, British Standard BS Code 8004 (1986), and Chinese code (TB10002.5-2005).The study focused on the design and analysis of bridge’s foundation manually with the three codes and found which code is better for design and controls the problem of high settlement due to the applied loads. The results showed the Chinese codes are costly that the number of reinforcement bars in the pile cap and piles is more than those with AASHTO code and BS code with the same dimensions. Settlement of the bridge was calculated depending on the data collected from the project site. The vertical ultimate bearing capacity of single pile for three codes is also discussed. Other analyses by using the two-dimensional Plaxis program and other programs like SAP2000 14, PROKON many parameters are calculated. The maximum values of the vertical displacement are close to the calculated ones. The results indicate that the AASHTO code is economics and safer in the bearing capacity of single pile. The purpose of this project is to study out the pier on the basis of the design of the pile foundation. There is a 32m simply supported beam of box section on top of the structure. The pier of bridge is round-type. The main component of the design is to calculate pile foundation and the settlement. According to the related data, we choose 1.0m in diameter bored pile of 48m. The pile is laid out in the rectangular pile cap. The dimension of the cap is 12m 9 m. Because of the interaction factors of pile groups, the load-bearing capacity of simple pile must be checked, the punching resistance of pile cap, the shearing strength of pile cap, and the part in bending of pile cap, all of them are very important to the structure stability. Also, checking soft sub-bearing capacity is necessary under the pile foundation. This project provides a deeper analysis and comparison about pile foundation design schemes. Firstly, here are brief instructions of the construction situation about the Bridge. With the actual construction geological features and the upper load on the Bridge, this paper analyzes the bearing capacity and settlement of single pile. In the paper the Equivalent Pier Method is used to calculate and analyze settlements of the piles.Keywords: pile foundation, settlement, bearing capacity, civil engineering
Procedia PDF Downloads 4223508 Modeling of the Flow through an Earth Dam and Geotechnical Slope Analyzes
Authors: Ahmed Ferhati, Arezki Adjrad, Ratiba Mitiche-Kettab, Hakim Djafer Khodja
Abstract:
The porous media are omnipresent around us that they are natural as sand, clay, rocks, or manufactured like concretes, cement, and ceramics. The variety of porous environment indicates a wide material range which can be very different from each other. Their common point is to be made up of a solid matrix and a porous space. In our case of study, we made the modeling of the flows in porous environments through the massives as in the case of an earth dam. The computer code used (PLAXIS) offer the possibility of modeling of various structures, in particular, the works in lands because that it deals with the pore water pressure due to the underground flow and the calculation of the plastic deformations. To confirm results obtained by PLAXIS, GeoStudio SEEP/W code was used. This work treats modeling of flows and mechanical and hydraulic behavior of earth dam. A general framework which can fit the calculation of this kind of structures and the coupling of the soil consolidation and free surface flows was defined. In this study; we have confronted a real case modeling of an earth dam. It was shown, in particular, that it is possible to entirely lead the calculation of real dam and to get encouraging results from the hydraulic and mechanical point of view.Keywords: analyzes, dam, flow, modeling, PLAXIS, seep/w, slope
Procedia PDF Downloads 3103507 Protection of Stakeholders under the Transitional Commercial Code of Eritrea: Comparative Analysis with the 2018 Company Law of Peoples Republic of China
Authors: Hayle Makda Gebru
Abstract:
Companies are inevitable for society. They are the building blocks of every development in a country aimed at producing continuous goods and services for the people and, in turn, obliged to pay taxes, which enhances the economy of the nation. For the proper functioning of companies, their relationship with their stakeholders must be secure. The major stakeholders are suppliers, consumers, employees, creditors, etc. The law plays an important role in enhancing the relationship between these different stakeholders. If the law fails to keep track of the relationship, both the company and stakeholders remain unprotected. As a result, the potential benefits are prejudiced. This paper makes a comparative analysis of the types and formation of companies under the Transitional Commercial Code of Eritrea and the Company Law of the Peoples Republic of China. In particular, the paper addresses the legal lacuna under the TCrCE on handling the failure of shareholders to pay the promised capital. So, the methodology of the study is entirely analyzing the two countries' laws using practical cases. After analyzing the practical problems on the ground using real cases, this paper calls on Eritrea to update its outdated Commercial Code to give proper protection to the stakeholders.Keywords: companies, company law of the People's Republic of China, transitional commercial code of Eritrea, protection of stakeholders, failure to pay the promised capital
Procedia PDF Downloads 713506 Computer Fraud from the Perspective of Iran's Law and International Documents
Authors: Babak Pourghahramani
Abstract:
One of the modern crimes against property and ownership in the cyber-space is the computer fraud. Despite being modern, the aforementioned crime has its roots in the principles of religious jurisprudence. In some cases, this crime is compatible with the traditional regulations and that is when the computer is considered as a crime commitment device and also some computer frauds that take place in the context of electronic exchanges are considered as crime based on the E-commerce Law (approved in 2003) but the aforementioned regulations are flawed and until recent years there was no comprehensive law in this regard; yet after some years the Computer Crime Act was approved in 2009/26/5 and partly solved the problem of legal vacuum. The present study intends to investigate the computer fraud according to Iran's Computer Crime Act and by taking into consideration the international documents.Keywords: fraud, cyber fraud, computer fraud, classic fraud, computer crime
Procedia PDF Downloads 3323505 Frequency of Nosocomial Infections in a Tertiary Hospital in Isfahan, Iran
Authors: Zahra Tolou-Ghamari
Abstract:
Objective: Health care associated with multiresistant pathogens is rising globally. It is well known that nosocomial infections increase hospital stay, morbidity, mortality, and disability. Therefore, the aim of this study was to define the occurrence of nosocomial infections in a tertiary hospital in Isfahan/Iran. Materials and Methods: The data were extracted from the official database of hospital nosocomial infections records that included 9152 vertical rows. For each patient, the reported infections were coded by number as UTI-SUTI; Code 55, VAE-PVAP; Code 56, BSI-LCBI Code 19, SSI-DIP; Code 14, and so on. For continuous variables, mean ± standard deviation and for categorical variables, the frequency was used. Results: The study population was 5542 patients, comprised of males (n=3282) and females (n=2260). With a minimum of 15 and a maximum of 99, the mean age in 5313 patients was 58.5 ± 19.1 years old. The highest reported nosocomial infections (n= 77%) were associated with the ages 30-80 years old. Sites of nosocomial infections in 87% were as: VAE-PVAP; 27.3%, VAE-IVAC; 7.7, UTI-SUTI; 29.5%, BSI-LCBI; 12.9%, SSI-DIP; 9.5% and other individual infection (13%) with the main pathogens klebsiella pneumonia, acinetobacter baumannii and staphylococcus. Conclusions: For an efficient surveillance system, adopting pharmacotherapy used antibiotics in terms of monotherapy or polypharmacy control policy, in addition to advanced infection control programs at regional and national levels in Iran recommended.Keywords: infection, nosocomial, ventilator, blood stream, Isfahan, Iran
Procedia PDF Downloads 793504 Mobile Platform’s Attitude Determination Based on Smoothed GPS Code Data and Carrier-Phase Measurements
Authors: Mohamed Ramdani, Hassen Abdellaoui, Abdenour Boudrassen
Abstract:
Mobile platform’s attitude estimation approaches mainly based on combined positioning techniques and developed algorithms; which aim to reach a fast and accurate solution. In this work, we describe the design and the implementation of an attitude determination (AD) process, using only measurements from GPS sensors. The major issue is based on smoothed GPS code data using Hatch filter and raw carrier-phase measurements integrated into attitude algorithm based on vectors measurement using least squares (LSQ) estimation method. GPS dataset from a static experiment is used to investigate the effectiveness of the presented approach and consequently to check the accuracy of the attitude estimation algorithm. Attitude results from GPS multi-antenna over short baselines are introduced and analyzed. The 3D accuracy of estimated attitude parameters using smoothed measurements is over 0.27°.Keywords: attitude determination, GPS code data smoothing, hatch filter, carrier-phase measurements, least-squares attitude estimation
Procedia PDF Downloads 1553503 Effect of Modeling of Hydraulic Form Loss Coefficient to Break on Emergency Core Coolant Bypass
Authors: Young S. Bang, Dong H. Yoon, Seung H. Yoo
Abstract:
Emergency Core Coolant Bypass (ECC Bypass) has been regarded as an important phenomenon to peak cladding temperature of large-break loss-of-coolant-accidents (LBLOCA) in nuclear power plants (NPP). A modeling scheme to address the ECC Bypass phenomena and the calculation of LBLOCA using that scheme are discussed in the present paper. A hydraulic form loss coefficient (HFLC) from the reactor vessel downcomer to the broken cold leg is predicted by the computational fluid dynamics (CFD) code with a variation of the void fraction incoming from the downcomer. The maximum, mean, and minimum values of FLC are derived from the CFD results and are incorporated into the LBLOCA calculation using a system thermal-hydraulic code, MARS-KS. As a relevant parameter addressing the ECC Bypass phenomena, the FLC to the break and its range are proposed.Keywords: CFD analysis, ECC bypass, hydraulic form loss coefficient, system thermal-hydraulic code
Procedia PDF Downloads 2313502 Establish a Company in Turkey for Foreigners
Authors: Mucahit Unal, Ibrahim Arslan
Abstract:
The New Turkish Commercial Code (TCC) No. 6102 was published in the Official Gazette on February 14, 2011. As stated in the New Turkish Commercial Code No. 6102 and Law No. 6103 on Validity and Application of the Turkish Commercial Code, TCC came into effect on July 1, 2012. The basic purpose of the TCC is to form corporate governance coherent with the international standards; to provide transparency in company management; to adjust the Turkish Commercial Code rules with European Union legislations and to simplify establishing a company for foreigner investors to move investments to Turkish market. In this context according to TCC, joint stock companies and limited liability companies can establish with only one single shareholder; the one single shareholder can be foreigner; all board of director members can be foreigner, also all shareholders and board of director members can be non-resident foreigners. Additionally, TCC does not require physical participation to the general shareholders and board members meetings. TCC allows that the general shareholders and board members meetings can hold in an electronic form and resolution of these meetings may also be approved via electronic signatures. Through this amendment, foreign investors no longer have to deal with red tapes. This amendment also means the TCC prevents foreign companies from incurring unnecessary travel expenses. In accordance with all this amendments about TCC, to invest in Turkish market is easy, simple and transparent for foreigner investors and also investors can establish a company in Turkey, irrespective of nationality or place of residence. This article aims to analyze ‘Establish a Company in Turkey for Foreigners’ and inform investors about investing (especially establishing a company) in the Turkish market.Keywords: establish a company, foreigner investors, invest in Turkish market, Turkish commercial code
Procedia PDF Downloads 2643501 Investigating the Use of English Arabic Codeswitching in EFL classroom Oral Discourse Case study: Middle school pupils of Ain Fekroun, Wilaya of Oum El Bouaghi Algeria
Authors: Fadila Hadjeris
Abstract:
The study aims at investigating the functions of English-Arabic code switching in English as a foreign language classroom oral discourse and the extent to which they can contribute to the flow of classroom interaction. It also seeks to understand the views, beliefs, and perceptions of teachers and learners towards this practice. We hypothesized that code switching is a communicative strategy which facilitates classroom interaction. Due to this fact, both teachers and learners support its use. The study draws on a key body of literature in bilingualism, second language acquisition, and classroom discourse in an attempt to provide a framework for considering the research questions. It employs a combination of qualitative and quantitative research methods which include classroom observations and questionnaires. The analysis of the recordings shows that teachers’ code switching to Arabic is not only used for academic and classroom management reasons. Rather, the data display instances in which code switching is used for social reasons. The analysis of the questionnaires indicates that teachers and pupils have different attitudes towards this phenomenon. Teachers reported their deliberate switching during EFL teaching, yet the majority was against this practice. According to them, the use of the mother has detrimental effects on the acquisition and the practice of the target language. In contrast, pupils showed their preference to their teachers’ code switching because it enhances and facilitates their understanding. These findings support the fact that the shift to pupils’ mother tongue is a strategy which aids and facilitates the teaching and the learning of the target language. This, in turn, necessitates recommendations which are suggested to teachers and course designers.Keywords: bilingualism, codeswitching, classroom interaction, classroom discourse, EFL learning/ teaching, SLA
Procedia PDF Downloads 4813500 Bit Error Rate (BER) Performance of Coherent Homodyne BPSK-OCDMA Network for Multimedia Applications
Authors: Morsy Ahmed Morsy Ismail
Abstract:
In this paper, the structure of a coherent homodyne receiver for the Binary Phase Shift Keying (BPSK) Optical Code Division Multiple Access (OCDMA) network is introduced based on the Multi-Length Weighted Modified Prime Code (ML-WMPC) for multimedia applications. The Bit Error Rate (BER) of this homodyne detection is evaluated as a function of the number of active users and the signal to noise ratio for different code lengths according to the multimedia application such as audio, voice, and video. Besides, the Mach-Zehnder interferometer is used as an external phase modulator in homodyne detection. Furthermore, the Multiple Access Interference (MAI) and the receiver noise in a shot-noise limited regime are taken into consideration in the BER calculations.Keywords: OCDMA networks, bit error rate, multiple access interference, binary phase-shift keying, multimedia
Procedia PDF Downloads 1763499 Performance Comparison of Non-Binary RA and QC-LDPC Codes
Abstract:
Repeat–Accumulate (RA) codes are subclass of LDPC codes with fast encoder structures. In this paper, we consider a nonbinary extension of binary LDPC codes over GF(q) and construct a non-binary RA code and a non-binary QC-LDPC code over GF(2^4), we construct non-binary RA codes with linear encoding method and non-binary QC-LDPC codes with algebraic constructions method. And the BER performance of RA and QC-LDPC codes over GF(q) are compared with BP decoding and by simulation over the Additive White Gaussian Noise (AWGN) channels.Keywords: non-binary RA codes, QC-LDPC codes, performance comparison, BP algorithm
Procedia PDF Downloads 3773498 Tertiary Level Teachers' Beliefs about Codeswitching
Authors: Hoa Pham
Abstract:
Code switching, which can be described as the use of students’ first language in second language classrooms, has long been a controversial topic in the area of language teaching and second language acquisition. While this has been widely investigated across different contexts, little empirical research has been undertaken in Vietnam. The findings of this study contribute to our understanding of bilingual discourse and code switching practices in content and language integrated classrooms, which has significant implications for language teaching and learning in general and in particular for language pedagogy at tertiary level in Vietnam. This study examines the accounts the teachers articulated for their code switching practices in content-based Business English in Vietnam. Data were collected from five teachers through the use of stimulated recall interviews facilitated by the video data to garner the teachers' cognitive reflection, and allowed them to vocalise the motivations behind their code switching behaviour in particular contexts. The literature has recommended that when participants are provided with a large amount of stimuli or cues, they will experience an original situation again in their imagination with great accuracy. This technique can also provide a valuable "insider" perspective on the phenomenon under investigation which complements the researcher’s "outsider" observation. This can create a relaxed atmosphere during the interview process, which in turn promotes the collection of rich and diverse data. Also, participants can be empowered by this technique as they can raise their own concerns and discuss instances which they find important or interesting. The data generated through this study were analysed using a constant comparative approach. The study found that the teachers indicated their support for the use of code switching in their pedagogical practices. Particularly, as a pedagogical resource, the teachers saw code switching to the L1 playing a key role in facilitating the students' comprehension of both content knowledge and the target language. They believed the use of the L1 accommodates the students' current language competence and content knowledge. They also expressed positive opinions about the role that code switching plays in stimulating students' schematic language and content knowledge, encouraging retention and interest in learning and promoting a positive affective environment in the classroom. The teachers perceived that their use of code switching to the L1 helps them meet the students' language needs and prepares them for their study in subsequent courses and addresses functional needs so that students can cope with English language use outside the classroom. Several factors shaped the teachers' perceptions of their code switching practices, including their accumulated teaching experience, their previous experience as language learners, their theoretical understanding of language teaching and learning, and their knowledge of the teaching context. Code switching was a typical phenomenon in the observed classes and was supported by the teachers in certain contexts. This study reinforces the call in the literature to recognise this practice as a useful instructional resource.Keywords: codeswitching, language teaching, teacher beliefs, tertiary level
Procedia PDF Downloads 4533497 Metaphorical Perceptions of Middle School Students regarding Computer Games
Authors: Ismail Celik, Ismail Sahin, Fetah Eren
Abstract:
The computer, among the most important inventions of the twentieth century, has become an increasingly important component in our everyday lives. Computer games also have become increasingly popular among people day-by-day, owing to their features based on realistic virtual environments, audio and visual features, and the roles they offer players. In the present study, the metaphors students have for computer games are investigated, as well as an effort to fill the gap in the literature. Students were asked to complete the sentence—‘Computer game is like/similar to….because….’— to determine the middle school students’ metaphorical images of the concept for ‘computer game’. The metaphors created by the students were grouped in six categories, based on the source of the metaphor. These categories were ordered as ‘computer game as a means of entertainment’, ‘computer game as a beneficial means’, ‘computer game as a basic need’, ‘computer game as a source of evil’, ‘computer game as a means of withdrawal’, and ‘computer game as a source of addiction’, according to the number of metaphors they included.Keywords: computer game, metaphor, middle school students, virtual environments
Procedia PDF Downloads 5353496 Use of Computer and Machine Learning in Facial Recognition
Authors: Neha Singh, Ananya Arora
Abstract:
Facial expression measurement plays a crucial role in the identification of emotion. Facial expression plays a key role in psychophysiology, neural bases, and emotional disorder, to name a few. The Facial Action Coding System (FACS) has proven to be the most efficient and widely used of the various systems used to describe facial expressions. Coders can manually code facial expressions with FACS and, by viewing video-recorded facial behaviour at a specified frame rate and slow motion, can decompose into action units (AUs). Action units are the most minor visually discriminable facial movements. FACS explicitly differentiates between facial actions and inferences about what the actions mean. Action units are the fundamental unit of FACS methodology. It is regarded as the standard measure for facial behaviour and finds its application in various fields of study beyond emotion science. These include facial neuromuscular disorders, neuroscience, computer vision, computer graphics and animation, and face encoding for digital processing. This paper discusses the conceptual basis for FACS, a numerical listing of discrete facial movements identified by the system, the system's psychometric evaluation, and the software's recommended training requirements.Keywords: facial action, action units, coding, machine learning
Procedia PDF Downloads 1063495 GPU Based High Speed Error Protection for Watermarked Medical Image Transmission
Authors: Md Shohidul Islam, Jongmyon Kim, Ui-pil Chong
Abstract:
Medical image is an integral part of e-health care and e-diagnosis system. Medical image watermarking is widely used to protect patients’ information from malicious alteration and manipulation. The watermarked medical images are transmitted over the internet among patients, primary and referred physicians. The images are highly prone to corruption in the wireless transmission medium due to various noises, deflection, and refractions. Distortion in the received images leads to faulty watermark detection and inappropriate disease diagnosis. To address the issue, this paper utilizes error correction code (ECC) with (8, 4) Hamming code in an existing watermarking system. In addition, we implement the high complex ECC on a graphics processing units (GPU) to accelerate and support real-time requirement. Experimental results show that GPU achieves considerable speedup over the sequential CPU implementation, while maintaining 100% ECC efficiency.Keywords: medical image watermarking, e-health system, error correction, Hamming code, GPU
Procedia PDF Downloads 2913494 A Qualitative Evidence of the Markedness of Code Switching during Commercial Bank Service Encounters in Ìbàdàn Metropolis
Authors: A. Robbin
Abstract:
In a multilingual setting like Nigeria, the success of service encounters is enhanced by the use of a language that ensures the linguistic and persuasive demands of the interlocutors. This study examined motivations for code switching as a negotiation strategy in bank-hall desk service encounters in Ìbàdàn metropolis using Myers-Scotton’s exploration on markedness in language use. The data consisted of transcribed audio recording of bank-hall service encounters, and direct observation of bank interactions in two purposively sampled commercial banks in Ìbàdàn metropolis. The data was subjected to descriptive linguistic analysis using Myers Scotton’s Markedness Model. Findings reveal that code switching is frequently employed during different stages of service encounter: greeting, transaction and closing to fulfil relational, bargaining and referential functions. Bank staff and customers code switch to make unmarked, marked and explanatory choices. A strategy used to identify with customer’s cultural affiliation, close status gap, and appeal to begrudged customer; or as an explanatory choice with non-literate customers for ease of communication. Bankers select English to maintain customers’ perceptions of prestige which is retained or diverged from depending on their linguistic preference or ability. Yoruba is seen as an efficient negotiation strategy with both bankers and their customers, making choices within conversation to achieve desired conversational and functional aims.Keywords: banking, bilingualism, code-switching, markedness, service encounter
Procedia PDF Downloads 2073493 Myanmar Character Recognition Using Eight Direction Chain Code Frequency Features
Authors: Kyi Pyar Zaw, Zin Mar Kyu
Abstract:
Character recognition is the process of converting a text image file into editable and searchable text file. Feature Extraction is the heart of any character recognition system. The character recognition rate may be low or high depending on the extracted features. In the proposed paper, 25 features for one character are used in character recognition. Basically, there are three steps of character recognition such as character segmentation, feature extraction and classification. In segmentation step, horizontal cropping method is used for line segmentation and vertical cropping method is used for character segmentation. In the Feature extraction step, features are extracted in two ways. The first way is that the 8 features are extracted from the entire input character using eight direction chain code frequency extraction. The second way is that the input character is divided into 16 blocks. For each block, although 8 feature values are obtained through eight-direction chain code frequency extraction method, we define the sum of these 8 feature values as a feature for one block. Therefore, 16 features are extracted from that 16 blocks in the second way. We use the number of holes feature to cluster the similar characters. We can recognize the almost Myanmar common characters with various font sizes by using these features. All these 25 features are used in both training part and testing part. In the classification step, the characters are classified by matching the all features of input character with already trained features of characters.Keywords: chain code frequency, character recognition, feature extraction, features matching, segmentation
Procedia PDF Downloads 3213492 The Connection between the Schwartz Theory of Basic Values and Ethical Principles in Clinical Psychology
Authors: Matej Stritesky
Abstract:
The research deals with the connection between the Schwartz Theory of Basic Values and the ethical principles in psychology, on which the meta-code of ethics the European Federation of Psychological Associations is based. The research focuses on ethically problematic situations in clinical psychology in the Czech Republic. Based on the analysis of papers that identified ethically problematic situations faced by clinical psychologists, a questionnaire of ethically problematic situations in clinical psychology (EPSCP) was created for the purposes of the research. The questionnaire was created to represent situations that correspond to the 4 principles on which the meta-code of ethics the European Federation of Psychological Associations is based. The questionnaire EPSCP consists of descriptions of 32 situations that respondents evaluate on a scale from 1 (psychologist's behaviour is ethically perfectly fine) to 10 (psychologist's behaviour is ethically completely unacceptable). The EPSCP questionnaire, together with Schwartz's PVQ questionnaire, will be presented to 60 psychology students. The relationship between principles in clinical psychology and the values on Schwartz´s value continuum will be described using multidimensional scaling. A positive correlation is assumed between the higher-order value of openness to change and problematic ethical situations related to the principle of integrity; a positive correlation between the value of the higher order of self-transcendence and the principle of respect and responsibility; a positive correlation between the value of the higher order of conservation and the principle of competence; and negative correlation between the value of the higher order of ego strengthening and sensitivity to ethically problematic situations. The research also includes an experimental part. The first half of the students are presented with the code of ethics of the Czech Association of Clinical Psychologists before completing the questionnaires, and to the second half of the students is the code of ethics presented after completing the questionnaires. In addition to reading the code of ethics, students describe the three rules of the code of ethics that they consider most important and state why they chose these rules. The output of the experimental part will be to determine whether the presentation of the code of ethics leads to greater sensitivity to ethically problematic situations.Keywords: clinical psychology, ethically problematic situations in clinical psychology, ethical principles in psychology, Schwartz theory of basic values
Procedia PDF Downloads 1133491 Detecting Heartbeat Architectural Tactic in Source Code Using Program Analysis
Authors: Ananta Kumar Das, Sujit Kumar Chakrabarti
Abstract:
Architectural tactics such as heartbeat, ping-echo, encapsulate, encrypt data are techniques that are used to achieve quality attributes of a system. Detecting architectural tactics has several benefits: it can aid system comprehension (e.g., legacy systems) and in the estimation of quality attributes such as safety, security, maintainability, etc. Architectural tactics are typically spread over the source code and are implicit. For large codebases, manual detection is often not feasible. Therefore, there is a need for automated methods of detection of architectural tactics. This paper presents a formalization of the heartbeat architectural tactic and a program analytic approach to detect this tactic in source code. The experiment of the proposed method is done on a set of Java applications. The outcome of the experiment strongly suggests that the method compares well with a manual approach in terms of its sensitivity and specificity, and far supersedes a manual exercise in terms of its scalability.Keywords: software architecture, architectural tactics, detecting architectural tactics, program analysis, AST, alias analysis
Procedia PDF Downloads 1613490 Prompt Design for Code Generation in Data Analysis Using Large Language Models
Authors: Lu Song Ma Li Zhi
Abstract:
With the rapid advancement of artificial intelligence technology, large language models (LLMs) have become a milestone in the field of natural language processing, demonstrating remarkable capabilities in semantic understanding, intelligent question answering, and text generation. These models are gradually penetrating various industries, particularly showcasing significant application potential in the data analysis domain. However, retraining or fine-tuning these models requires substantial computational resources and ample downstream task datasets, which poses a significant challenge for many enterprises and research institutions. Without modifying the internal parameters of the large models, prompt engineering techniques can rapidly adapt these models to new domains. This paper proposes a prompt design strategy aimed at leveraging the capabilities of large language models to automate the generation of data analysis code. By carefully designing prompts, data analysis requirements can be described in natural language, which the large language model can then understand and convert into executable data analysis code, thereby greatly enhancing the efficiency and convenience of data analysis. This strategy not only lowers the threshold for using large models but also significantly improves the accuracy and efficiency of data analysis. Our approach includes requirements for the precision of natural language descriptions, coverage of diverse data analysis needs, and mechanisms for immediate feedback and adjustment. Experimental results show that with this prompt design strategy, large language models perform exceptionally well in multiple data analysis tasks, generating high-quality code and significantly shortening the data analysis cycle. This method provides an efficient and convenient tool for the data analysis field and demonstrates the enormous potential of large language models in practical applications.Keywords: large language models, prompt design, data analysis, code generation
Procedia PDF Downloads 433489 Simulation of Reflectometry in Alborz Tokamak
Authors: S. Kohestani, R. Amrollahi, P. Daryabor
Abstract:
Microwave diagnostics such as reflectometry are receiving growing attention in magnetic confinement fusionresearch. In order to obtain the better understanding of plasma confinement physics, more detailed measurements on density profile and its fluctuations might be required. A 2D full-wave simulation of ordinary mode propagation has been written in an effort to model effects seen in reflectometry experiment. The code uses the finite-difference-time-domain method with a perfectly-matched-layer absorption boundary to solve Maxwell’s equations.The code has been used to simulate the reflectometer measurement in Alborz Tokamak.Keywords: reflectometry, simulation, ordinary mode, tokamak
Procedia PDF Downloads 4203488 Evaluation of a Data Fusion Algorithm for Detecting and Locating a Radioactive Source through Monte Carlo N-Particle Code Simulation and Experimental Measurement
Authors: Hadi Ardiny, Amir Mohammad Beigzadeh
Abstract:
Through the utilization of a combination of various sensors and data fusion methods, the detection of potential nuclear threats can be significantly enhanced by extracting more information from different data. In this research, an experimental and modeling approach was employed to track a radioactive source by combining a surveillance camera and a radiation detector (NaI). To run this experiment, three mobile robots were utilized, with one of them equipped with a radioactive source. An algorithm was developed in identifying the contaminated robot through correlation between camera images and camera data. The computer vision method extracts the movements of all robots in the XY plane coordinate system, and the detector system records the gamma-ray count. The position of the robots and the corresponding count of the moving source were modeled using the MCNPX simulation code while considering the experimental geometry. The results demonstrated a high level of accuracy in finding and locating the target in both the simulation model and experimental measurement. The modeling techniques prove to be valuable in designing different scenarios and intelligent systems before initiating any experiments.Keywords: nuclear threats, radiation detector, MCNPX simulation, modeling techniques, intelligent systems
Procedia PDF Downloads 1263487 Redefining Infrastructure as Code Orchestration Using AI
Authors: Georges Bou Ghantous
Abstract:
This research delves into the transformative impact of Artificial Intelligence (AI) on Infrastructure as Code (IaaC) practices, specifically focusing on the redefinition of infrastructure orchestration. By harnessing AI technologies such as machine learning algorithms and predictive analytics, organizations can achieve unprecedented levels of efficiency and optimization in managing their infrastructure resources. AI-driven IaaC introduces proactive decision-making through predictive insights, enabling organizations to anticipate and address potential issues before they arise. Dynamic resource scaling, facilitated by AI, ensures that infrastructure resources can seamlessly adapt to fluctuating workloads and changing business requirements. Through case studies and best practices, this paper sheds light on the tangible benefits and challenges associated with AI-driven IaaC transformation, providing valuable insights for organizations navigating the evolving landscape of digital infrastructure management.Keywords: artificial intelligence, infrastructure as code, efficiency optimization, predictive insights, dynamic resource scaling, proactive decision-making
Procedia PDF Downloads 363486 Feature Engineering Based Detection of Buffer Overflow Vulnerability in Source Code Using Deep Neural Networks
Authors: Mst Shapna Akter, Hossain Shahriar
Abstract:
One of the most important challenges in the field of software code audit is the presence of vulnerabilities in software source code. Every year, more and more software flaws are found, either internally in proprietary code or revealed publicly. These flaws are highly likely exploited and lead to system compromise, data leakage, or denial of service. C and C++ open-source code are now available in order to create a largescale, machine-learning system for function-level vulnerability identification. We assembled a sizable dataset of millions of opensource functions that point to potential exploits. We developed an efficient and scalable vulnerability detection method based on deep neural network models that learn features extracted from the source codes. The source code is first converted into a minimal intermediate representation to remove the pointless components and shorten the dependency. Moreover, we keep the semantic and syntactic information using state-of-the-art word embedding algorithms such as glove and fastText. The embedded vectors are subsequently fed into deep learning networks such as LSTM, BilSTM, LSTM-Autoencoder, word2vec, BERT, and GPT-2 to classify the possible vulnerabilities. Furthermore, we proposed a neural network model which can overcome issues associated with traditional neural networks. Evaluation metrics such as f1 score, precision, recall, accuracy, and total execution time have been used to measure the performance. We made a comparative analysis between results derived from features containing a minimal text representation and semantic and syntactic information. We found that all of the deep learning models provide comparatively higher accuracy when we use semantic and syntactic information as the features but require higher execution time as the word embedding the algorithm puts on a bit of complexity to the overall system.Keywords: cyber security, vulnerability detection, neural networks, feature extraction
Procedia PDF Downloads 913485 Regional Review of Outcome of Cervical Smears Reported with Cytological Features of Non Cervical Glandular Neoplasia
Authors: Uma Krishnamoorthy, Vivienne Beavers, Janet Marshall
Abstract:
Introduction: Cervical cytology showing features raising the suspicion of non cervical glandular neoplasia are reported as code 0 under the United Kingdom National Health Service Cervical screening programme ( NHSCSP). As the suspicion is regarding non cervical neoplasia, smear is reported as normal and patient informed that cervical screening result is normal. GP receives copy of results where it states further referral is indicated in small font within text of report. Background: There were several incidents of delayed diagnosis of endometrial cancer in Lancashire which prompted this Northwest Regional review to enable an understanding of underlying pathology outcome of code zero smears to raise awareness and also to review whether further action on wording of smear results was indicated to prevent such delay. Methodology: All Smears reported at the Manchester cytology centre who process cytology for Lancashire population from March 2013 to March 2014 were reviewed and histological diagnosis outcome of women in whom smear was reported as code zero was reviewed retrospectively . Results: Total smears reported by the cytology centre during this period was approximately 109400. Reports issued with result code 0 among this during this time period was 49.Results revealed that among three fourth (37) of women with code zero smear (N=49), evidence of underlying pathology of non cervical origin was confirmed. Of this, 73 % (36) were due to endometrial pathology with 49 % (24) endometrial carcinoma, 12 % (6)polyp, 4 % atypical endometrial hyperplasia (2), 6 % endometrial hyperplasia without atypia (3), and 2 % adenomyosis (1 case) and 2 % ( 1 case) due to ovarian adenocarcinoma. Conclusion: This review demonstrated that more than half (51 %) of women with a code 0 smear report were diagnosed with underlying carcinoma and 75 % had a confirmed underlying pathology contributory to code 0 smear findings. Recommendations and Action Plan: A local rapid access referral and management pathway for this group of women was implemented as a result of this in our unit. The findings and Pathway were shared with other regional units served by the cytology centre through the Pan Lancashire cervical screening board and through the Cytology centre. Locally, the smear report wording was updated to include a rubber stamp/ print in "Red Bold letters" stating that " URGENT REFERRAL TO GYNAECOLOGY IS INDICATED". Findings were also shared through the Pan Lancashire board with National cervical screening programme board, and revisions to wording of code zero smear reports to highlight the need for Urgent referral has now been agreed at National level to be implemented.Keywords: code zero smears, endometrial cancer, non cervical glandular neoplasia, ovarian cancer
Procedia PDF Downloads 2993484 Steady State Modeling and Simulation of an Industrial Steam Boiler
Authors: Amina Lyria Deghal Cheridi, Abla Chaker, Ahcene Loubar
Abstract:
Relap5 system code is one among powerful tools, which is used in the area of design and safety evaluation. This work aims to simulate the behavior of a radiant steam boiler at the steady-state conditions using Relap5 code system. To perform this study, a detailed Relap5 model is built including all the parts of the steam boiler. The control and regulation systems are also considered. To reproduce the most important parameters and phenomena with an acceptable accuracy and fidelity, a strong qualification work is undertaken concerning the facility nodalization. It consists of making a comparison between the code results and the plant available data in steady-state operation mode. Therefore, the model qualification results at the steady-state are in good agreement with the steam boiler experimental data. The steam boiler Relap5 model has proved satisfactory; and the model was capable of predicting the main thermal-hydraulic steady-state conditions of the steam boiler.Keywords: industrial steam boiler, model qualification, natural circulation, relap5/mod3.2, steady state simulation
Procedia PDF Downloads 273