Search results for: pen and paper or computer
25732 Implementation of an IoT Sensor Data Collection and Analysis Library
Authors: Jihyun Song, Kyeongjoo Kim, Minsoo Lee
Abstract:
Due to the development of information technology and wireless Internet technology, various data are being generated in various fields. These data are advantageous in that they provide real-time information to the users themselves. However, when the data are accumulated and analyzed, more various information can be extracted. In addition, development and dissemination of boards such as Arduino and Raspberry Pie have made it possible to easily test various sensors, and it is possible to collect sensor data directly by using database application tools such as MySQL. These directly collected data can be used for various research and can be useful as data for data mining. However, there are many difficulties in using the board to collect data, and there are many difficulties in using it when the user is not a computer programmer, or when using it for the first time. Even if data are collected, lack of expert knowledge or experience may cause difficulties in data analysis and visualization. In this paper, we aim to construct a library for sensor data collection and analysis to overcome these problems.Keywords: clustering, data mining, DBSCAN, k-means, k-medoids, sensor data
Procedia PDF Downloads 37825731 Features of Normative and Pathological Realizations of Sibilant Sounds for Computer-Aided Pronunciation Evaluation in Children
Authors: Zuzanna Miodonska, Michal Krecichwost, Pawel Badura
Abstract:
Sigmatism (lisping) is a speech disorder in which sibilant consonants are mispronounced. The diagnosis of this phenomenon is usually based on the auditory assessment. However, the progress in speech analysis techniques creates a possibility of developing computer-aided sigmatism diagnosis tools. The aim of the study is to statistically verify whether specific acoustic features of sibilant sounds may be related to pronunciation correctness. Such knowledge can be of great importance while implementing classifiers and designing novel tools for automatic sibilants pronunciation evaluation. The study covers analysis of various speech signal measures, including features proposed in the literature for the description of normative sibilants realization. Amplitudes and frequencies of three fricative formants (FF) are extracted based on local spectral maxima of the friction noise. Skewness, kurtosis, four normalized spectral moments (SM) and 13 mel-frequency cepstral coefficients (MFCC) with their 1st and 2nd derivatives (13 Delta and 13 Delta-Delta MFCC) are included in the analysis as well. The resulting feature vector contains 51 measures. The experiments are performed on the speech corpus containing words with selected sibilant sounds (/ʃ, ʒ/) pronounced by 60 preschool children with proper pronunciation or with natural pathologies. In total, 224 /ʃ/ segments and 191 /ʒ/ segments are employed in the study. The Mann-Whitney U test is employed for the analysis of stigmatism and normative pronunciation. Statistically, significant differences are obtained in most of the proposed features in children divided into these two groups at p < 0.05. All spectral moments and fricative formants appear to be distinctive between pathology and proper pronunciation. These metrics describe the friction noise characteristic for sibilants, which makes them particularly promising for the use in sibilants evaluation tools. Correspondences found between phoneme feature values and an expert evaluation of the pronunciation correctness encourage to involve speech analysis tools in diagnosis and therapy of sigmatism. Proposed feature extraction methods could be used in a computer-assisted stigmatism diagnosis or therapy systems.Keywords: computer-aided pronunciation evaluation, sigmatism diagnosis, speech signal analysis, statistical verification
Procedia PDF Downloads 30125730 A Case Study on the Impact of Technology Readiness in a Department of Clinical Nurses
Authors: Julie Delany
Abstract:
To thrive in today’s digital climate, it is vital that organisations adopt new technology and prepare for rising digital trends. This proves more difficult in government where, traditionally, people lack change readiness. While individuals may have a desire to work smarter, this does not necessarily mean embracing technology. This paper discusses the rollout of an application into a small department of highly experienced nurses. The goal was to both streamline the department's workflow and provide a platform for gathering essential business metrics. The biggest challenges were adoption and motivating the nurses to change their routines and learn new computer skills. Two-thirds struggled with the change, and as a result, some jeopardised the validity of the business metrics. In conclusion, there are lessons learned and recommendations for similar projects.Keywords: change ready, information technology, end-user, iterative method, rollout plan, data analytics
Procedia PDF Downloads 14525729 Biogas Control: Methane Production Monitoring Using Arduino
Authors: W. Ait Ahmed, M. Aggour, M. Naciri
Abstract:
Extracting energy from biomass is an important alternative to produce different types of energy (heat, electricity, or both) assuring low pollution and better efficiency. It is a new yet reliable approach to reduce green gas emission by extracting methane from industry effluents and use it to power machinery. We focused in our project on using paper and mill effluents, treated in a UASB reactor. The methane produced is used in the factory’s power supply. The aim of this work is to develop an electronic system using Arduino platform connected to a gas sensor, to measure and display the curve of daily methane production on processing. The sensor will send the gas values in ppm to the Arduino board so that the later sends the RS232 hardware protocol. The code developed with processing will transform the values into a curve and display it on the computer screen.Keywords: biogas, Arduino, processing, code, methane, gas sensor, program
Procedia PDF Downloads 32125728 3D Object Detection for Autonomous Driving: A Comprehensive Review
Authors: Ahmed Soliman Nagiub, Mahmoud Fayez, Heba Khaled, Said Ghoniemy
Abstract:
Accurate perception is a critical component in enabling autonomous vehicles to understand their driving environment. The acquisition of 3D information about objects, including their location and pose, is essential for achieving this understanding. This survey paper presents a comprehensive review of 3D object detection techniques specifically tailored for autonomous vehicles. The survey begins with an introduction to 3D object detection, elucidating the significance of the third dimension in perceiving the driving environment. It explores the types of sensors utilized in this context and the corresponding data extracted from these sensors. Additionally, the survey investigates the different types of datasets employed, including their formats, sizes, and provides a comparative analysis. Furthermore, the paper categorizes and thoroughly examines the perception methods employed for 3D object detection based on the diverse range of sensors utilized. Each method is evaluated based on its effectiveness in accurately detecting objects in a three-dimensional space. Additionally, the evaluation metrics used to assess the performance of these methods are discussed. By offering a comprehensive overview of 3D object detection techniques for autonomous vehicles, this survey aims to advance the field of perception systems. It serves as a valuable resource for researchers and practitioners, providing insights into the techniques, sensors, and evaluation metrics employed in 3D object detection for autonomous vehicles.Keywords: computer vision, 3D object detection, autonomous vehicles, deep learning
Procedia PDF Downloads 6225727 Can 3D Virtual Prototyping Conquers the Apparel Industry?
Authors: Evridiki Papachristou, Nikolaos Bilalis
Abstract:
Imagine an apparel industry where fashion design does not begin with a paper-and-pen drawing which is then translated into pattern and later to a 3D model where the designer tries out different fabrics, colours and contrasts. Instead, imagine a fashion designer in the future who produces that initial fashion drawing in a three-dimensional space and won’t leave that environment until the product is done, communicating his/her ideas with the entire development team in true to life 3D. Three-dimensional (3D) technology - while well established in many other industrial sectors like automotive, aerospace, architecture and industrial design, has only just started to open up a whole range of new opportunities for apparel designers. The paper will discuss the process of 3D simulation technology enhanced by high quality visualization of data and its capability to ensure a massive competitiveness in the market. Secondly, it will underline the most frequent problems & challenges that occur in the process chain when various partners in the production of textiles and apparel are working together. Finally, it will offer a perspective of how the Virtual Prototyping Technology will make the global textile and apparel industry change to a level where designs will be visualized on a computer and various scenarios modeled without even having to produce a physical prototype. This state-of-the-art 3D technology has been described as transformative and“disruptive”comparing to the process of the way apparel companies develop their fashion products today. It provides the benefit of virtual sampling not only for quick testing of design ideas, but also reducing process steps and having more visibility.A so called“digital asset” that can be used for other purposes such as merchandising or marketing.Keywords: 3D visualization, apparel, virtual prototyping, prototyping technology
Procedia PDF Downloads 59025726 Architectural Wind Data Maps Using an Array of Wireless Connected Anemometers
Authors: D. Serero, L. Couton, J. D. Parisse, R. Leroy
Abstract:
In urban planning, an increasing number of cities require wind analysis to verify comfort of public spaces and around buildings. These studies are made using computer fluid dynamic simulation (CFD). However, this technique is often based on wind information taken from meteorological stations located at several kilometers of the spot of analysis. The approximated input data on project surroundings produces unprecise results for this type of analysis. They can only be used to get general behavior of wind in a zone but not to evaluate precise wind speed. This paper presents another approach to this problem, based on collecting wind data and generating an urban wind cartography using connected ultrasound anemometers. They are wireless devices that send immediate data on wind to a remote server. Assembled in array, these devices generate geo-localized data on wind such as speed, temperature, pressure and allow us to compare wind behavior on a specific site or building. These Netatmo-type anemometers communicate by wifi with central equipment, which shares data acquired by a wide variety of devices such as wind speed, indoor and outdoor temperature, rainfall, and sunshine. Beside its precision, this method extracts geo-localized data on any type of site that can be feedback looped in the architectural design of a building or a public place. Furthermore, this method allows a precise calibration of a virtual wind tunnel using numerical aeraulic simulations (like STAR CCM + software) and then to develop the complete volumetric model of wind behavior over a roof area or an entire city block. The paper showcases connected ultrasonic anemometers, which were implanted for an 18 months survey on four study sites in the Grand Paris region. This case study focuses on Paris as an urban environment with multiple historical layers whose diversity of typology and buildings allows considering different ways of capturing wind energy. The objective of this approach is to categorize the different types of wind in urban areas. This, particularly the identification of the minimum and maximum wind spectrum, helps define the choice and performance of wind energy capturing devices that could be implanted there. The localization on the roof of a building, the type of wind, the altimetry of the device in relation to the levels of the roofs, the potential nuisances generated. The method allows identifying the characteristics of wind turbines in order to maximize their performance in an urban site with turbulent wind.Keywords: computer fluid dynamic simulation in urban environment, wind energy harvesting devices, net-zero energy building, urban wind behavior simulation, advanced building skin design methodology
Procedia PDF Downloads 10125725 Behavioral Experiments of Small Societies in Social Media: Facebook Expressions of Anchored Relationships
Authors: Nuran Öze
Abstract:
Communities and societies have been changing towards computer mediated communication. This paper explores online and offline identities and how relationships are formed and negotiated within internet environments which offer opportunities for people who know each other offline and move into relationships online. The expectations and norms of behavior within everyday life cause people to be embodied self. According to the age categories of Turkish Cypriots, their measurements of attitudes in Facebook will be investigated. Face-to-face field research and semi-structured interview methods are used in the study. Face-to-face interview has been done with Turkish Cypriots who are using Facebook already. According to the study, in constructing a linkage between real and virtual identities mostly affected from societal relations serves as a societal grooming tool for Turkish Cypriots.Keywords: facebook, identity, social media, virtual reality
Procedia PDF Downloads 30225724 Effect of Hybrid Learning in Higher Education
Authors: A. Meydanlioglu, F. Arikan
Abstract:
In recent years, thanks to the development of information and communication technologies, the computer and internet have been used widely in higher education. Internet-based education is impacting traditional higher education as online components increasingly become integrated into face-to-face (FTF) courses. The goal of combined internet-based and traditional education is to take full advantage of the benefits of each platform in order to provide an educational opportunity that can promote student learning better than can either platform alone. Research results show that the use of hybrid learning is more effective than online or FTF models in higher education. Due to the potential benefits, an increasing number of institutions are interested in developing hybrid courses, programs, and degrees. Future research should evaluate the effectiveness of hybrid learning. This paper is designed to determine the impact of hybrid learning on higher education.Keywords: e-learning, higher education, hybrid learning, online education
Procedia PDF Downloads 90925723 Interoperability Model Design of Smart Grid Power System
Authors: Seon-Hack Hong, Tae-Il Choi
Abstract:
Interoperability is defined as systems, components, and devices developed by different entities smoothly exchanging information and functioning organically without mutual consultation, being able to communicate with each other and computer systems of the same type or different types, and exchanging information or the ability of two or more systems to exchange information and use the information exchanged without extra effort. Insufficiencies such as duplication of functions when developing systems and applications due to lack of interoperability in the electric power system and low efficiency due to a lack of mutual information transmission system between the inside of the application program and the design is improved, and the seamless linkage of newly developed systems is improved. Since it is necessary to secure interoperability for this purpose, we designed the smart grid-based interoperability standard model in this paper.Keywords: interoperability, power system, common information model, SCADA, IEEE2030, Zephyr
Procedia PDF Downloads 12325722 Stress Solitary Waves Generated by a Second-Order Polynomial Constitutive Equation
Authors: Tsun-Hui Huang, Shyue-Cheng Yang, Chiou-Fen Shieha
Abstract:
In this paper, a nonlinear constitutive law and a curve fitting, two relationships between the stress-strain and the shear stress-strain for sandstone material were used to obtain a second-order polynomial constitutive equation. Based on the established polynomial constitutive equations and Newton’s second law, a mathematical model of the non-homogeneous nonlinear wave equation under an external pressure was derived. The external pressure can be assumed as an impulse function to simulate a real earthquake source. A displacement response under nonlinear two-dimensional wave equation was determined by a numerical method and computer-aided software. The results show that a suit pressure in the sandstone generates the phenomenon of stress solitary waves.Keywords: polynomial constitutive equation, solitary, stress solitary waves, nonlinear constitutive law
Procedia PDF Downloads 49725721 An Integrated Architecture of E-Learning System to Digitize the Learning Method
Authors: M. Touhidul Islam Sarker, Mohammod Abul Kashem
Abstract:
The purpose of this paper is to improve the e-learning system and digitize the learning method in the educational sector. The learner will login into e-learning platform and easily access the digital content, the content can be downloaded and take an assessment for evaluation. Learner can get access to these digital resources by using tablet, computer, and smart phone also. E-learning system can be defined as teaching and learning with the help of multimedia technologies and the internet by access to digital content. E-learning replacing the traditional education system through information and communication technology-based learning. This paper has designed and implemented integrated e-learning system architecture with University Management System. Moodle (Modular Object-Oriented Dynamic Learning Environment) is the best e-learning system, but the problem of Moodle has no school or university management system. In this research, we have not considered the school’s student because they are out of internet facilities. That’s why we considered the university students because they have the internet access and used technologies. The University Management System has different types of activities such as student registration, account management, teacher information, semester registration, staff information, etc. If we integrated these types of activity or module with Moodle, then we can overcome the problem of Moodle, and it will enhance the e-learning system architecture which makes effective use of technology. This architecture will give the learner to easily access the resources of e-learning platform anytime or anywhere which digitizes the learning method.Keywords: database, e-learning, LMS, Moodle
Procedia PDF Downloads 18825720 Wrist Pain, Technological Device Used, and Perceived Academic Performance Among the College of Computer Studies Students
Authors: Maquiling Jhuvie Jane R., Ojastro Regine B., Peroja Loreille Marie B., Pinili Joy Angela., Salve Genial Gail M., Villavicencio Marielle Irene B., Yap Alther Francis Garth B.
Abstract:
Introduction: This study investigated the impact of prolonged device usage on wrist pain and perceived academic performance among college students in Computer Studies. The research aims to explore the correlation between the frequency of technological device use and the incidence of wrist pain, as well as how this pain affects students' academic performance. The study seeks to provide insights that could inform interventions to promote better musculoskeletal health among students engaged in intensive technology use to further improve their academic performance. Method: The study utilized descriptive-correlational and comparative design, focusing on bona fide students from Silliman University’s College of Computer Studies during the second semester of 2023-2024. Participants were recruited through a survey sent via school email, with responses collected until March 30, 2024. Data was gathered using a password-protected device and Google Forms, ensuring restricted access to raw data. The demographic profile was summarized, and the prevalence of wrist pain and device usage were analyzed using percentages and weighted means. Statistical analyses included Spearman’s rank correlation coefficient to assess the relationship between wrist pain and device usage and an Independent T-test to evaluate differences in academic performance based on wrist pain presence. Alpha was set at 0.05. Results: The study revealed that 40% of College of Computer Studies students experience wrist pain, with 2 out of every 5 students affected. Laptops and desktops were the most frequently used devices for academic work, achieving a weighted mean of 4.511, while mobile phones and tablets received lower means of 4.183 and 1.911, respectively. The average academic performance score among students was 29.7, classified as ‘Good Performance.’ Notably, there was no significant relationship between the frequency of device usage and wrist pain, as indicated by p-values exceeding 0.05. However, a significant difference in perceived academic performance was observed, with students without wrist pain scoring an average of 30.39 compared to 28.72 for those with wrist pain and a p-value of 0.0134 confirming this distinction. Conclusion: The study revealed that about 40% of students in the College of Computer Studies experience wrist pain, but there is no significant link between device usage and pain occurrence. However, students without wrist pain demonstrated better academic performance than those with pain, suggesting that wrist health may impact academic success. These findings imply that physical therapy practices in the Philippines should focus on preventive strategies and ergonomic education to improve student health and performance.Keywords: wrist pain, frequency of use of technological devices, perceived academic performance, physical therapy
Procedia PDF Downloads 1225719 Clustering and Modelling Electricity Conductors from 3D Point Clouds in Complex Real-World Environments
Authors: Rahul Paul, Peter Mctaggart, Luke Skinner
Abstract:
Maintaining public safety and network reliability are the core objectives of all electricity distributors globally. For many electricity distributors, managing vegetation clearances from their above ground assets (poles and conductors) is the most important and costly risk mitigation control employed to meet these objectives. Light Detection And Ranging (LiDAR) is widely used by utilities as a cost-effective method to inspect their spatially-distributed assets at scale, often captured using high powered LiDAR scanners attached to fixed wing or rotary aircraft. The resulting 3D point cloud model is used by these utilities to perform engineering grade measurements that guide the prioritisation of vegetation cutting programs. Advances in computer vision and machine-learning approaches are increasingly applied to increase automation and reduce inspection costs and time; however, real-world LiDAR capture variables (e.g., aircraft speed and height) create complexity, noise, and missing data, reducing the effectiveness of these approaches. This paper proposes a method for identifying each conductor from LiDAR data via clustering methods that can precisely reconstruct conductors in complex real-world configurations in the presence of high levels of noise. It proposes 3D catenary models for individual clusters fitted to the captured LiDAR data points using a least square method. An iterative learning process is used to identify potential conductor models between pole pairs. The proposed method identifies the optimum parameters of the catenary function and then fits the LiDAR points to reconstruct the conductors.Keywords: point cloud, LİDAR data, machine learning, computer vision, catenary curve, vegetation management, utility industry
Procedia PDF Downloads 9925718 A Comprehensive Study of Camouflaged Object Detection Using Deep Learning
Authors: Khalak Bin Khair, Saqib Jahir, Mohammed Ibrahim, Fahad Bin, Debajyoti Karmaker
Abstract:
Object detection is a computer technology that deals with searching through digital images and videos for occurrences of semantic elements of a particular class. It is associated with image processing and computer vision. On top of object detection, we detect camouflage objects within an image using Deep Learning techniques. Deep learning may be a subset of machine learning that's essentially a three-layer neural network Over 6500 images that possess camouflage properties are gathered from various internet sources and divided into 4 categories to compare the result. Those images are labeled and then trained and tested using vgg16 architecture on the jupyter notebook using the TensorFlow platform. The architecture is further customized using Transfer Learning. Methods for transferring information from one or more of these source tasks to increase learning in a related target task are created through transfer learning. The purpose of this transfer of learning methodologies is to aid in the evolution of machine learning to the point where it is as efficient as human learning.Keywords: deep learning, transfer learning, TensorFlow, camouflage, object detection, architecture, accuracy, model, VGG16
Procedia PDF Downloads 14925717 Best Practices to Enhance Patient Security and Confidentiality When Using E-Health in South Africa
Authors: Lethola Tshikose, Munyaradzi Katurura
Abstract:
Information and Communication Technology (ICT) plays a critical role in improving daily healthcare processes. The South African healthcare organizations have adopted Information Systems to integrate their patient records. This has made it much easier for healthcare organizations because patient information can now be accessible at any time. The primary purpose of this research study was to investigate the best practices that can be applied to enhance patient security and confidentiality when using e-health systems in South Africa. Security and confidentiality are critical in healthcare organizations as they ensure safety in EHRs. The research study used an inductive research approach that included a thorough literature review; therefore, no data was collected. The research paper’s scope included patient data and possible security threats associated with healthcare systems. According to the study, South African healthcare organizations discovered various patient data security and confidentiality issues. The study also revealed that when it comes to handling patient data, health professionals sometimes make mistakes. Some may not be computer literate, which posed issues and caused data to be tempered with. The research paper recommends that healthcare organizations ensure that security measures are adequately supported and promoted by their IT department. This will ensure that adequate resources are distributed to keep patient data secure and confidential. Healthcare organizations must correctly use standards set up by IT specialists to solve patient data security and confidentiality issues. Healthcare organizations must make sure that their organizational structures are adaptable to improve security and confidentiality.Keywords: E-health, EHR, security, confidentiality, healthcare
Procedia PDF Downloads 5725716 Condition Based Assessment of Power Transformer with Modern Techniques
Authors: Piush Verma, Y. R. Sood
Abstract:
This paper provides the information on the diagnostics techniques for condition monitoring of power transformer (PT). This paper deals with the practical importance of the transformer diagnostic in the Electrical Engineering field. The life of the transformer depends upon its insulation i.e paper and oil. The major testing techniques applies on transformer oil and paper i.e dissolved gas analysis, furfural analysis, radio interface, acoustic emission, infra-red emission, frequency response analysis, power factor, polarization spectrum, magnetizing currents, turn and winding ratio. A review has been made on the modern development of this practical technology.Keywords: temperature, condition monitoring, diagnostics methods, paper analysis techniques, oil analysis techniques
Procedia PDF Downloads 43325715 Optimization of Hydraulic Fracturing for Horizontal Wells in Enhanced Geothermal Reservoirs
Authors: Qudratullah Muradi
Abstract:
Geothermal energy is a renewable energy source that can be found in abundance on our planet. Only a small fraction of it is currently converted to electrical power, though in recent years installed geothermal capacity has increased considerably all over the world. In this paper, we assumed a model for designing of Enhanced Geothermal System, EGS. We used computer modeling group, CMG reservoir simulation software to create the typical Hot Dry Rock, HDR reservoir. In this research two wells, one injection of cold water and one production of hot water are included in the model. There are some hydraulic fractures created by the mentioned software. And cold water is injected in order to produce energy from the reservoir. The result of injecting cold water to the reservoir and extracting geothermal energy is defined by some graphs at the end of this research. The production of energy is quantified in a period of 10 years.Keywords: geothermal energy, EGS, HDR, hydraulic fracturing
Procedia PDF Downloads 19925714 20 Definitions in 20 Years: Exploring the Evolution of Blended Learning Definitions from 2003-2022
Authors: Damian Gordon, Paul Doyle, Anna Becevel, Tina Baloh
Abstract:
The goal of this research is to explore the evolution of the concept of “blended learning” over a twenty-year period, to see whether or not the conceptualization has remained consistent or if it has become either more specific or more general. To achieve this goal, the term “blended learning” (and variations) was searched for in various bibliographical repositories for each year 2003-2022 to locate a highly cited paper that is not behind a paywall, to locate unique definitions that would be freely available to all academics each year. Each of the twenty unique definitions is explored to identify how they categorize both the Classroom Component and the Computer Component of blended learning, as well as identify which discipline each definition originates from and which country it comes from to see if there are any significant geographical variations. Based on this analysis, trends that appear in the definitions are noted, as well as an overall interpretation of the notion of “Blended Learning.”Keywords: blended learning, definitions of blended learning, e-learning, thematic searches
Procedia PDF Downloads 12925713 The Post-Hegemony of Post-Capitalism: Towards a Political Theory of Open Cooperativism
Authors: Vangelis Papadimitropoulos
Abstract:
The paper is part of the research project “Techno-Social Innovation in the Collaborative Economy'', funded by the Hellenic Foundation of Research and Innovation for the years 2022-2024. The research project examines the normative and empirical conditions of grassroots technologically driven innovation, potentially enabling the transition towards a commons-oriented post-capitalist economy. The project carries out a conceptually led and empirically grounded multi-case study of the digital commons, open-source technologies, platform cooperatives, open cooperatives and Distributed Autonomous Organizations (DAOs) on the Blockchain. The methodological scope of research is interdisciplinary inasmuch as it comprises political theory, economics, sustainability science and computer science, among others. The research draws specifically on Michel Bauwens and Vasilis Kostakis' model of open cooperativism between the commons, ethical market entities and a partner state. Bauwens and Kostakis advocate for a commons-based counter-hegemonic post-capitalist transition beyond and against neoliberalism. The research further employs Laclau and Mouffe's discourse theory of hegemony to introduce a post-hegemonic conceptualization of the model of open cooperativism. Thus, the paper aims to outline the theoretical contribution of the research project to contemporary political theory debates on post-capitalism and the collaborative economy.Keywords: open cooperativism, techno-social innovation, post-hegemony, post-capitalism
Procedia PDF Downloads 6625712 NFC Kenaf Core Graphene Paper: In-situ Method Application
Authors: M. A. Izzati, R. Rosazley, A. W. Fareezal, M. Z. Shazana, I. Rushdan, M. Jani
Abstract:
Ultrasonic probe were using to produce nanofibrillated cellulose (NFC) kenaf core. NFC kenaf core and graphene was mixed using in-situ method with the 5V voltage for 24 hours. The resulting NFC graphene paper was characterized by field emission scanning electron microscopy (FESEM), fourier transformed infrared (FTIR) spectra and thermogavimetric analysis (TGA). The properties of NFC kenaf core graphene paper are compared with properties of pure NFC kenaf core paper.Keywords: NFC, kenaf core, graphene, in-situ method
Procedia PDF Downloads 39425711 Mathematical Modeling of a Sub-Wet Bulb Temperature Evaporative Cooling Using Porous Ceramic Materials
Authors: Meryem Kanzari, Rabah Boukhanouf, Hatem G. Ibrahim
Abstract:
Indirect Evaporative Cooling process has the advantage of supplying cool air at constant moisture content. However, such system can only supply air at temperatures above wet bulb temperature. This paper presents a mathematical model for a sub-wet bulb temperature indirect evaporative cooling arrangement that can overcome this limitation and supply cool air at temperatures approaching dew point and without increasing its moisture content. In addition, the use of porous ceramics as wet media materials offers the advantage of integration into building elements. Results of the computer show that the proposed design is capable of cooling air to temperatures lower than the ambient wet bulb temperature and achieving wet bulb effectiveness of about 1.17.Keywords: indirect evaporative cooling, porous ceramic, sub-wet bulb temperature, mathematical modeling
Procedia PDF Downloads 29425710 Video Games Technologies Approach for Their Use in the Classroom
Authors: Daniel Vargas-Herrera, Ivette Caldelas, Fernando Brambila-Paz, Rodrigo Montufar-Chaveznava
Abstract:
In this paper, we present the advances corresponding to the implementation of a set of educational materials based on video games technologies. Essentially these materials correspond to projects developed and under development as bachelor thesis of some Computer Engineering students of the Engineering School. All materials are based on the Unity SDK; integrating some devices such as kinect, leap motion, oculus rift, data gloves and Google cardboard. In detail, we present a virtual reality application for neurosciences students (suitable for neural rehabilitation), and virtual scenes for the Google cardboard, which will be used by the psychology students for phobias treatment. The objective is these materials will be located at a server to be available for all students, in the classroom or in the cloud, considering the use of smartphones has been widely extended between students.Keywords: virtual reality, interactive technologies, video games, educational materials
Procedia PDF Downloads 65725709 The Paralinguistic Function of Emojis in Twitter Communication
Authors: Yasmin Tantawi, Mary Beth Rosson
Abstract:
In response to the dearth of information about emoji use for different purposes in different settings, this paper investigates the paralinguistic function of emojis within Twitter communication in the United States. To conduct this investigation, the Twitter feeds from 16 population centers spread throughout the United States were collected from the Twitter public API. One hundred tweets were collected from each population center, totaling to 1,600 tweets. Tweets containing emojis were next extracted using the “emot” Python package; these were then analyzed via the IBM Watson API Natural Language Understanding module to identify the topics discussed. A manual content analysis was then conducted to ascertain the paralinguistic and emotional features of the emojis used in these tweets. We present our characterization of emoji usage in Twitter and discuss implications for the design of Twitter and other text-based communication tools.Keywords: computer-mediated communication, content analysis, paralinguistics, sociology
Procedia PDF Downloads 16025708 Digital Musical Organology: The Audio Games: The Question of “A-Musicological” Interfaces
Authors: Hervé Zénouda
Abstract:
This article seeks to shed light on an emerging creative field: "Audio games," at the crossroads between video games and computer music. Indeed, many applications, which propose entertaining audio-visual experiences with the objective of musical creation, are available today for different supports (game consoles, computers, cell phones). The originality of this field is the use of the gameplay of video games applied to music composition. Thus, composing music using interfaces but also cognitive logics that we qualify as "a-musicological" seem to us particularly interesting from the perspective of musical digital organology. This field raises questions about the representation of sound and musical structures and develops new instrumental gestures and strategies of musical composition. We will try in this article to define the characteristics of this field by highlighting some historical milestones (abstract cinema, game theory in music, actions, and graphic scores) as well as the novelties brought by digital technologies.Keywords: audio-games, video games, computer generated music, gameplay, interactivity, synesthesia, sound interfaces, relationships image/sound, audiovisual music
Procedia PDF Downloads 11225707 Diagnosis of Diabetes Using Computer Methods: Soft Computing Methods for Diabetes Detection Using Iris
Authors: Piyush Samant, Ravinder Agarwal
Abstract:
Complementary and Alternative Medicine (CAM) techniques are quite popular and effective for chronic diseases. Iridology is more than 150 years old CAM technique which analyzes the patterns, tissue weakness, color, shape, structure, etc. for disease diagnosis. The objective of this paper is to validate the use of iridology for the diagnosis of the diabetes. The suggested model was applied in a systemic disease with ocular effects. 200 subject data of 100 each diabetic and non-diabetic were evaluated. Complete procedure was kept very simple and free from the involvement of any iridologist. From the normalized iris, the region of interest was cropped. All 63 features were extracted using statistical, texture analysis, and two-dimensional discrete wavelet transformation. A comparison of accuracies of six different classifiers has been presented. The result shows 89.66% accuracy by the random forest classifier.Keywords: complementary and alternative medicine, classification, iridology, iris, feature extraction, disease prediction
Procedia PDF Downloads 40725706 Examining Relationship between Programming Performance, Programming Self Efficacy and Math Success
Authors: Mustafa Ekici, Sacide Güzin Mazman
Abstract:
Programming is the one of ability in computer science fields which is generally perceived difficult by students and various individual differences have been implicated in that ability success. Although several factors that affect programming ability have been identified over the years, there is not still a full understanding of why some students learn to program easily and quickly while others find it complex and difficult. Programming self-efficacy and mathematic success are two of those essential individual differences which are handled as having important effect on the programming success. This study aimed to identify the relationship between programming performance, programming self efficacy and mathematics success. The study group is consisted of 96 undergraduates from Department of Econometrics of Uşak University. 38 (39,58%) of the participants are female while 58 (60,41%) of them are male. Study was conducted in the programming-I course during 2014-2015 fall term. Data collection tools are comprised of programming course final grades, programming self efficacy scale and a mathematics achievement test. Data was analyzed through correlation analysis. The result of study will be reported in the full text of the study.Keywords: programming performance, self efficacy, mathematic success, computer science
Procedia PDF Downloads 50225705 On a Continuous Formulation of Block Method for Solving First Order Ordinary Differential Equations (ODEs)
Authors: A. M. Sagir
Abstract:
The aim of this paper is to investigate the performance of the developed linear multistep block method for solving first order initial value problem of Ordinary Differential Equations (ODEs). The method calculates the numerical solution at three points simultaneously and produces three new equally spaced solution values within a block. The continuous formulations enable us to differentiate and evaluate at some selected points to obtain three discrete schemes, which were used in block form for parallel or sequential solutions of the problems. A stability analysis and efficiency of the block method are tested on ordinary differential equations involving practical applications, and the results obtained compared favorably with the exact solution. Furthermore, comparison of error analysis has been developed with the help of computer software.Keywords: block method, first order ordinary differential equations, linear multistep, self-starting
Procedia PDF Downloads 30625704 Automated Computer-Vision Analysis Pipeline of Calcium Imaging Neuronal Network Activity Data
Authors: David Oluigbo, Erik Hemberg, Nathan Shwatal, Wenqi Ding, Yin Yuan, Susanna Mierau
Abstract:
Introduction: Calcium imaging is an established technique in neuroscience research for detecting activity in neural networks. Bursts of action potentials in neurons lead to transient increases in intracellular calcium visualized with fluorescent indicators. Manual identification of cell bodies and their contours by experts typically takes 10-20 minutes per calcium imaging recording. Our aim, therefore, was to design an automated pipeline to facilitate and optimize calcium imaging data analysis. Our pipeline aims to accelerate cell body and contour identification and production of graphical representations reflecting changes in neuronal calcium-based fluorescence. Methods: We created a Python-based pipeline that uses OpenCV (a computer vision Python package) to accurately (1) detect neuron contours, (2) extract the mean fluorescence within the contour, and (3) identify transient changes in the fluorescence due to neuronal activity. The pipeline consisted of 3 Python scripts that could both be easily accessed through a Python Jupyter notebook. In total, we tested this pipeline on ten separate calcium imaging datasets from murine dissociate cortical cultures. We next compared our automated pipeline outputs with the outputs of manually labeled data for neuronal cell location and corresponding fluorescent times series generated by an expert neuroscientist. Results: Our results show that our automated pipeline efficiently pinpoints neuronal cell body location and neuronal contours and provides a graphical representation of neural network metrics accurately reflecting changes in neuronal calcium-based fluorescence. The pipeline detected the shape, area, and location of most neuronal cell body contours by using binary thresholding and grayscale image conversion to allow computer vision to better distinguish between cells and non-cells. Its results were also comparable to manually analyzed results but with significantly reduced result acquisition times of 2-5 minutes per recording versus 10-20 minutes per recording. Based on these findings, our next step is to precisely measure the specificity and sensitivity of the automated pipeline’s cell body and contour detection to extract more robust neural network metrics and dynamics. Conclusion: Our Python-based pipeline performed automated computer vision-based analysis of calcium image recordings from neuronal cell bodies in neuronal cell cultures. Our new goal is to improve cell body and contour detection to produce more robust, accurate neural network metrics and dynamic graphs.Keywords: calcium imaging, computer vision, neural activity, neural networks
Procedia PDF Downloads 8225703 U-Net Based Multi-Output Network for Lung Disease Segmentation and Classification Using Chest X-Ray Dataset
Authors: Jaiden X. Schraut
Abstract:
Medical Imaging Segmentation of Chest X-rays is used for the purpose of identification and differentiation of lung cancer, pneumonia, COVID-19, and similar respiratory diseases. Widespread application of computer-supported perception methods into the diagnostic pipeline has been demonstrated to increase prognostic accuracy and aid doctors in efficiently treating patients. Modern models attempt the task of segmentation and classification separately and improve diagnostic efficiency; however, to further enhance this process, this paper proposes a multi-output network that follows a U-Net architecture for image segmentation output and features an additional CNN module for auxiliary classification output. The proposed model achieves a final Jaccard Index of .9634 for image segmentation and a final accuracy of .9600 for classification on the COVID-19 radiography database.Keywords: chest X-ray, deep learning, image segmentation, image classification
Procedia PDF Downloads 144