Search results for: user interface.
2852 An Ultrasonic Signal Processing System for Tomographic Imaging of Reinforced Concrete Structures
Authors: Edwin Forero-Garcia, Jaime Vitola, Brayan Cardenas, Johan Casagua
Abstract:
This research article presents the integration of electronic and computer systems, which developed an ultrasonic signal processing system that performs the capture, adaptation, and analog-digital conversion to later carry out its processing and visualization. The capture and adaptation of the signal were carried out from the design and implementation of an analog electronic system distributed in stages: 1. Coupling of impedances; 2. Analog filter; 3. Signal amplifier. After the signal conditioning was carried out, the ultrasonic information was digitized using a digital microcontroller to carry out its respective processing. The digital processing of the signals was carried out in MATLAB software for the elaboration of A-Scan, B and D-Scan types of ultrasonic images. Then, advanced processing was performed using the SAFT technique to improve the resolution of the Scan-B-type images. Thus, the information from the ultrasonic images was displayed in a user interface developed in .Net with Visual Studio. For the validation of the system, ultrasonic signals were acquired, and in this way, the non-invasive inspection of the structures was carried out and thus able to identify the existing pathologies in them.Keywords: acquisition, signal processing, ultrasound, SAFT, HMI
Procedia PDF Downloads 1072851 A Software Tool for Computer Forensic Investigation Using Client-Side Web History Visualization
Authors: Francisca Onaolapo Oladipo, Peter Afam Ugwu
Abstract:
Records of user activities which are valuable for forensic investigation purposes are provided by web browsers -these records in most cases are not in visual formats that are easily understood, thereby requiring some extra processes. This paper describes the implementation of a software tool for client-side web history visualization providing suitable forensic evidence for investigative purposes. Visual C#, Perl and gnuplot were deployed on Windows Operating System (OS) environment to implement the system and the resulting tool parses and transforms a web browser history into a visual format that enables an investigator to quickly and efficiently explore, understand, and interpret the user online activities in the context of a specific investigation. The system was tested using two forensic cases: the client-side web history files generated by Mozilla Firefox browser was extracted using MozillaHistoryView utility, then parsed and visualized using bar and stacked column charts. From the visual representation, results of user web activities across various productive and non-productive websites were obtained.Keywords: history, forensics, visualization, web activities
Procedia PDF Downloads 2952850 Privacy Preservation Concerns and Information Disclosure on Social Networks: An Ongoing Research
Authors: Aria Teimourzadeh, Marc Favier, Samaneh Kakavand
Abstract:
The emergence of social networks has revolutionized the exchange of information. Every behavior on these platforms contributes to the generation of data known as social network data that are processed, stored and published by the social network service providers. Hence, it is vital to investigate the role of these platforms in user data by considering the privacy measures, especially when we observe the increased number of individuals and organizations engaging with the current virtual platforms without being aware that the data related to their positioning, connections and behavior is uncovered and used by third parties. Performing analytics on social network datasets may result in the disclosure of confidential information about the individuals or organizations which are the members of these virtual environments. Analyzing separate datasets can reveal private information about relationships, interests and more, especially when the datasets are analyzed jointly. Intentional breaches of privacy is the result of such analysis. Addressing these privacy concerns requires an understanding of the nature of data being accumulated and relevant data privacy regulations, as well as motivations for disclosure of personal information on social network platforms. Some significant points about how user's online information is controlled by the influence of social factors and to what extent the users are concerned about future use of their personal information by the organizations, are highlighted in this paper. Firstly, this research presents a short literature review about the structure of a network and concept of privacy in Online Social Networks. Secondly, the factors of user behavior related to privacy protection and self-disclosure on these virtual communities are presented. In other words, we seek to demonstrates the impact of identified variables on user information disclosure that could be taken into account to explain the privacy preservation of individuals on social networking platforms. Thirdly, a few research directions are discussed to address this topic for new researchers.Keywords: information disclosure, privacy measures, privacy preservation, social network analysis, user experience
Procedia PDF Downloads 2812849 Assisted Video Colorization Using Texture Descriptors
Authors: Andre Peres Ramos, Franklin Cesar Flores
Abstract:
Colorization is the process of add colors to a monochromatic image or video. Usually, the process involves to segment the image in regions of interest and then apply colors to each one, for videos, this process is repeated for each frame, which makes it a tedious and time-consuming job. We propose a new assisted method for video colorization; the user only has to colorize one frame, and then the colors are propagated to following frames. The user can intervene at any time to correct eventual errors in color assignment. The method consists of to extract intensity and texture descriptors from the frames and then perform a feature matching to determine the best color for each segment. To reduce computation time and give a better spatial coherence we narrow the area of search and give weights for each feature to emphasize texture descriptors. To give a more natural result, we use an optimization algorithm to make the color propagation. Experimental results in several image sequences, compared to others existing methods, demonstrates that the proposed method perform a better colorization with less time and user interference.Keywords: colorization, feature matching, texture descriptors, video segmentation
Procedia PDF Downloads 1622848 Ultrasensitive Hepatitis B Virus Detection in Blood Using Nano-Porous Silicon Oxide: Towards POC Diagnostics
Authors: N. Das, N. Samanta, L. Pandey, C. Roy Chaudhuri
Abstract:
Early diagnosis of infection like Hep-B virus in blood is important for low cost medical treatment. For this purpose, it is desirable to develop a point of care device which should be able to detect trace quantities of the target molecule in blood. In this paper, we report a nanoporous silicon oxide sensor which is capable of detecting down to 1fM concentration of Hep-B surface antigen in blood without the requirement of any centrifuge or pre-concentration. This has been made possible by the presence of resonant peak in the sensitivity characteristics. This peak is observed to be dependent only on the concentration of the specific antigen and not on the interfering species in blood serum. The occurrence of opposite impedance change within the pores and at the bottom of the pore is responsible for this effect. An electronic interface has also been designed to provide a display of the virus concentration.Keywords: impedance spectroscopy, ultrasensitive detection in blood, peak frequency, electronic interface
Procedia PDF Downloads 4002847 Exploring Time-Series Phosphoproteomic Datasets in the Context of Network Models
Authors: Sandeep Kaur, Jenny Vuong, Marcel Julliard, Sean O'Donoghue
Abstract:
Time-series data are useful for modelling as they can enable model-evaluation. However, when reconstructing models from phosphoproteomic data, often non-exact methods are utilised, as the knowledge regarding the network structure, such as, which kinases and phosphatases lead to the observed phosphorylation state, is incomplete. Thus, such reactions are often hypothesised, which gives rise to uncertainty. Here, we propose a framework, implemented via a web-based tool (as an extension to Minardo), which given time-series phosphoproteomic datasets, can generate κ models. The incompleteness and uncertainty in the generated model and reactions are clearly presented to the user via the visual method. Furthermore, we demonstrate, via a toy EGF signalling model, the use of algorithmic verification to verify κ models. Manually formulated requirements were evaluated with regards to the model, leading to the highlighting of the nodes causing unsatisfiability (i.e. error causing nodes). We aim to integrate such methods into our web-based tool and demonstrate how the identified erroneous nodes can be presented to the user via the visual method. Thus, in this research we present a framework, to enable a user to explore phosphorylation proteomic time-series data in the context of models. The observer can visualise which reactions in the model are highly uncertain, and which nodes cause incorrect simulation outputs. A tool such as this enables an end-user to determine the empirical analysis to perform, to reduce uncertainty in the presented model - thus enabling a better understanding of the underlying system.Keywords: κ-models, model verification, time-series phosphoproteomic datasets, uncertainty and error visualisation
Procedia PDF Downloads 2552846 Secure Text Steganography for Microsoft Word Document
Authors: Khan Farhan Rafat, M. Junaid Hussain
Abstract:
Seamless modification of an entity for the purpose of hiding a message of significance inside its substance in a manner that the embedding remains oblivious to an observer is known as steganography. Together with today's pervasive registering frameworks, steganography has developed into a science that offers an assortment of strategies for stealth correspondence over the globe that must, however, need a critical appraisal from security breach standpoint. Microsoft Word is amongst the preferably used word processing software, which comes as a part of the Microsoft Office suite. With a user-friendly graphical interface, the richness of text editing, and formatting topographies, the documents produced through this software are also most suitable for stealth communication. This research aimed not only to epitomize the fundamental concepts of steganography but also to expound on the utilization of Microsoft Word document as a carrier for furtive message exchange. The exertion is to examine contemporary message hiding schemes from security aspect so as to present the explorative discoveries and suggest enhancements which may serve a wellspring of information to encourage such futuristic research endeavors.Keywords: hiding information in plain sight, stealth communication, oblivious information exchange, conceal, steganography
Procedia PDF Downloads 2412845 Bone Fracture Detection with X-Ray Images Using Mobilenet V3 Architecture
Authors: Ashlesha Khanapure, Harsh Kashyap, Abhinav Anand, Sanjana Habib, Anupama Bidargaddi
Abstract:
Technologies that are developing quickly are being developed daily in a variety of disciplines, particularly the medical field. For the purpose of detecting bone fractures in X-ray pictures of different body segments, our work compares the ResNet-50 and MobileNetV3 architectures. It evaluates accuracy and computing efficiency with X-rays of the elbow, hand, and shoulder from the MURA dataset. Through training and validation, the models are evaluated on normal and fractured images. While ResNet-50 showcases superior accuracy in fracture identification, MobileNetV3 showcases superior speed and resource optimization. Despite ResNet-50’s accuracy, MobileNetV3’s swifter inference makes it a viable choice for real-time clinical applications, emphasizing the importance of balancing computational efficiency and accuracy in medical imaging. We created a graphical user interface (GUI) for MobileNet V3 model bone fracture detection. This research underscores MobileNetV3’s potential to streamline bone fracture diagnoses, potentially revolutionizing orthopedic medical procedures and enhancing patient care.Keywords: CNN, MobileNet V3, ResNet-50, healthcare, MURA, X-ray, fracture detection
Procedia PDF Downloads 632844 Interface Fracture of Sandwich Composite Influenced by Multiwalled Carbon Nanotube
Authors: Alak Kumar Patra, Nilanjan Mitra
Abstract:
Higher strength to weight ratio is the main advantage of sandwich composite structures. Interfacial delamination between the face sheet and core is a major problem in these structures. Many research works are devoted to improve the interfacial fracture toughness of composites majorities of which are on nano and laminated composites. Work on influence of multiwalled carbon nano-tubes (MWCNT) dispersed resin system on interface fracture of glass-epoxy PVC core sandwich composite is extremely limited. Finite element study is followed by experimental investigation on interface fracture toughness of glass-epoxy (G/E) PVC core sandwich composite with and without MWCNT. Results demonstrate an improvement in interface fracture toughness values (Gc) of samples with a certain percentages of MWCNT. In addition, dispersion of MWCNT in epoxy resin through sonication followed by mixing of hardener and vacuum resin infusion (VRI) technology used in this study is an easy and cost effective methodology in comparison to previously adopted other methods limited to laminated composites. The study also identifies the optimum weight percentage of MWCNT addition in the resin system for maximum performance gain in interfacial fracture toughness. The results agree with finite element study, high-resolution transmission electron microscope (HRTEM) analysis and fracture micrograph of field emission scanning electron microscope (FESEM) investigation. Interface fracture toughness (GC) of the DCB sandwich samples is calculated using the compliance calibration (CC) method considering the modification due to shear. Compliance (C) vs. crack length (a) data of modified sandwich DCB specimen is fitted to a power function of crack length. The calculated mean value of the exponent n from the plots of experimental results is 2.22 and is different from the value (n=3) prescribed in ASTM D5528-01for mode 1 fracture toughness of laminate composites (which is the basis for modified compliance calibration method). Differentiating C with respect to crack length (a) and substituting it in the expression GC provides its value. The research demonstrates improvement of 14.4% in peak load carrying capacity and 34.34% in interface fracture toughness GC for samples with 1.5 wt% MWCNT (weight % being taken with respect to weight of resin) in comparison to samples without MWCNT. The paper focuses on significant improvement in experimentally determined interface fracture toughness of sandwich samples with MWCNT over the samples without MWCNT using much simpler method of sonication. Good dispersion of MWCNT was observed in HRTEM with 1.5 wt% MWCNT addition in comparison to other percentages of MWCNT. FESEM studies have also demonstrated good dispersion and fiber bridging of MWCNT in resin system. Ductility is also observed to be higher for samples with MWCNT in comparison to samples without.Keywords: carbon nanotube, epoxy resin, foam, glass fibers, interfacial fracture, sandwich composite
Procedia PDF Downloads 3032843 A Case Study on Post-Occupancy Evaluation of User Satisfaction in Higher Educational Buildings
Authors: Yuanhong Zhao, Qingping Yang, Andrew Fox, Tao Zhang
Abstract:
Post-occupancy evaluation (POE) is a systematic approach to assess the actual building performance after the building has been occupied for some time. In this paper, a structured POE assessment was conducted using the building use survey (BUS) methodology in two higher educational buildings in the United Kingdom. This study aims to help close the building performance gap, provide optimized building operation suggestions, and to improve occupants’ satisfaction level. In this research, the questionnaire survey investigated the influences of environmental factors on user satisfaction from the main aspects of building overall design, thermal comfort, perceived control, indoor environment quality for noise, lighting, ventilation, and other non-environmental factors, such as the background information about age, sex, time in buildings, workgroup size, and so on. The results indicate that the occupant satisfaction level with the main aspects of building overall design, indoor environment quality, and thermal comfort in summer and winter on both two buildings, which is lower than the benchmark data. The feedback of this POE assessment has been reported to the building management team to allow managers to develop high-performance building operation plans. Finally, this research provided improvement suggestions to the building operation system to narrow down the performance gap and improve the user work experience satisfaction and productivity level.Keywords: building performance assessment systems, higher educational buildings, post-occupancy evaluation, user satisfaction
Procedia PDF Downloads 1522842 The OLOS® Way to Cultural Heritage: User Interface with Anthropomorphic Characteristics
Authors: Daniele Baldacci, Remo Pareschi
Abstract:
Augmented Reality and Augmented Intelligence are radically changing information technology. The path that starts from the keyboard and then, passing through milestones such as Siri, Alexa and other vocal avatars, reaches a more fluid and natural communication with computers, thus converting the dichotomy between man and machine into a harmonious interaction, now heads unequivocally towards a new IT paradigm, where holographic computing will play a key role. The OLOS® platform contributes substantially to this trend in that it infuses computers with human features, by transferring the gestures and expressions of persons of flesh and bones to anthropomorphic holographic interfaces which in turn will use them to interact with real-life humans. In fact, we could say, boldly but with a solid technological background to back the statement, that OLOS® gives reality to an altogether new entity, placed at the exact boundary between nature and technology, namely the holographic human being. Holographic humans qualify as the perfect carriers for the virtual reincarnation of characters handed down from history and tradition. Thus, they provide for an innovative and highly immersive way of experiencing our cultural heritage as something alive and pulsating in the present.Keywords: digital cinematography, human-computer interfaces, holographic simulation, interactive museum exhibits
Procedia PDF Downloads 1152841 Reduction of Multiple User Interference for Optical CDMA Systems Using Successive Interference Cancellation Scheme
Authors: Tawfig Eltaif, Hesham A. Bakarman, N. Alsowaidi, M. R. Mokhtar, Malek Harbawi
Abstract:
In Commonly, it is primary problem that there is multiple user interference (MUI) noise resulting from the overlapping among the users in optical code-division multiple access (OCDMA) system. In this article, we aim to mitigate this problem by studying an interference cancellation scheme called successive interference cancellation (SIC) scheme. This scheme will be tested on two different detection schemes, spectral amplitude coding (SAC) and direct detection systems (DS), using partial modified prime (PMP) as the signature codes. It was found that SIC scheme based on both SAC and DS methods had a potential to suppress the intensity noise, that is to say, it can mitigate MUI noise. Furthermore, SIC/DS scheme showed much lower bit error rate (BER) performance relative to SIC/SAC scheme for different magnitude of effective power. Hence, many more users can be supported by SIC/DS receiver system.Keywords: optical code-division multiple access (OCDMA), successive interference cancellation (SIC), multiple user interference (MUI), spectral amplitude coding (SAC), partial modified prime code (PMP)
Procedia PDF Downloads 5212840 Improving System Performance through User's Resource Access Patterns
Authors: K. C. Wong
Abstract:
This paper demonstrates a number of examples in the hope to shed some light on the possibility of designing future operating systems in a more adaptation-based manner. A modern operating system, we conceive, should possess the capability of 'learning' in such a way that it can dynamically adjust its services and behavior according to the current status of the environment in which it operates. In other words, a modern operating system should play a more proactive role during the session of providing system services to users. As such, a modern operating system is expected to create a computing environment, in which its users are provided with system services more matching their dynamically changing needs. The examples demonstrated in this paper show that user's resource access patterns 'learned' and determined during a session can be utilized to improve system performance and hence to provide users with a better and more effective computing environment. The paper also discusses how to use the frequency, the continuity, and the duration of resource accesses in a session to quantitatively measure and determine user's resource access patterns for the examples shown in the paper.Keywords: adaptation-based systems, operating systems, resource access patterns, system performance
Procedia PDF Downloads 1452839 Application of Chinese Remainder Theorem to Find The Messages Sent in Broadcast
Authors: Ayubi Wirara, Ardya Suryadinata
Abstract:
Improper application of the RSA algorithm scheme can cause vulnerability to attacks. The attack utilizes the relationship between broadcast messages sent to the user with some fixed polynomial functions that belong to each user. Scheme attacks carried out by applying the Chinese Remainder Theorem to obtain a general polynomial equation with the same modulus. The formation of the general polynomial becomes a first step to get back the original message. Furthermore, to solve these equations can use Coppersmith's theorem.Keywords: RSA algorithm, broadcast message, Chinese Remainder Theorem, Coppersmith’s theorem
Procedia PDF Downloads 3412838 Decomposition of Solidification Carbides during Cyclic Thermal Treatments in a Co-Based Alloy Deposit Applied to Stainless Steel
Authors: Sellidj Abdelaziz, Lebaili Soltane
Abstract:
A cobalt-based alloy type Co-Cr-Ni-WC was deposited by plasma transferred arc projection (PTA) on a stainless steel valve. The alloy is characterized at the equilibrium by a solid solution Co (γ) mainly dendritic, and eutectic carbides M₇C₃ and ηM₆C. At the deposit/substrate interface, this microstructure is modified by the fast cooling mode of the alloy when applied in the liquid state on the relatively cold steel substrate. The structure formed in this case is heterogeneous and metastable phases can occur and evolve over temperature service. Coating properties and reliability are directly related to microstructures formed during deposition. We were interested more particularly in this microstructure formed during the solidification of the deposit in the region of the interface joining the soldered couple and its evolution during cyclic heat treatments at temperatures similar to those of the thermal environment of the valve. The characterization was carried out by SEM-EDS microprobe CAMECA, XRD, and micro hardness profiles. The deposit obtained has a linear and regular appearance that is free of cracks and with little porosity. The morphology of the microstructure represents solidification stages that are relatively fast with a temperature gradient high at the beginning of the interface by forming a plane front solid solution Co (γ). It gradually changes with the decreasing temperature gradient by getting farther from the junction towards the outer limit of the deposit. The matrix takes the forms: cellular, mixed (cells and dendrites) and dendritic. Dendritic growth is done according to primary ramifications in the direction of the heat removal which takes place in the direction perpendicular to the interface, towards the external surface of the deposit, following secondary and tertiary undeveloped arms. The eutectic carbides M₇C₃ and ηM₆C formed are very thin and are located in the intercellular and interdendritic spaces of the solid solution Co (γ).Keywords: Co-Ni-Cr-W-C alloy, solid deposit, microstructure, carbides, cyclic heat treatment
Procedia PDF Downloads 1162837 Using LTE-Sim in New Hanover Decision Algorithm for 2-Tier Macrocell-Femtocell LTE Network
Authors: Umar D. M., Aminu A. M., Izaddeen K. Y.
Abstract:
Deployments of mini macrocell base stations also referred to as femtocells, improve the quality of service of indoor and outdoor users. Nevertheless, mobility management remains a key issue with regards to their deployment. This paper is leaned towards this issue, with an in-depth focus on the most important aspect of mobility management -handover. In handover management, making a handover decision in the LTE two-tier macrocell femtocell network is a crucial research area. Decision algorithms in this research are classified and comparatively analyzed according to received signal strength, user equipment speed, cost function, and interference. However, it was observed that most of the discussed decision algorithms fail to consider cell selection with hybrid access policy in a single macrocell multiple femtocell scenario, another observation was a majority of these algorithms lack the incorporation of user equipment residence parameter. Not including this parameter boosts the number of unnecessary handover occurrence. To deal with these issues, a sophisticated handover decision algorithm is proposed. The proposed algorithm considers the user’s velocity, received signal strength, residence time, as well as the femtocell base station’s access policy. Simulation results have shown that the proposed algorithm reduces the number of unnecessary handovers when compared to conventional received signal strength-based handover decision algorithm.Keywords: user-equipment, radio signal service, long term evolution, mobility management, handoff
Procedia PDF Downloads 1242836 Outline of a Technique for the Recommendation of Tourism Products in Cuba Using GIS
Authors: Jesse D. Cano, Marlon J. Remedios
Abstract:
Cuban tourism has developed so much in the last 30 years to the point of becoming one of the engines of the Cuban economy. With such a development, Cuban companies opting for e-tourism as a way to publicize their products and attract customers has also grown. Despite this fact, the majority of Cuban tourism-themed websites simply provide information on the different products and services they offer which results in many cases, in the user getting overwhelmed with the amount of information available which results in the user abandoning the search before he can find a product that fits his needs. Customization has been recognized as a critical factor for successful electronic tourism business and the use of recommender systems is the best approach to address the problem of personalization. This paper aims to outline a preliminary technique to obtain predictions about which products a particular user would give a better evaluation; these products would be those which the website would show in the first place. To achieve this, the theoretical elements of the Cuban tourism environment are discussed; recommendation systems and geographic information systems as tools for information representation are also discussed. Finally, for each structural component identified, we define a set of rules that allows obtaining an electronic tourism system that handles the personalization of the service provided effectively.Keywords: geographic information system, technique, tourism products, recommendation
Procedia PDF Downloads 5032835 Relationship between Interfacial Instabilities and Mechanical Strength of Multilayer Symmetric Polymer Melts
Authors: Mohammad Ranjbaran Madiseh
Abstract:
In this research, an experimental apparatus has been developed for observing interfacial stability and deformation of multilayer pressure-driven channel flows. The interface instability of the co-extrusion flow of polyethylene and polypropylene is studied experimentally in a slit geometry. By investigating the growing interfacial wave (IW) and tensile stress of extrudate samples, a relationship between interfacial instability (II) and mechanical properties of polypropylene (PP) and high-density polyethylene (HDPE) has been established. It is shown that the mechanism of interfacial strength is related to interfacial instabilities as well as interfacial strength. It is shown that there is an ability to forecast the quality of final products in the co-extrusion process. In this study, it is found that the instability is controlled by its dominant wave number, which is associated with maximum tensile stress at the interface.Keywords: interfacial instability, interfacial strength, wave number, interfacial wave
Procedia PDF Downloads 922834 Development of an Automatic Control System for ex vivo Heart Perfusion
Authors: Pengzhou Lu, Liming Xin, Payam Tavakoli, Zhonghua Lin, Roberto V. P. Ribeiro, Mitesh V. Badiwala
Abstract:
Ex vivo Heart Perfusion (EVHP) has been developed as an alternative strategy to expand cardiac donation by enabling resuscitation and functional assessment of hearts donated from marginal donors, which were previously not accepted. EVHP parameters, such as perfusion flow (PF) and perfusion pressure (PP) are crucial for optimal organ preservation. However, with the heart’s constant physiological changes during EVHP, such as coronary vascular resistance, manual control of these parameters is rendered imprecise and cumbersome for the operator. Additionally, low control precision and the long adjusting time may lead to irreversible damage to the myocardial tissue. To solve this problem, an automatic heart perfusion system was developed by applying a Human-Machine Interface (HMI) and a Programmable-Logic-Controller (PLC)-based circuit to control PF and PP. The PLC-based control system collects the data of PF and PP through flow probes and pressure transducers. It has two control modes: the RPM-flow mode and the pressure mode. The RPM-flow control mode is an open-loop system. It influences PF through providing and maintaining the desired speed inputted through the HMI to the centrifugal pump with a maximum error of 20 rpm. The pressure control mode is a closed-loop system where the operator selects a target Mean Arterial Pressure (MAP) to control PP. The inputs of the pressure control mode are the target MAP, received through the HMI, and the real MAP, received from the pressure transducer. A PID algorithm is applied to maintain the real MAP at the target value with a maximum error of 1mmHg. The precision and control speed of the RPM-flow control mode were examined by comparing the PLC-based system to an experienced operator (EO) across seven RPM adjustment ranges (500, 1000, 2000 and random RPM changes; 8 trials per range) tested in a random order. System’s PID algorithm performance in pressure control was assessed during 10 EVHP experiments using porcine hearts. Precision was examined through monitoring the steady-state pressure error throughout perfusion period, and stabilizing speed was tested by performing two MAP adjustment changes (4 trials per change) of 15 and 20mmHg. A total of 56 trials were performed to validate the RPM-flow control mode. Overall, the PLC-based system demonstrated the significantly faster speed than the EO in all trials (PLC 1.21±0.03, EO 3.69±0.23 seconds; p < 0.001) and greater precision to reach the desired RPM (PLC 10±0.7, EO 33±2.7 mean RPM error; p < 0.001). Regarding pressure control, the PLC-based system has the median precision of ±1mmHg error and the median stabilizing times in changing 15 and 20mmHg of MAP are 15 and 19.5 seconds respectively. The novel PLC-based control system was 3 times faster with 60% less error than the EO for RPM-flow control. In pressure control mode, it demonstrates a high precision and fast stabilizing speed. In summary, this novel system successfully controlled perfusion flow and pressure with high precision, stability and a fast response time through a user-friendly interface. This design may provide a viable technique for future development of novel heart preservation and assessment strategies during EVHP.Keywords: automatic control system, biomedical engineering, ex-vivo heart perfusion, human-machine interface, programmable logic controller
Procedia PDF Downloads 1752833 Urban Big Data: An Experimental Approach to Building-Value Estimation Using Web-Based Data
Authors: Sun-Young Jang, Sung-Ah Kim, Dongyoun Shin
Abstract:
Current real-estate value estimation, difficult for laymen, usually is performed by specialists. This paper presents an automated estimation process based on big data and machine-learning technology that calculates influences of building conditions on real-estate price measurement. The present study analyzed actual building sales sample data for Nonhyeon-dong, Gangnam-gu, Seoul, Korea, measuring the major influencing factors among the various building conditions. Further to that analysis, a prediction model was established and applied using RapidMiner Studio, a graphical user interface (GUI)-based tool for derivation of machine-learning prototypes. The prediction model is formulated by reference to previous examples. When new examples are applied, it analyses and predicts accordingly. The analysis process discerns the crucial factors effecting price increases by calculation of weighted values. The model was verified, and its accuracy determined, by comparing its predicted values with actual price increases.Keywords: apartment complex, big data, life-cycle building value analysis, machine learning
Procedia PDF Downloads 3742832 The Use of Artificial Intelligence in Language Learning and Teaching: A New Frontier in Education
Authors: Abdulaziz Fageeh
Abstract:
This study investigates the integration of artificial intelligence (AI) within the landscape of language learning and teaching, exploring its potential benefits and challenges. Employing a mixed-methods approach, the research draws upon a comprehensive literature review, case studies, user reviews, and in-depth interviews with educators and students. Findings demonstrate that AI tools, including language learning apps and writing assistants, can enhance personalization, improve writing skills, and increase accessibility to language learning resources. However, the study also highlights concerns regarding over-reliance on AI, potential accuracy and reliability issues, and ethical implications such as data privacy and potential bias. User and educator perspectives emphasize the importance of balancing AI with traditional teaching methods, fostering critical thinking skills, and addressing potential misuse. The study concludes by underscoring the need for ongoing research and development to ensure responsible AI integration in language learning, focusing on pedagogical strategies, ethical frameworks, and the long-term impact of AI on learning outcomes.Keywords: artificial intelligence, language learning, education, technology, ethical considerations, user perceptions
Procedia PDF Downloads 152831 Quantifying User-Related, System-Related, and Context-Related Patterns of Smartphone Use
Authors: Andrew T. Hendrickson, Liven De Marez, Marijn Martens, Gytha Muller, Tudor Paisa, Koen Ponnet, Catherine Schweizer, Megan Van Meer, Mariek Vanden Abeele
Abstract:
Quantifying and understanding the myriad ways people use their phones and how that impacts their relationships, cognitive abilities, mental health, and well-being is increasingly important in our phone-centric society. However, most studies on the patterns of phone use have focused on theory-driven tests of specific usage hypotheses using self-report questionnaires or analyses of smaller datasets. In this work we present a series of analyses from a large corpus of over 3000 users that combine data-driven and theory-driven analyses to identify reliable smartphone usage patterns and clusters of similar users. Furthermore, we compare the stability of user clusters across user- and system-initiated sessions, as well as during the hypothesized ritualized behavior times directly before and after sleeping. Our results indicate support for some hypothesized usage patterns but present a more complete and nuanced view of how people use smartphones.Keywords: data mining, experience sampling, smartphone usage, health and well being
Procedia PDF Downloads 1632830 Surface Tension and Bulk Density of Ammonium Nitrate Solutions: A Molecular Dynamics Study
Authors: Sara Mosallanejad, Bogdan Z. Dlugogorski, Jeff Gore, Mohammednoor Altarawneh
Abstract:
Ammonium nitrate (NH₄NO₃, AN) is commonly used as the main component of AN emulsion and fuel oil (ANFO) explosives, that use extensively in civilian and mining operations for underground development and tunneling applications. The emulsion formulation and wettability of AN prills, which affect the physical stability and detonation of ANFO, highly depend on the surface tension, density, viscosity of the used liquid. Therefore, for engineering applications of this material, the determination of density and surface tension of concentrated aqueous solutions of AN is essential. The molecular dynamics (MD) simulation method have been used to investigate the density and the surface tension of high concentrated ammonium nitrate solutions; up to its solubility limit in water. Non-polarisable models for water and ions have carried out the simulations, and the electronic continuum correction model (ECC) uses a scaling of the charges of the ions to apply the polarisation implicitly into the non-polarisable model. The results of calculated density and the surface tension of the solutions have been compared to available experimental values. Our MD simulations show that the non-polarisable model with full-charge ions overestimates the experimental results while the reduce-charge model for the ions fits very well with the experimental data. Ions in the solutions show repulsion from the interface using the non-polarisable force fields. However, when charges of the ions in the original model are scaled in line with the scaling factor of the ECC model, the ions create a double ionic layer near the interface by the migration of anions toward the interface while cations stay in the bulk of the solutions. Similar ions orientations near the interface were observed when polarisable models were used in simulations. In conclusion, applying the ECC model to the non-polarisable force field yields the density and surface tension of the AN solutions with high accuracy in comparison to the experimental measurements.Keywords: ammonium nitrate, electronic continuum correction, non-polarisable force field, surface tension
Procedia PDF Downloads 2302829 A Rapid Code Acquisition Scheme in OOC-Based CDMA Systems
Authors: Keunhong Chae, Seokho Yoon
Abstract:
We propose a code acquisition scheme called improved multiple-shift (IMS) for optical code division multiple access systems, where the optical orthogonal code is used instead of the pseudo noise code. Although the IMS algorithm has a similar process to that of the conventional MS algorithm, it has a better code acquisition performance than the conventional MS algorithm. We analyze the code acquisition performance of the IMS algorithm and compare the code acquisition performances of the MS and the IMS algorithms in single-user and multi-user environments.Keywords: code acquisition, optical CDMA, optical orthogonal code, serial algorithm
Procedia PDF Downloads 5392828 Indoor Visible Light Communication Channel Characterization for User Mobility: A Use-Case Study
Authors: Pooja Sanathkumar, Srinidhi Murali, Sethuraman TV, Saravanan M, Paventhan Arumugam, Ashwin Ashok
Abstract:
The last decade has witnessed a significant interest in visible light communication (VLC) technology, as VLC can potentially achieve high data rate links and secure communication channels. However, the use of VLC under mobile settings is fundamentally limited as its a line-of-sight (LOS) technology and there has been limited breakthroughs in realizing VLC for mobile settings. In this regard, this work targets to study the VLC channel under mobility. Through a use-case study analysis with experiment data traces this paper presents an empirical VLC channel study considering the application of VLC for smart lighting in an indoor room environment. This paper contributes a calibration study of a prototype VLC smart lighting system in an indoor environment and through the inferences gained from the calibration, and considering a user is carrying a mobile device fit with a VLC receiver, this work presents recommendations for user's position adjustments, with the goal to ensure maximum connectivity across the room.Keywords: visible light communication, mobility, empirical study, channel characterization
Procedia PDF Downloads 1262827 Value in Exchange: The Importance of Users Interaction as the Center of User Experiences
Authors: Ramlan Jantan, Norfadilah Kamaruddin, Shahriman Zainal Abidin
Abstract:
In this era of technology, the co-creation method has become a new development trend. In this light, most design businesses have currently transformed their development strategy from being goods-dominant into service-dominant where more attention is given to the end-users and their roles in the development process. As a result, the conventional development process has been replaced with a more cooperative one. Consequently, numerous studies have been conducted to explore the extension of co-creation method in the design development process and most studies have focused on issues found during the production process. In the meantime, this study aims to investigate potential values established during the pre-production process, which is also known as the ‘circumstances value creation’. User involvement is questioned and crucially debate at the entry level of pre-production process in value in-exchange jointly spheres; thus user experiences took place. Thus, this paper proposed a potential framework of the co-creation method for Malaysian interactive product development. The framework is formulated from both parties involved: the users and designers. The framework will clearly give an explanation of the value of the co-creation method, and it could assist relevant design industries/companies in developing a blueprint for the design process. This paper further contributes to the literature on the co-creation of value and digital ecosystems.Keywords: co-creation method, co-creation framework, co-creation, co-production
Procedia PDF Downloads 1782826 Synchronization of Bus Frames during Universal Serial Bus Transfer
Authors: Petr Šimek
Abstract:
This work deals with the problem of synchronization of bus frames during transmission using USB (Universal Serial Bus). The principles for synchronization between USB and the non-deterministic CAN (Controller Area Network) bus will be described here. Furthermore, the work deals with ensuring the time sequence of communication frames when receiving from multiple communication bus channels. The structure of a general object for storing frames from different types of communication buses, such as CAN and LIN (Local Interconnect Network), will be described here. Finally, an evaluation of the communication throughput of bus frames for USB High speed will be performed. The creation of this architecture was based on the analysis of the communication of control units with a large number of communication buses. For the design of the architecture, a test HW with a USB-HS interface was used, which received previously known messages, which were compared with the received result. The result of this investigation is the block architecture of the control program for test HW ensuring correct data transmission via the USB bus.Keywords: analysis, CAN, interface, LIN, synchronization, USB
Procedia PDF Downloads 622825 Improving Monitoring and Fault Detection of Solar Panels Using Arduino Mega in WSN
Authors: Ali Al-Dahoud, Mohamed Fezari, Thamer Al-Rawashdeh, Ismail Jannoud
Abstract:
Monitoring and detecting faults on a set of Solar panels, using a wireless sensor network (WNS) is our contribution in this paper, This work is part of the project we are working on at Al-Zaytoonah University. The research problem has been exposed by engineers and technicians or operators dealing with PV panels maintenance, in order to monitor and detect faults within solar panels which affect considerably the energy produced by the solar panels. The proposed solution is based on installing WSN nodes with appropriate sensors for more often occurred faults on the 45 solar panels installed on the roof of IT faculty. A simulation has been done on nodes distribution and a study for the design of a node with appropriate sensors taking into account the priorities of the processing faults. Finally, a graphic user interface is designed and adapted to telemonitoring panels using WSN. The primary tests of hardware implementation gave interesting results, the sensors calibration and interference transmission problem have been solved. A friendly GUI using high level language Visial Basic was developed to carry out the monitoring process and to save data on Exel File.Keywords: Arduino Mega microcnotroller, solar panels, fault-detection, simulation, node design
Procedia PDF Downloads 4652824 User-Awareness from Eye Line Tracing During Specification Writing to Improve Specification Quality
Authors: Yoshinori Wakatake
Abstract:
Many defects after the release of software packages are caused due to omissions of sufficient test items in test specifications. Poor test specifications are detected by manual review, which imposes a high human load. The prevention of omissions depends on the end-user awareness of test specification writers. If test specifications were written while envisioning the behavior of end-users, the number of omissions in test items would be greatly reduced. The paper pays attention to the point that writers who can achieve it differ from those who cannot in not only the description richness but also their gaze information. It proposes a method to estimate the degree of user-awareness of writers through the analysis of their gaze information when writing test specifications. We conduct an experiment to obtain the gaze information of a writer of the test specifications. Test specifications are automatically classified using gaze information. In this method, a Random Forest model is constructed for the classification. The classification is highly accurate. By looking at the explanatory variables which turn out to be important variables, we know behavioral features to distinguish test specifications of high quality from others. It is confirmed they are pupil diameter size and the number and the duration of blinks. The paper also investigates test specifications automatically classified with gaze information to discuss features in their writing ways in each quality level. The proposed method enables us to automatically classify test specifications. It also prevents test item omissions, because it reveals writing features that test specifications of high quality should satisfy.Keywords: blink, eye tracking, gaze information, pupil diameter, quality improvement, specification document, user-awareness
Procedia PDF Downloads 642823 Sentinel-2 Based Burn Area Severity Assessment Tool in Google Earth Engine
Authors: D. Madhushanka, Y. Liu, H. C. Fernando
Abstract:
Fires are one of the foremost factors of land surface disturbance in diverse ecosystems, causing soil erosion and land-cover changes and atmospheric effects affecting people's lives and properties. Generally, the severity of the fire is calculated as the Normalized Burn Ratio (NBR) index. This is performed manually by comparing two images obtained afterward. Then by using the bitemporal difference of the preprocessed satellite images, the dNBR is calculated. The burnt area is then classified as either unburnt (dNBR<0.1) or burnt (dNBR>= 0.1). Furthermore, Wildfire Severity Assessment (WSA) classifies burnt areas and unburnt areas using classification levels proposed by USGS and comprises seven classes. This procedure generates a burn severity report for the area chosen by the user manually. This study is carried out with the objective of producing an automated tool for the above-mentioned process, namely the World Wildfire Severity Assessment Tool (WWSAT). It is implemented in Google Earth Engine (GEE), which is a free cloud-computing platform for satellite data processing, with several data catalogs at different resolutions (notably Landsat, Sentinel-2, and MODIS) and planetary-scale analysis capabilities. Sentinel-2 MSI is chosen to obtain regular processes related to burnt area severity mapping using a medium spatial resolution sensor (15m). This tool uses machine learning classification techniques to identify burnt areas using NBR and to classify their severity over the user-selected extent and period automatically. Cloud coverage is one of the biggest concerns when fire severity mapping is performed. In WWSAT based on GEE, we present a fully automatic workflow to aggregate cloud-free Sentinel-2 images for both pre-fire and post-fire image compositing. The parallel processing capabilities and preloaded geospatial datasets of GEE facilitated the production of this tool. This tool consists of a Graphical User Interface (GUI) to make it user-friendly. The advantage of this tool is the ability to obtain burn area severity over a large extent and more extended temporal periods. Two case studies were carried out to demonstrate the performance of this tool. The Blue Mountain national park forest affected by the Australian fire season between 2019 and 2020 is used to describe the workflow of the WWSAT. This site detected more than 7809 km2, using Sentinel-2 data, giving an error below 6.5% when compared with the area detected on the field. Furthermore, 86.77% of the detected area was recognized as fully burnt out, of which high severity (17.29%), moderate-high severity (19.63%), moderate-low severity (22.35%), and low severity (27.51%). The Arapaho and Roosevelt National Forest Park, California, the USA, which is affected by the Cameron peak fire in 2020, is chosen for the second case study. It was found that around 983 km2 had burned out, of which high severity (2.73%), moderate-high severity (1.57%), moderate-low severity (1.18%), and low severity (5.45%). These spots also can be detected through the visual inspection made possible by cloud-free images generated by WWSAT. This tool is cost-effective in calculating the burnt area since satellite images are free and the cost of field surveys is avoided.Keywords: burnt area, burnt severity, fires, google earth engine (GEE), sentinel-2
Procedia PDF Downloads 235