Search results for: incidental information processing
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 13664

Search results for: incidental information processing

11894 Information System Management Factors Related to Behavioral Trend of Online Accommodation Services

Authors: Supattra Kanchanopast

Abstract:

The purpose of this research was to study the customers’ behavioral trend for online accommodation system at Bangkonthi District, Samutsongkhram province. The research collected data from 400 online users. A questionnaire was utilized as the tool in collecting information. Descriptive statistics included frequency, percentage, mean and standard deviation. Independent- sample t- test, analysis of variance and Pearson Correlation were also used. The findings of this research revealed that the majority of the respondents were male, 25-32 years old, and graduated a bachelor degree. The respondents mostly worked in private sectors and had monthly income between 10,001-15,000 baht. The regular online users, visiting this system between 3-4 times/month, spending 1-2 hours/time, searched for online accommodation information. This result showed that the users had good and high attitude towards the system. According to the hypothesis testing, the number of online usage had positive related to the behavioral trends: accommodation purchasing intention and recommend the accommodation to others. Furthermore, both the number of online usage and overall attitude had a significant correlation to accommodation purchase intention and recommend the accommodation to others.

Keywords: customer behavior, information system management, online accommodation services, behavioral trend

Procedia PDF Downloads 243
11893 The Management Information System for Convenience Stores: Case Study in 7 Eleven Shop in Bangkok

Authors: Supattra Kanchanopast

Abstract:

The purpose of this research is to develop and design a management information system for 7 eleven shop in Bangkok. The system was designed and developed to meet users’ requirements via the internet network by use of application software such as My SQL for database management, Apache HTTP Server for Web Server and PHP Hypertext Preprocessor for an interface between web server, database and users. The system was designed into two subsystems as the main system, or system for head office, and the branch system for branch shops. These consisted of three parts which are classified by user management as shop management, inventory management and Point of Sale (POS) management. The implementation of the MIS for the mini-mart shop, can lessen the amount of paperwork and reduce repeating tasks so it may decrease the capital of the business and support an extension of branches in the future as well.

Keywords: convenience store, the management information system, inventory management, 7 eleven shop

Procedia PDF Downloads 484
11892 Procedure for Recommendation of Archival Documents

Authors: Marlon J. Remedios, Maria T. Morell, Jesse D. Cano

Abstract:

Diffusion and accessibility of historical collections is one of the main objectives of the institutions that aim to safeguard archival documents (General Archives). Several countries have Web applications that try to make accessible and public the large number of documents that they guard. Each of these sites has a set of features in order to facilitate access, navigability, and search for information. Different sources of information include Recommender Systems as a way of customizing content. This paper aims at describing a process for the production of archival documents relevant to the user. To comply with this, the characteristics ruling archival description, elements and main techniques that establishes the design of Recommender Systems, a set of rules to follow, and how these rules operate and the way in which take advantage of the domain knowledge are discussed. Finally, relevant issues are discussed in the design of the proposed tests and the results obtained are shown.

Keywords: archival document, recommender system, procedure, information management

Procedia PDF Downloads 516
11891 Deployment of Matrix Transpose in Digital Image Encryption

Authors: Okike Benjamin, Garba E J. D.

Abstract:

Encryption is used to conceal information from prying eyes. Presently, information and data encryption are common due to the volume of data and information in transit across the globe on daily basis. Image encryption is yet to receive the attention of the researchers as deserved. In other words, video and multimedia documents are exposed to unauthorized accessors. The authors propose image encryption using matrix transpose. An algorithm that would allow image encryption is developed. In this proposed image encryption technique, the image to be encrypted is split into parts based on the image size. Each part is encrypted separately using matrix transpose. The actual encryption is on the picture elements (pixel) that make up the image. After encrypting each part of the image, the positions of the encrypted images are swapped before transmission of the image can take place. Swapping the positions of the images is carried out to make the encrypted image more robust for any cryptanalyst to decrypt.

Keywords: image encryption, matrices, pixel, matrix transpose

Procedia PDF Downloads 421
11890 Harmonization of Financial Information Systems in Latin America in Light of International Public Sector Accounting Standards Using the Herfindahl-Hirschman Index

Authors: Laura Sour

Abstract:

Government accounting is an essential instrument of transparency and accountability in public administration, which allows connecting internal management with the implementation of policies and their evaluation by third parties through the construction of indicators on the cost of government. Several countries have adopted the International Public Sector Accounting Standards (IPSAS) as part of their modernization strategy. This document will evaluate the quantity and harmonization of the financial information published in the financial statements of 12 Latin American countries based on what is established in IPSAS 1, 2 and 17. For this, seven types of financial statements are analyzed. published during the period from 2015 to 2019. Based on this information, it will be possible to describe the evolution in the government financial publication to carry out a detailed analysis of the items that have been most transparent in these countries. Finally, the level of harmonization of the financial statements will be studied using the Herfindahl-Hirschman index (IHH) to determine the degree of comparability of the information. To date, the results indicate that the public sector has increased the quantity and harmonization of the financial information published during the study period, but in a heterogeneous way: From the data collected, it has been found that the financial statement published with greater frequency and quantity is the Income Statement (classification of expenses by nature). On the other hand, the most complete reports were published by Costa Rica (2017 to 2019) and Mexico (2016 to 2018), periods during which these countries complied with 92.9 percent of the items analyzed. Although 2017 and 2018 are the years in which the most financial statements were reported, it is important to mention that Mexico is the country that has published the most financial information throughout the entire study period. The use of the IHH is expected to provide accurate information on the quality with which countries have adopted IPSAS within their government accounting systems to promote transparency and accountability in the continent.

Keywords: accounting and auditing, government policy and regulation, harmonization, public sector accounting and audits IPSAS

Procedia PDF Downloads 91
11889 Neuropsychological Aspects in Adolescents Victims of Sexual Violence with Post-Traumatic Stress Disorder

Authors: Fernanda Mary R. G. Da Silva, Adriana C. F. Mozzambani, Marcelo F. Mello

Abstract:

Introduction: Sexual assault against children and adolescents is a public health problem with serious consequences on their quality of life, especially for those who develop post-traumatic stress disorder (PTSD). The broad literature in this research area points to greater losses in verbal learning, explicit memory, speed of information processing, attention and executive functioning in PTSD. Objective: To compare the neuropsychological functions of adolescents from 14 to 17 years of age, victims of sexual violence with PTSD with those of healthy controls. Methodology: Application of a neuropsychological battery composed of the following subtests: WASI vocabulary and matrix reasoning; Digit subtests (WISC-IV); verbal auditory learning test RAVLT; Spatial Span subtest of the WMS - III scale; abbreviated version of the Wisconsin test; concentrated attention test - D2; prospective memory subtest of the NEUPSILIN scale; five-digit test - FDT and the Stroop test (Trenerry version) in adolescents with a history of sexual violence in the previous six months, referred to the Prove (Violence Care and Research Program of the Federal University of São Paulo), for further treatment. Results: The results showed a deficit in the word coding process in the RAVLT test, with impairment in A3 (p = 0.004) and A4 (p = 0.016) measures, which compromises the verbal learning process (p = 0.010) and the verbal recognition memory (p = 0.012), seeming to present a worse performance in the acquisition of verbal information that depends on the support of the attentional system. A worse performance was found in list B (p = 0.047), a lower priming effect p = 0.026, that is, lower evocation index of the initial words presented and less perseveration (p = 0.002), repeated words. Therefore, there seems to be a failure in the creation of strategies that help the mnemonic process of retention of the verbal information necessary for learning. Sustained attention was found to be impaired, with greater loss of setting in the Wisconsin test (p = 0.023), a lower rate of correct responses in stage C of the Stroop test (p = 0.023) and, consequently, a higher index of erroneous responses in C of the Stroop test (p = 0.023), besides more type II errors in the D2 test (p = 0.008). A higher incidence of total errors was observed in the reading stage of the FDT test p = 0.002, which suggests fatigue in the execution of the task. Performance is compromised in executive functions in the cognitive flexibility ability, suggesting a higher index of total errors in the alternating step of the FDT test (p = 0.009), as well as a greater number of persevering errors in the Wisconsin test (p = 0.004). Conclusion: The data from this study suggest that sexual violence and PTSD cause significant impairment in the neuropsychological functions of adolescents, evidencing risk to quality of life in stages that are fundamental for the development of learning and cognition.

Keywords: adolescents, neuropsychological functions, PTSD, sexual violence

Procedia PDF Downloads 135
11888 An Integrated Framework for Engaging Stakeholders in the Circular Economy Processes Using Building Information Modeling and Virtual Reality

Authors: Erisasadat Sahebzamani, Núria Forcada, Francisco Lendinez

Abstract:

Global climate change has become increasingly problematic over the past few decades. The construction industry has contributed to greenhouse gas emissions in recent decades. Considering these issues and the high demand for materials in the construction industry, Circular Economy (CE) is considered necessary to keep materials in the loop and extend their useful lives. By providing tangible benefits, Construction 4.0 facilitates the adoption of CE by reducing waste, updating standard work, sharing knowledge, and increasing transparency and stability. This study aims to present a framework for integrating CE and digital tools like Building Information Modeling (BIM) and Virtual Reality (VR) to examine the impact on the construction industry based on stakeholders' perspectives.

Keywords: circular economy, building information modeling, virtual reality, stakeholder engagement

Procedia PDF Downloads 111
11887 Integrating Nursing Informatics to Improve Patient-Centered Care: A Project to Reduce Patient Waiting Time at the Blood Pressure Counter

Authors: Pi-Chi Wu, Tsui-Ping Chu, Hsiu-Hung Wang

Abstract:

Background: The ability to provide immediate medical service in outpatient departments is one of the keys to patient satisfaction. Objectives: This project used electronic equipment to integrate nursing care information to patient care at a blood pressure diagnostic counter. Through process reengineering, the average patient waiting time decreased from 35 minutes to 5 minutes, while service satisfaction increased from a score of 2.7 to 4.6. Methods: Data was collected from a local hospital in Southern Taiwan from a daily average of 2,200 patients in the outpatient department. Previous waiting times were affected by (1) space limitations, (2) the need to help guide patient mobility, (3) the need for nurses to appease irate patients and give instructions, (4), the need for patients to replace lost counter tickets, (5) the need to re-enter information, (6) the replacement of missing patient information. An ad hoc group was established to enhance patient satisfaction and shorten waiting times for patients to see a doctor. A four step strategy consisting of (1) counter relocation, (2) queue reorganization, (3) electronic information integration, (4) process reengineering was implemented. Results: Implementation of the developed strategy decreased patient waiting time from 35 minutes to an average of 5 minutes, and increased patient satisfaction scores from 2.7 to 6.4. Conclusion: Through the integration of information technology and process transformation, waiting times were drastically reduced, patient satisfaction increased, and nurses were allowed more time to engage in more cost-effective services. This strategy was simultaneously enacted in separate hospitals throughout Taiwan.

Keywords: process reengineering, electronic information integration, patient satisfaction, patient waiting time

Procedia PDF Downloads 378
11886 A Review of Spatial Analysis as a Geographic Information Management Tool

Authors: Chidiebere C. Agoha, Armstong C. Awuzie, Chukwuebuka N. Onwubuariri, Joy O. Njoku

Abstract:

Spatial analysis is a field of study that utilizes geographic or spatial information to understand and analyze patterns, relationships, and trends in data. It is characterized by the use of geographic or spatial information, which allows for the analysis of data in the context of its location and surroundings. It is different from non-spatial or aspatial techniques, which do not consider the geographic context and may not provide as complete of an understanding of the data. Spatial analysis is applied in a variety of fields, which includes urban planning, environmental science, geosciences, epidemiology, marketing, to gain insights and make decisions about complex spatial problems. This review paper explores definitions of spatial analysis from various sources, including examples of its application and different analysis techniques such as Buffer analysis, interpolation, and Kernel density analysis (multi-distance spatial cluster analysis). It also contrasts spatial analysis with non-spatial analysis.

Keywords: aspatial technique, buffer analysis, epidemiology, interpolation

Procedia PDF Downloads 319
11885 Design of Personal Job Recommendation Framework on Smartphone Platform

Authors: Chayaporn Kaensar

Abstract:

Recently, Job Recommender Systems have gained much attention in industries since they solve the problem of information overload on the recruiting website. Therefore, we proposed Extended Personalized Job System that has the capability of providing the appropriate jobs for job seeker and recommending some suitable information for them using Data Mining Techniques and Dynamic User Profile. On the other hands, company can also interact to the system for publishing and updating job information. This system have emerged and supported various platforms such as web application and android mobile application. In this paper, User profiles, Implicit User Action, User Feedback, and Clustering Techniques in WEKA libraries have gained attention and implemented for this application. In additions, open source tools like Yii Web Application Framework, Bootstrap Front End Framework and Android Mobile Technology were also applied.

Keywords: recommendation, user profile, data mining, web and mobile technology

Procedia PDF Downloads 313
11884 The Causes and Potential Solutions for Foodborne Illness, Food Security, and Food Safety: In the Case of the East Harerghe Region of Oromia, Ethiopia

Authors: Tuji Jemal Ahmed, Abdi Mohammed, Geremew Geidare Kailo

Abstract:

Food security, foodborne illness, and food safety are critical issues that affect the East Harerghe region of Oromia, Ethiopia. Despite the region's potential for agriculture, food insecurity remains a significant problem, with many households experiencing chronic hunger and malnutrition. The region also experiences high rates of foodborne illnesses, including cholera, typhoid, and diarrhea, which are caused by poor hygiene and sanitation practices. Additionally, food safety is a significant challenge, particularly in rural areas, where there is a lack of infrastructure, inadequate food storage facilities, and limited access to information about food safety. There are several factors that contribute to the current situation in the East Harerghe region; firstly, the region is susceptible to natural disasters, for instance, drought, which affects crop yields and livestock production. Secondly, the region also experiences poor infrastructure, which affects the storage and transportation of food, particularly in rural areas. Thirdly, there is a lack of awareness and knowledge on good hygiene and sanitation practices, specifically during food handling, processing, and storage. Fourthly, unitability due to conflict and other forms of land degradation exacerbates food insecurity and malnutrition. Finally, limited access to financial resources and markets commonly affects smallholder farmers by their ability to produce and sell food. To address the current situation in that area, several potential solutions can be implemented; investment in infrastructure is necessary, especially in rural areas, to improve the storage and transportation of food. Education and awareness programs on good hygiene and sanitation practices should target local communities, smallholder farmers, and food vendors. Financial resources and markets should be made more accessible to smallholder farmers, particularly through the provision of credit and improved access to markets. Addressing the underlying causes of conflict and promoting peaceful coexistence can help to reduce displacement and loss of livelihoods. Finally, the enforcement of food safety regulations and the implementation of standards for food processing and storage facilities are necessary to ensure food safety. In conclusion, addressing the challenges of food security, foodborne illness, and food safety in the East Harerghe region requires a coordinated effort from various stakeholders, including the government, non-governmental organizations, and local communities. By implementing the solutions outlined above, the region can improve its food security, prevent foodborne illnesses, and keep food safe for its population. Eventually, building the resilience of communities to shocks such as droughts, floods, and conflict is necessary to ensure long-term food security in the region.

Keywords: foodborne illness, food handling, food safety, food security

Procedia PDF Downloads 101
11883 Geographic Information System for District Level Energy Performance Simulations

Authors: Avichal Malhotra, Jerome Frisch, Christoph van Treeck

Abstract:

The utilization of semantic, cadastral and topological data from geographic information systems (GIS) has exponentially increased for building and urban-scale energy performance simulations. Urban planners, simulation scientists, and researchers use virtual 3D city models for energy analysis, algorithms and simulation tools. For dynamic energy simulations at city and district level, this paper provides an overview of the available GIS data models and their levels of detail. Adhering to different norms and standards, these models also intend to describe building and construction industry data. For further investigations, CityGML data models are considered for simulations. Though geographical information modelling has considerably many different implementations, extensions of virtual city data can also be made for domain specific applications. Highlighting the use of the extended CityGML models for energy researches, a brief introduction to the Energy Application Domain Extension (ADE) along with its significance is made. Consequently, addressing specific input simulation data, a workflow using Modelica underlining the usage of GIS information and the quantification of its significance over annual heating energy demand is presented in this paper.

Keywords: CityGML, EnergyADE, energy performance simulation, GIS

Procedia PDF Downloads 169
11882 Context Detection in Spreadsheets Based on Automatically Inferred Table Schema

Authors: Alexander Wachtel, Michael T. Franzen, Walter F. Tichy

Abstract:

Programming requires years of training. With natural language and end user development methods, programming could become available to everyone. It enables end users to program their own devices and extend the functionality of the existing system without any knowledge of programming languages. In this paper, we describe an Interactive Spreadsheet Processing Module (ISPM), a natural language interface to spreadsheets that allows users to address ranges within the spreadsheet based on inferred table schema. Using the ISPM, end users are able to search for values in the schema of the table and to address the data in spreadsheets implicitly. Furthermore, it enables them to select and sort the spreadsheet data by using natural language. ISPM uses a machine learning technique to automatically infer areas within a spreadsheet, including different kinds of headers and data ranges. Since ranges can be identified from natural language queries, the end users can query the data using natural language. During the evaluation 12 undergraduate students were asked to perform operations (sum, sort, group and select) using the system and also Excel without ISPM interface, and the time taken for task completion was compared across the two systems. Only for the selection task did users take less time in Excel (since they directly selected the cells using the mouse) than in ISPM, by using natural language for end user software engineering, to overcome the present bottleneck of professional developers.

Keywords: natural language processing, natural language interfaces, human computer interaction, end user development, dialog systems, data recognition, spreadsheet

Procedia PDF Downloads 311
11881 Studying the Effect of Reducing Thermal Processing over the Bioactive Composition of Non-Centrifugal Cane Sugar: Towards Natural Products with High Therapeutic Value

Authors: Laura Rueda-Gensini, Jader Rodríguez, Juan C. Cruz, Carolina Munoz-Camargo

Abstract:

There is an emerging interest in botanicals and plant extracts for medicinal practices due to their widely reported health benefits. A large variety of phytochemicals found in plants have been correlated with antioxidant, immunomodulatory, and analgesic properties, which makes plant-derived products promising candidates for modulating the progression and treatment of numerous diseases. Non-centrifugal cane sugar (NCS), in particular, has been known for its high antioxidant and nutritional value, but composition-wise variability due to changing environmental and processing conditions have considerably limited its use in the nutraceutical and biomedical fields. This work is therefore aimed at assessing the effect of thermal exposure during NCS production over its bioactive composition and, in turn, its therapeutic value. Accordingly, two modified dehydration methods are proposed that employ: (i) vacuum-aided evaporation, which reduces the necessary temperatures to dehydrate the sample, and (ii) window refractance evaporation, which reduces thermal exposure time. The biochemical composition of NCS produced under these two methods was compared to traditionally-produced NCS by estimating their total polyphenolic and protein content with Folin-Ciocalteu and Bradford assays, as well as identifying the major phenolic compounds in each sample via HPLC-coupled mass spectrometry. Their antioxidant activities were also compared as measured by their scavenging potential of ABTS and DPPH radicals. Results show that the two modified production methods enhance polyphenolic and protein yield in resulting NCS samples when compared to traditional production methods. In particular, reducing employed temperatures with vacuum-aided evaporation demonstrated to be superior at preserving polyphenolic compounds, as evidenced both in the total and individual polyphenol concentrations. However, antioxidant activities were not significantly different between these. Although additional studies should be performed to determine if the observed compositional differences affect other therapeutic activities (e.g., anti-inflammatory, analgesic, and immunoprotective), these results suggest that reducing thermal exposure holds great promise for the production of natural products with enhanced nutritional value.

Keywords: non-centrifugal cane sugar, polyphenolic compounds, thermal processing, antioxidant activity

Procedia PDF Downloads 92
11880 Enhancing Temporal Extrapolation of Wind Speed Using a Hybrid Technique: A Case Study in West Coast of Denmark

Authors: B. Elshafei, X. Mao

Abstract:

The demand for renewable energy is significantly increasing, major investments are being supplied to the wind power generation industry as a leading source of clean energy. The wind energy sector is entirely dependable and driven by the prediction of wind speed, which by the nature of wind is very stochastic and widely random. This s0tudy employs deep multi-fidelity Gaussian process regression, used to predict wind speeds for medium term time horizons. Data of the RUNE experiment in the west coast of Denmark were provided by the Technical University of Denmark, which represent the wind speed across the study area from the period between December 2015 and March 2016. The study aims to investigate the effect of pre-processing the data by denoising the signal using empirical wavelet transform (EWT) and engaging the vector components of wind speed to increase the number of input data layers for data fusion using deep multi-fidelity Gaussian process regression (GPR). The outcomes were compared using root mean square error (RMSE) and the results demonstrated a significant increase in the accuracy of predictions which demonstrated that using vector components of the wind speed as additional predictors exhibits more accurate predictions than strategies that ignore them, reflecting the importance of the inclusion of all sub data and pre-processing signals for wind speed forecasting models.

Keywords: data fusion, Gaussian process regression, signal denoise, temporal extrapolation

Procedia PDF Downloads 136
11879 Robust Variable Selection Based on Schwarz Information Criterion for Linear Regression Models

Authors: Shokrya Saleh A. Alshqaq, Abdullah Ali H. Ahmadini

Abstract:

The Schwarz information criterion (SIC) is a popular tool for selecting the best variables in regression datasets. However, SIC is defined using an unbounded estimator, namely, the least-squares (LS), which is highly sensitive to outlying observations, especially bad leverage points. A method for robust variable selection based on SIC for linear regression models is thus needed. This study investigates the robustness properties of SIC by deriving its influence function and proposes a robust SIC based on the MM-estimation scale. The aim of this study is to produce a criterion that can effectively select accurate models in the presence of vertical outliers and high leverage points. The advantages of the proposed robust SIC is demonstrated through a simulation study and an analysis of a real dataset.

Keywords: influence function, robust variable selection, robust regression, Schwarz information criterion

Procedia PDF Downloads 141
11878 The Interaction of Country-of-Manufacturing with Country-of-Design within Different Consumption Context

Authors: Ebru Genc, Shih-Ching Wang

Abstract:

In today’s globalized world, while companies move their production centers to developing countries in order to gain cost advantage, they receive negative responses from consumers because of the weak image of those countries. In this study, we looked at this tradeoff faced by multinational companies. Some companies that have headquarters in developed countries have devised a strategy of manipulating country-of-origin (COO) information by introducing the concept of country of design (COD). We analyzed the impact of country-of-manufacturing (COM) information on consumers’ product evaluation and purchase intention in the presence of different levels of COD information, namely, in terms of developed and developing countries. We found that it is not advantageous for a firm to publish a design location with a strong image if the firm is producing in a country that has a weak image. On the other hand, revealing COD information has a reinforcing effect on consumers’ product evaluation and purchase intention if the firm is producing in a country with a strong image. Second, we studied the impact of consumption context on this relationship (in terms of public or private use) and found that for products that are typically used in public, COM has significantly shown higher importance on product evaluation and purchase intention, compared to products typically used in private. However, our results show that consumption context shows no effect of an impact resulting from COD information.

Keywords: consumption context, country of design, country of manufacturing, country of origin

Procedia PDF Downloads 249
11877 Electronic Payment Recording with Payment History Retrieval Module: A System Software

Authors: Adrian Forca, Simeon Cainday III

Abstract:

The Electronic Payment Recording with Payment History Retrieval Module is developed intendedly for the College of Science and Technology. This system software innovates the manual process of recording the payments done in the department through the development of electronic payment recording system software shifting from the slow and time-consuming procedure to quick yet reliable and accurate way of recording payments because it immediately generates receipts for every transaction. As an added feature to its software process, generation of recorded payment report is integrated eliminating the manual reporting to a more easy and consolidated report. As an added feature to the system, all recorded payments of the students can be retrieved immediately making the system transparent and reliable payment recording software. Viewing the whole process, the system software will shift from the manual process to an organized software technology because the information will be stored in a logically correct and normalized database. Further, the software will be developed using the modern programming language and implement strict programming methods to validate all users accessing the system, evaluate all data passed into the system and information retrieved to ensure data accuracy and reliability. In addition, the system will identify the user and limit its access privilege to establish boundaries of the specific access to information allowed for the store, modify, and update making the information secure against unauthorized data manipulation. As a result, the System software will eliminate the manual procedure and replace with an innovative modern information technology resulting to the improvement of the whole process of payment recording fast, secure, accurate and reliable software innovations.

Keywords: collection, information system, manual procedure, payment

Procedia PDF Downloads 167
11876 Experiences of Pediatric Cancer Patients and Their Families: A Focus Group Interview

Authors: Bu Kyung Park

Abstract:

Background: The survival rate of pediatric cancer patients has been increased. Thus, the needs of long-term management and follow-up education after discharge continue to grow. Purpose: The purpose of this study was to explore the experiences of pediatric cancer patients and their families from first diagnosis to returning their social life. The ultimate goal of this study was to assess which information and intervention did pediatric cancer patients and their families required and needed, so that this could provide fundamental information for developing educational content of web-based intervention program for pediatric cancer patients. Research Approach: This study was based on a descriptive qualitative research design using semi-structured focus group interview. Participants: Twelve pediatric cancer patients and 12 family members participated in a total six focus group interview sessions. Methods: All interviews were audiotaped after obtaining participants’ approval. The recordings were transcribed. Qualitative Content analysis using the inductive coding approach was performed on the transcriptions by three coders. Findings: Eighteen categories emerged from the six main themes: 1) Information needs, 2) Support system, 3) Barriers to treatment, 4) Facilitators to treatment, 5) Return to social life, 6) Healthcare system issues. Each theme had both pediatric cancer patients’ codes and their family members’ codes. Patients and family members had high information needs through the whole process of treatment, not only the first diagnosis but also after completion of treatment. Hospitals provided basic information on chemo therapy, medication, and various examinations. However, they were more likely to rely on information from other patients and families by word of mouth. Participants’ information needs were different according to their treatment stage (e.g., first admitted patients versus cancer survivors returning to their social life). Even newly diagnosed patients worried about social adjustment after completion of all treatment, such as return to school and diet and physical activity at home. Most family members had unpleasant experiences while they were admitted in hospitals and concerned about healthcare system issues, such as medical error and patient safety. Conclusions: In conclusion, pediatric cancer patients and their family members wanted information source which can provide tailored information based on their needs. Different information needs with patients and their family members based on their diagnosis, progress, stage of treatment were identified. Findings from this study will be used to develop a patient-centered online health intervention program for pediatric cancer patients. Pediatric cancer patients and their family members had variety fields of education needs and soak the information from various sources. Web-based health intervention program for them is required to satisfy their inquiries to provide reliable information.

Keywords: focus group interview, family caregivers, pediatric cancer patients, qualitative content analysis

Procedia PDF Downloads 181
11875 IT Skills and Soft Skills for Accountants in Thailand

Authors: Manirath Wongsim

Abstract:

Information technology management has become important for the achievement of organisations. An increase in the pace of technological change has revolutionised the way accountants perform their jobs. In response to this challenge, the identification of a new comprehensive set of information technology competencies combined with information technology skills and other skills (namely, soft skills) are necessary. Thus, this study aims to investigate IT competencies among professional accountants to enhance firm performance. This research was conducted with 42 respondents at ten organisations in Thailand. This research used qualitative, interpretive evidence.The results indicate that the factor IT competencies within the organizational issues defines19 factors. Specifically, these new factors, based on the research findings and the literature and unique to IT competences for professional accountants, include ERP software skills, BI software skills and accounting law and legal skills. The evidence in this study suggests that ERP software, spreadsheets, BI software and accounting software were ranked as much-needed skills to be acquired by accountants while communication skills were ranked as the most required skills, and delegation skills as the least required. The findings of the research’s empirical evidence suggest that organizations should understand appropriate into developing information technology related competencies for knowledge workers in general and professional accountants in particular and provide assistance in all processes of decision making.

Keywords: IT competencies, IT competencies for accountants, IT skills for accounting, soft skills for accountants

Procedia PDF Downloads 415
11874 The Relation between Cognitive Fluency and Utterance Fluency in Second Language Spoken Fluency: Studying Fluency through a Psycholinguistic Lens

Authors: Tannistha Dasgupta

Abstract:

This study explores the aspects of second language (L2) spoken fluency that are related to L2 linguistic knowledge and processing skill. It draws on Levelt’s ‘blueprint’ of the L2 speaker which discusses the cognitive issues underlying the act of speaking. However, L2 speaking assessments have largely neglected the underlying mechanism involved in language production; emphasis is given on the relationship between subjective ratings of L2 speech sample and objectively measured aspects of fluency. Hence, in this study, the relation between L2 linguistic knowledge and processing skill i.e. Cognitive Fluency (CF), and objectively measurable aspects of L2 spoken fluency i.e. Utterance Fluency (UF) is examined. The participants of the study are L2 learners of English, studying at high school level in Hyderabad, India. 50 participants with intermediate level of proficiency in English performed several lexical retrieval tasks and attention-shifting tasks to measure CF, and 8 oral tasks to measure UF. Each aspect of UF (speed, pause, and repair) were measured against the scores of CF to find out those aspects of UF which are reliable indicators of CF. Quantitative analysis of the data shows that among the three aspects of UF; speed is the best predictor of CF, and pause is weakly related to CF. The study suggests that including the speed aspect of UF could make L2 fluency assessment more reliable, valid, and objective. Thus, incorporating the assessment of psycholinguistic mechanisms into L2 spoken fluency testing, could result in fairer evaluation.

Keywords: attention-shifting, cognitive fluency, lexical retrieval, utterance fluency

Procedia PDF Downloads 711
11873 Digitalisation of the Railway Industry: Recent Advances in the Field of Dialogue Systems: Systematic Review

Authors: Andrei Nosov

Abstract:

This paper discusses the development directions of dialogue systems within the digitalisation of the railway industry, where technologies based on conversational AI are already potentially applied or will be applied. Conversational AI is one of the popular natural language processing (NLP) tasks, as it has great prospects for real-world applications today. At the same time, it is a challenging task as it involves many areas of NLP based on complex computations and deep insights from linguistics and psychology. In this review, we focus on dialogue systems and their implementation in the railway domain. We comprehensively review the state-of-the-art research results on dialogue systems and analyse them from three perspectives: type of problem to be solved, type of model, and type of system. In particular, from the perspective of the type of tasks to be solved, we discuss characteristics and applications. This will help to understand how to prioritise tasks. In terms of the type of models, we give an overview that will allow researchers to become familiar with how to apply them in dialogue systems. By analysing the types of dialogue systems, we propose an unconventional approach in contrast to colleagues who traditionally contrast goal-oriented dialogue systems with open-domain systems. Our view focuses on considering retrieval and generative approaches. Furthermore, the work comprehensively presents evaluation methods and datasets for dialogue systems in the railway domain to pave the way for future research. Finally, some possible directions for future research are identified based on recent research results.

Keywords: digitalisation, railway, dialogue systems, conversational AI, natural language processing, natural language understanding, natural language generation

Procedia PDF Downloads 63
11872 Creating Inclusive Information Services: Librarians’ Design-Thinking Approach to Helping Students Succeed in the Digital Age

Authors: Yi Ding

Abstract:

With the rapid development of educational technologies, higher education institutions are facing the challenge of creating an inclusive learning environment for students from diverse backgrounds. Academic libraries, the hubs of research, instruction, and innovation at higher educational institutions, are facing the same challenge. While academic librarians worldwide have been working hard to provide services for emerging information technology such as information literacy education, online learning support, and scholarly communication advocacy, the problem of digital exclusion remains a difficult one at higher education institutions. Information services provided by academic libraries can result in the digital exclusion of students from diverse backgrounds, such as students with various digital readiness levels, students with disabilities, as well as English-as-a-Second-Language learners. This research study shows how academic librarians can design digital learning objects that are cognizant of differences in learner traits and student profiles through the lens of design thinking. By demonstrating how the design process of digital learning objects can take into consideration users’ needs, experiences, and engagement with different technologies, this research study explains design principles of accessibility, connectivity, and scalability in creating inclusive digital learning objects as shown in various case studies. Equipped with the mindset and techniques to be mindful of diverse student learning traits and profiles when designing information services, academic libraries can improve the digital inclusion and ultimately student success at higher education institutions.

Keywords: academic librarians, digital inclusion, information services, digital learning objects, student success

Procedia PDF Downloads 216
11871 3D Human Body Reconstruction Based on Multiple Viewpoints

Authors: Jiahe Liu, HongyangYu, Feng Qian, Miao Luo

Abstract:

The aim of this study was to improve the effects of human body 3D reconstruction. The MvP algorithm was adopted to obtain key point information from multiple perspectives. This algorithm allowed the capture of human posture and joint positions from multiple angles, providing more comprehensive and accurate data. The study also incorporated the SMPL-X model, which has been widely used for human body modeling, to achieve more accurate 3D reconstruction results. The use of the MvP algorithm made it possible to observe the reconstructed object from multiple angles, thus reducing the problems of blind spots and missing information. This algorithm was able to effectively capture key point information, including the position and rotation angle of limbs, providing key data for subsequent 3D reconstruction. Compared with traditional single-view methods, the method of multi-view fusion significantly improved the accuracy and stability of reconstruction. By combining the MvP algorithm with the SMPL-X model, we successfully achieved better human body 3D reconstruction effects. The SMPL-X model is highly scalable and can generate highly realistic 3D human body models, thus providing more detail and shape information.

Keywords: 3D human reconstruction, multi-view, joint point, SMPL-X

Procedia PDF Downloads 70
11870 Peer-Assisted Learning of Ebm in, a UK Medical School: Evaluation of the NICE Evidence Search Student Champion Scheme

Authors: Emily Jin, Harry Sharples, Anne Weist

Abstract:

Introduction: NICE Evidence Search Student Champion Scheme is a peer-assisted learning scheme that aims to improve the routine use of evidence-based information by future health and social care staff. The focus is on the NICE evidence search portal that provides selected information from more than 800 reliable health, social care, and medicines sources, including up-to-date guidelines and information for the public. This paper aims to evaluate the effectiveness of the scheme when implemented in Liverpool School of Medicine and to understand the experiences of those attending. Methods: Twelve student champions were recruited and trained in February 2020 as peer tutors during a workshop facilitated by NICE. Cascade sessions were then organised and delivered on an optional basis for students, in small groups of < 10 to approximately 70 attendees. Surveys were acquired immediately before and 8-12 weeks after cascade sessions (n=47 and 45 respectively). Data from these surveys facilitated the analysis of the scheme. Results: Surveys demonstrated 74% of all attendees frequently searched for health and social care information online as a part of their studies. However, only 15% of attendees reported having prior formal training on searching for health information, despite receiving such training earlier on in the curriculum. After attending cascade sessions, students reported a 58% increase in confidence when searching for information using evidence search, from a pre-session a baseline of 36%. Conclusion: NICE Evidence Search Student Champion Scheme provided clear benefits for attending students, increasing confidence in searching for peer-reviewed, mainly secondary sources of health information. The lack of reported training represents the unmet need that the champion scheme satisfies, and this likely benefits student champions as well as attendees. Increasing confidence in searching for healthcare information online may support future evidence-based decision-making.

Keywords: evidence-based medicine, NICE, medical education, medical school, peer-assisted learning

Procedia PDF Downloads 130
11869 Predictive Analysis of Chest X-rays Using NLP and Large Language Models with the Indiana University Dataset and Random Forest Classifier

Authors: Azita Ramezani, Ghazal Mashhadiagha, Bahareh Sanabakhsh

Abstract:

This study researches the combination of Random. Forest classifiers with large language models (LLMs) and natural language processing (NLP) to improve diagnostic accuracy in chest X-ray analysis using the Indiana University dataset. Utilizing advanced NLP techniques, the research preprocesses textual data from radiological reports to extract key features, which are then merged with image-derived data. This improved dataset is analyzed with Random Forest classifiers to predict specific clinical results, focusing on the identification of health issues and the estimation of case urgency. The findings reveal that the combination of NLP, LLMs, and machine learning not only increases diagnostic precision but also reliability, especially in quickly identifying critical conditions. Achieving an accuracy of 99.35%, the model shows significant advancements over conventional diagnostic techniques. The results emphasize the large potential of machine learning in medical imaging, suggesting that these technologies could greatly enhance clinician judgment and patient outcomes by offering quicker and more precise diagnostic approximations.

Keywords: natural language processing (NLP), large language models (LLMs), random forest classifier, chest x-ray analysis, medical imaging, diagnostic accuracy, indiana university dataset, machine learning in healthcare, predictive modeling, clinical decision support systems

Procedia PDF Downloads 46
11868 Sentiment Analysis of Chinese Microblog Comments: Comparison between Support Vector Machine and Long Short-Term Memory

Authors: Xu Jiaqiao

Abstract:

Text sentiment analysis is an important branch of natural language processing. This technology is widely used in public opinion analysis and web surfing recommendations. At present, the mainstream sentiment analysis methods include three parts: sentiment analysis based on a sentiment dictionary, based on traditional machine learning, and based on deep learning. This paper mainly analyzes and compares the advantages and disadvantages of the SVM method of traditional machine learning and the Long Short-term Memory (LSTM) method of deep learning in the field of Chinese sentiment analysis, using Chinese comments on Sina Microblog as the data set. Firstly, this paper classifies and adds labels to the original comment dataset obtained by the web crawler, and then uses Jieba word segmentation to classify the original dataset and remove stop words. After that, this paper extracts text feature vectors and builds document word vectors to facilitate the training of the model. Finally, SVM and LSTM models are trained respectively. After accuracy calculation, it can be obtained that the accuracy of the LSTM model is 85.80%, while the accuracy of SVM is 91.07%. But at the same time, LSTM operation only needs 2.57 seconds, SVM model needs 6.06 seconds. Therefore, this paper concludes that: compared with the SVM model, the LSTM model is worse in accuracy but faster in processing speed.

Keywords: sentiment analysis, support vector machine, long short-term memory, Chinese microblog comments

Procedia PDF Downloads 94
11867 The Impact of Electronic Marketing on the Quality Banking Services

Authors: Ahmed Ghalem

Abstract:

The research to be explained is a collection of information about several public and private economic institutions. This information is represented in highlighting the large and useful role in adopting the method of electronic marketing. Which is widespread and easy to use among community members at the local and international levels. Which generates large sums of money with little effort and little time, and also satisfies the customers. Do these things, despite what we have said, run the risk of losing large amounts of money in a moment or a short time.

Keywords: economic, finance, bank, development, marketing

Procedia PDF Downloads 91
11866 SISSLE in Consensus-Based Ripple: Some Improvements in Speed, Security, Last Mile Connectivity and Ease of Use

Authors: Mayank Mundhra, Chester Rebeiro

Abstract:

Cryptocurrencies are rapidly finding wide application in areas such as Real Time Gross Settlements and Payments Systems. Ripple is a cryptocurrency that has gained prominence with banks and payment providers. It solves the Byzantine General’s Problem with its Ripple Protocol Consensus Algorithm (RPCA), where each server maintains a list of servers, called Unique Node List (UNL) that represents the network for the server, and will not collectively defraud it. The server believes that the network has come to a consensus when members of the UNL come to a consensus on a transaction. In this paper we improve Ripple to achieve better speed, security, last mile connectivity and ease of use. We implement guidelines and automated systems for building and maintaining UNLs for resilience, robustness, improved security, and efficient information propagation. We enhance the system so as to ensure that each server receives information from across the whole network rather than just from the UNL members. We also introduce the paradigm of UNL overlap as a function of information propagation and the trust a server assigns to its own UNL. Our design not only reduces vulnerabilities such as eclipse attacks, but also makes it easier to identify malicious behaviour and entities attempting to fraudulently Double Spend or stall the system. We provide experimental evidence of the benefits of our approach over the current Ripple scheme. We observe ≥ 4.97x and 98.22x in speedup and success rate for information propagation respectively, and ≥ 3.16x and 51.70x in speedup and success rate in consensus.

Keywords: Ripple, Kelips, unique node list, consensus, information propagation

Procedia PDF Downloads 146
11865 Combined Optical Coherence Microscopy and Spectrally Resolved Multiphoton Microscopy

Authors: Bjorn-Ole Meyer, Dominik Marti, Peter E. Andersen

Abstract:

A multimodal imaging system, combining spectrally resolved multiphoton microscopy (MPM) and optical coherence microscopy (OCM) is demonstrated. MPM and OCM are commonly integrated into multimodal imaging platforms to combine functional and morphological information. The MPM signals, such as two-photon fluorescence emission (TPFE) and signals created by second harmonic generation (SHG) are biomarkers which exhibit information on functional biological features such as the ratio of pyridine nucleotide (NAD(P)H) and flavin adenine dinucleotide (FAD) in the classification of cancerous tissue. While the spectrally resolved imaging allows for the study of biomarkers, using a spectrometer as a detector limits the imaging speed of the system significantly. To overcome those limitations, an OCM setup was added to the system, which allows for fast acquisition of structural information. Thus, after rapid imaging of larger specimens, navigation within the sample is possible. Subsequently, distinct features can be selected for further investigation using MPM. Additionally, by probing a different contrast, complementary information is obtained, and different biomarkers can be investigated. OCM images of tissue and cell samples are obtained, and distinctive features are evaluated using MPM to illustrate the benefits of the system.

Keywords: optical coherence microscopy, multiphoton microscopy, multimodal imaging, two-photon fluorescence emission

Procedia PDF Downloads 511