Search results for: enterprise intelligence
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2113

Search results for: enterprise intelligence

253 Implementation of Smart Card Automatic Fare Collection Technology in Small Transit Agencies for Standards Development

Authors: Walter E. Allen, Robert D. Murray

Abstract:

Many large transit agencies have adopted RFID technology and electronic automatic fare collection (AFC) or smart card systems, but small and rural agencies remain tied to obsolete manual, cash-based fare collection. Small countries or transit agencies can benefit from the implementation of smart card AFC technology with the promise of increased passenger convenience, added passenger satisfaction and improved agency efficiency. For transit agencies, it reduces revenue loss, improves passenger flow and bus stop data. For countries, further implementation into security, distribution of social services or currency transactions can provide greater benefits. However, small countries or transit agencies cannot afford expensive proprietary smart card solutions typically offered by the major system suppliers. Deployment of Contactless Fare Media System (CFMS) Standard eliminates the proprietary solution, ultimately lowering the cost of implementation. Acumen Building Enterprise, Inc. chose the Yuma County Intergovernmental Public Transportation Authority (YCIPTA) existing proprietary YCAT smart card system to implement CFMS. The revised system enables the purchase of fare product online with prepaid debit or credit cards using the Payment Gateway Processor. Open and interoperable smart card standards for transit have been developed. During the 90-day Pilot Operation conducted, the transit agency gathered the data from the bus AcuFare 200 Card Reader, loads (copies) the data to a USB Thumb Drive and uploads the data to the Acumen Host Processing Center for consolidation of the data into the transit agency master data file. The transition from the existing proprietary smart card data format to the new CFMS smart card data format was transparent to the transit agency cardholders. It was proven that open standards and interoperability design can work and reduce both implementation and operational costs for small transit agencies or countries looking to expand smart card technology. Acumen was able to avoid the implementation of the Payment Card Industry (PCI) Data Security Standards (DSS) which is expensive to develop and costly to operate on a continuing basis. Due to the substantial additional complexities of implementation and the variety of options presented to the transit agency cardholder, Acumen chose to implement only the Directed Autoload. To improve the implementation efficiency and the results for a similar undertaking, it should be considered that some passengers lack credit cards and are averse to technology. There are more than 1,300 small and rural agencies in the United States. This grows by 10 fold when considering small countries or rural locations throughout Latin American and the world. Acumen is evaluating additional countries, sites or transit agency that can benefit from the smart card systems. Frequently, payment card systems require extensive security procedures for implementation. The Project demonstrated the ability to purchase fare value, rides and passes with credit cards on the internet at a reasonable cost without highly complex security requirements.

Keywords: automatic fare collection, near field communication, small transit agencies, smart cards

Procedia PDF Downloads 283
252 Towards a Reinvented Cash Management Function: Mobilising Innovative Advances for Enhanced Performance and Optimised Cost Management: Insights from Large Moroccan Companies in the Casablanca-Settat Region

Authors: Badrane Nohayla, Bamousse Zineb

Abstract:

Financial crises, exchange rate volatility, fluctuations in commodity prices, increased competitive pressures, and environmental issues are all threats that businesses face. In light of these diverse challenges, proactive, agile, and innovative cash management becomes an indispensable financial shield, allowing companies to thrive despite the adverse conditions of the global environment. In the same spirit, uncertainty, turbulence, volatility, and competitiveness continue to disrupt economic environments, compelling companies to swiftly master innovative breakthroughs that provide added value. In such a context, innovation emerges as a catalytic vector for performance, aiming to reduce costs, strengthen growth, and ultimately ensure the sustainability of Moroccan companies in the national arena. Moreover, innovation in treasury management promises to be one of the key pillars of financial stability, enabling companies to navigate the tumultuous waters of a globalized environment. Therefore, the objective of this study is to better understand the impact of innovative treasury management on cost optimization and, by extension, performance improvement. To elucidate this relationship, we conducted an exploratory qualitative study with 20 large Moroccan companies operating in the Casablanca-Settat region. The results highlight that innovation at the heart of treasury management is a guarantee of sustainability against the risks of failure and stands as a true pivot of the performance of Moroccan companies, an important parameter of their financial balance and a catalytic vector of their growth in the national economic landscape. In this regard, the present study aims to explore the extent to which innovation at the core of the treasury function serves as an indispensable tool for boosting performance while optimising costs in large Moroccan companies.

Keywords: innovative cash management, artificial intelligence, financial performance, risk management, cost savings

Procedia PDF Downloads 29
251 Big Data Analytics and Public Policy: A Study in Rural India

Authors: Vasantha Gouri Prathapagiri

Abstract:

Innovations in ICT sector facilitate qualitative life style for citizens across the globe. Countries that facilitate usage of new techniques in ICT, i.e., big data analytics find it easier to fulfil the needs of their citizens. Big data is characterised by its volume, variety, and speed. Analytics involves its processing in a cost effective way in order to draw conclusion for their useful application. Big data also involves into the field of machine learning, artificial intelligence all leading to accuracy in data presentation useful for public policy making. Hence using data analytics in public policy making is a proper way to march towards all round development of any country. The data driven insights can help the government to take important strategic decisions with regard to socio-economic development of her country. Developed nations like UK and USA are already far ahead on the path of digitization with the support of Big Data analytics. India is a huge country and is currently on the path of massive digitization being realised through Digital India Mission. Internet connection per household is on the rise every year. This transforms into a massive data set that has the potential to improvise the public services delivery system into an effective service mechanism for Indian citizens. In fact, when compared to developed nations, this capacity is being underutilized in India. This is particularly true for administrative system in rural areas. The present paper focuses on the need for big data analytics adaptation in Indian rural administration and its contribution towards development of the country on a faster pace. Results of the research focussed on the need for increasing awareness and serious capacity building of the government personnel working for rural development with regard to big data analytics and its utility for development of the country. Multiple public policies are framed and implemented for rural development yet the results are not as effective as they should be. Big data has a major role to play in this context as can assist in improving both policy making and implementation aiming at all round development of the country.

Keywords: Digital India Mission, public service delivery system, public policy, Indian administration

Procedia PDF Downloads 159
250 Smart Signature - Medical Communication without Barrier

Authors: Chia-Ying Lin

Abstract:

This paper explains how to enhance doctor-patient communication and nurse-patient communication through multiple intelligence signing methods and user-centered. It is hoped that through the implementation of the "electronic consent", the problems faced by the paper consent can be solved: storage methods, resource utilization, convenience, correctness of information, integrated management, statistical analysis and other related issues. Make better use and allocation of resources to provide better medical quality. First, invite the medical records department to assist in the inventory of paper consent in the hospital: organising, classifying, merging, coding, and setting. Second, plan the electronic consent configuration file: set the form number, consent form group, fields and templates, and the corresponding doctor's order code. Next, Summarize four types of rapid methods of electronic consent: according to the doctor's order, according to the medical behavior, according to the schedule, and manually generate the consent form. Finally, system promotion and adjustment: form an "electronic consent promotion team" to improve, follow five major processes: planning, development, testing, release, and feedback, and invite clinical units to raise the difficulties faced in the promotion, and make improvements to the problems. The electronic signature rate of the whole hospital will increase from 4% in January 2022 to 79% in November 2022. Use the saved resources more effectively, including: reduce paper usage (reduce carbon footprint), reduce the cost of ink cartridges, re-plan and use the space for paper medical records, and save human resources to provide better services. Through the introduction of information technology and technology, the main spirit of "lean management" is implemented. Transforming and reengineering the process to eliminate unnecessary waste is also the highest purpose of this project.

Keywords: smart signature, electronic consent, electronic medical records, user-centered, doctor-patient communication, nurse-patient communication

Procedia PDF Downloads 126
249 Insecurity and Insurgency on Economic Development of Nigeria

Authors: Uche Lucy Onyekwelu, Uche B. Ugwuanyi

Abstract:

Suffice to say that socio-economic disruptions of any form is likely to affect the wellbeing of the citizenry. The upsurge of social disequilibrium caused by the incessant disruptive tendencies exhibited by youths and some others in Nigeria are not helping matters. In Nigeria the social unrest has caused different forms of draw backs in Socio Economic Development. This study has empirically evaluated the impact of insecurity and insurgency on the Economic Development of Nigeria. The paper noted that the different forms of insecurity in Nigeria are namely: Insurgency and Banditry as witnessed in Northern Nigeria; Militancy: Niger Delta area and self-determination groups pursuing various forms of agenda such as Sit –at- Home Syndrome in the South Eastern Nigeria and other secessionist movements. All these have in one way or the other hampered Economic development in Nigeria. Data for this study were collected through primary and secondary sources using questionnaire and some existing documentations. Cost of investment in different aspects of security outfits in Nigeria represents the independent variable while the differentials in the Gross Domestic Product(GDP) and Human Development Index(HDI) are the measures of the dependent variable. Descriptive statistics and Simple Linear Regression analytical tool were employed in the data analysis. The result revealed that Insurgency/Insecurity negatively affect the economic development of the different parts of Nigeria. Following the findings, a model to analyse the effect of insecurity and insurgency was developed, named INSECUREDEVNIG. It implies that the economic development of Nigeria will continue to deteriorate if insurgency and insecurity continue. The study therefore recommends that the government should do all it could to nurture its human capital, adequately fund the state security apparatus and employ individuals of high integrity to manage the various security outfits in Nigeria. The government should also as a matter of urgency train the security personnel in intelligence cum Information and Communications Technology to enable them ensure the effectiveness of implementation of security policies needed to sustain Gross Domestic Product and Human Capital Index of Nigeria.

Keywords: insecurity, insurgency, gross domestic product, human development index, Nigeria

Procedia PDF Downloads 102
248 Healthcare-SignNet: Advanced Video Classification for Medical Sign Language Recognition Using CNN and RNN Models

Authors: Chithra A. V., Somoshree Datta, Sandeep Nithyanandan

Abstract:

Sign Language Recognition (SLR) is the process of interpreting and translating sign language into spoken or written language using technological systems. It involves recognizing hand gestures, facial expressions, and body movements that makeup sign language communication. The primary goal of SLR is to facilitate communication between hearing- and speech-impaired communities and those who do not understand sign language. Due to the increased awareness and greater recognition of the rights and needs of the hearing- and speech-impaired community, sign language recognition has gained significant importance over the past 10 years. Technological advancements in the fields of Artificial Intelligence and Machine Learning have made it more practical and feasible to create accurate SLR systems. This paper presents a distinct approach to SLR by framing it as a video classification problem using Deep Learning (DL), whereby a combination of Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs) has been used. This research targets the integration of sign language recognition into healthcare settings, aiming to improve communication between medical professionals and patients with hearing impairments. The spatial features from each video frame are extracted using a CNN, which captures essential elements such as hand shapes, movements, and facial expressions. These features are then fed into an RNN network that learns the temporal dependencies and patterns inherent in sign language sequences. The INCLUDE dataset has been enhanced with more videos from the healthcare domain and the model is evaluated on the same. Our model achieves 91% accuracy, representing state-of-the-art performance in this domain. The results highlight the effectiveness of treating SLR as a video classification task with the CNN-RNN architecture. This approach not only improves recognition accuracy but also offers a scalable solution for real-time SLR applications, significantly advancing the field of accessible communication technologies.

Keywords: sign language recognition, deep learning, convolution neural network, recurrent neural network

Procedia PDF Downloads 27
247 Transforming Data Science Curriculum Through Design Thinking

Authors: Samar Swaid

Abstract:

Today, corporates are moving toward the adoption of Design-Thinking techniques to develop products and services, putting their consumer as the heart of the development process. One of the leading companies in Design-Thinking, IDEO (Innovation, Design, Engineering Organization), defines Design-Thinking as an approach to problem-solving that relies on a set of multi-layered skills, processes, and mindsets that help people generate novel solutions to problems. Design thinking may result in new ideas, narratives, objects or systems. It is about redesigning systems, organizations, infrastructures, processes, and solutions in an innovative fashion based on the users' feedback. Tim Brown, president and CEO of IDEO, sees design thinking as a human-centered approach that draws from the designer's toolkit to integrate people's needs, innovative technologies, and business requirements. The application of design thinking has been witnessed to be the road to developing innovative applications, interactive systems, scientific software, healthcare application, and even to utilizing Design-Thinking to re-think business operations, as in the case of Airbnb. Recently, there has been a movement to apply design thinking to machine learning and artificial intelligence to ensure creating the "wow" effect on consumers. The Association of Computing Machinery task force on Data Science program states that" Data scientists should be able to implement and understand algorithms for data collection and analysis. They should understand the time and space considerations of algorithms. They should follow good design principles developing software, understanding the importance of those principles for testability and maintainability" However, this definition hides the user behind the machine who works on data preparation, algorithm selection and model interpretation. Thus, the Data Science program includes design thinking to ensure meeting the user demands, generating more usable machine learning tools, and developing ways of framing computational thinking. Here, describe the fundamentals of Design-Thinking and teaching modules for data science programs.

Keywords: data science, design thinking, AI, currculum, transformation

Procedia PDF Downloads 81
246 A Readiness Framework for Digital Innovation in Education: The Context of Academics and Policymakers in Higher Institutions of Learning to Assess the Preparedness of Their Institutions to Adopt and Incorporate Digital Innovation

Authors: Lufungula Osembe

Abstract:

The field of education has witnessed advances in technology and digital transformation. The methods of teaching have undergone significant changes in recent years, resulting in effects on various areas such as pedagogies, curriculum design, personalized teaching, gamification, data analytics, cloud-based learning applications, artificial intelligence tools, advanced plug-ins in LMS, and the emergence of multimedia creation and design. The field of education has not been immune to the changes brought about by digital innovation in recent years, similar to other fields such as engineering, health, science, and technology. There is a need to look at the variables/elements that digital innovation brings to education and develop a framework for higher institutions of learning to assess their readiness to create a viable environment for digital innovation to be successfully adopted. Given the potential benefits of digital innovation in education, it is essential to develop a framework that can assist academics and policymakers in higher institutions of learning to evaluate the effectiveness of adopting and adapting to the evolving landscape of digital innovation in education. The primary research question addressed in this study is to establish the preparedness of higher institutions of learning to adopt and adapt to the evolving landscape of digital innovation. This study follows a Design Science Research (DSR) paradigm to develop a framework for academics and policymakers in higher institutions of learning to evaluate the readiness of their institutions to adopt digital innovation in education. The Design Science Research paradigm is proposed to aid in developing a readiness framework for digital innovation in education. This study intends to follow the Design Science Research (DSR) methodology, which includes problem awareness, suggestion, development, evaluation, and conclusion. One of the major contributions of this study will be the development of the framework for digital innovation in education. Given the various opportunities offered by digital innovation in recent years, the need to create a readiness framework for digital innovation will play a crucial role in guiding academics and policymakers in their quest to align with emerging technologies facilitated by digital innovation in education.

Keywords: digital innovation, DSR, education, opportunities, research

Procedia PDF Downloads 69
245 Hand Gesture Detection via EmguCV Canny Pruning

Authors: N. N. Mosola, S. J. Molete, L. S. Masoebe, M. Letsae

Abstract:

Hand gesture recognition is a technique used to locate, detect, and recognize a hand gesture. Detection and recognition are concepts of Artificial Intelligence (AI). AI concepts are applicable in Human Computer Interaction (HCI), Expert systems (ES), etc. Hand gesture recognition can be used in sign language interpretation. Sign language is a visual communication tool. This tool is used mostly by deaf societies and those with speech disorder. Communication barriers exist when societies with speech disorder interact with others. This research aims to build a hand recognition system for Lesotho’s Sesotho and English language interpretation. The system will help to bridge the communication problems encountered by the mentioned societies. The system has various processing modules. The modules consist of a hand detection engine, image processing engine, feature extraction, and sign recognition. Detection is a process of identifying an object. The proposed system uses Canny pruning Haar and Haarcascade detection algorithms. Canny pruning implements the Canny edge detection. This is an optimal image processing algorithm. It is used to detect edges of an object. The system employs a skin detection algorithm. The skin detection performs background subtraction, computes the convex hull, and the centroid to assist in the detection process. Recognition is a process of gesture classification. Template matching classifies each hand gesture in real-time. The system was tested using various experiments. The results obtained show that time, distance, and light are factors that affect the rate of detection and ultimately recognition. Detection rate is directly proportional to the distance of the hand from the camera. Different lighting conditions were considered. The more the light intensity, the faster the detection rate. Based on the results obtained from this research, the applied methodologies are efficient and provide a plausible solution towards a light-weight, inexpensive system which can be used for sign language interpretation.

Keywords: canny pruning, hand recognition, machine learning, skin tracking

Procedia PDF Downloads 185
244 Income Generation and Employment Opportunity of the Entrepreneurs and Farmers Through Production, Processing, and Marketing of Medicinal Plants in Bangladesh

Authors: Md. Nuru Miah, A. F. M. Akhter Uddin

Abstract:

Medicinal plants are grown naturally in a tropical environment in Bangladesh and used as drug and therapeutic agents in the health care system. According to Bangladesh Agricultural Research Institute (BARI), there are 722 species of medicinal plants in the country. Of them, 255 plants are utilized by the manufacturers of Ayurvedic and Unani medicines. Medicinal plants like Aloevera, Ashwagonda, shotomul,Tulsi, Vuikumra, Misridana are extensively cultivated in some selected areas as well; where Aloevera scored the highest position in production. In the early 1980, Ayurvedic and Unani companies procured 80 percent of medicinal plants from natural forests, and the rest 20 percent was imported. Now the scenario has changed; 80 percent is imported, and the rest 20 percent is collected from local products(Source: Astudy on sectorbased need assessment of Business promotion council-Herbal products and medicinal plants, page-4). Uttara Development Program Society, a leading Non- Government development organization in Bangladesh, has been implementing a value chain development project under promoting Agricultural commercialization and Enterprises of Pally Karma Sahayak Foundation (PKSF) funded by the International Fund for Agricultural Development (IFAD) in Natore Sadar Upazila from April 2017 to sustainably develop the technological interventions for products and market development. The ultimate goal of the project is to increase income, generate employment and develop this sector as a sustainable business enterprise. Altogether 10,000 farmers (Nursery owners, growers, input supplier, processors, whole sellers, and retailers) are engaged in different activities of the project. The entrepreneurs engaged in medicinal plant cultivation did not know and follow environmental and good agricultural practices. They used to adopt traditional methodology in production and processing. Locally the farmers didn’t have any positive initiative to expand their business as well as developvalue added products. A lot of diversified products could be possible to develop and marketed with the introduction of post-harvest processing technology and market linkage with the local and global buyer. Training is imparted to the nursery owners and herbal growers on production technologies, sowing method, use of organic fertilizers/compost/pesticides, harvesting procedures, and storage facilities. Different types of herbal tea like Rosella, Moringa, Tulshi, and Basak are being produced and packed locally with a good scope of its marketing in different cities of the country. The project has been able to achieve a significant impact in the development of production technologies, but still, there is room for further improvement in processing, packaging, and marketing level. The core intervention of the current project to develop some entrepreneurs for branding, packaging, promotion, and marketing while considering environment friendly practices. The present strategies will strengthen the knowledge and skills of the entrepreneurs for the production and marketing of their products, maintaining worldwide accepted compliance system for easy access to the global market.

Keywords: aloe vera, herbs and shrubs, market, interventions

Procedia PDF Downloads 96
243 Design and Development of an Autonomous Beach Cleaning Vehicle

Authors: Mahdi Allaoua Seklab, Süleyman BaşTürk

Abstract:

In the quest to enhance coastal environmental health, this study introduces a fully autonomous beach cleaning machine, a breakthrough in leveraging green energy and advanced artificial intelligence for ecological preservation. Designed to operate independently, the machine is propelled by a solar-powered system, underscoring a commitment to sustainability and the use of renewable energy in autonomous robotics. The vehicle's autonomous navigation is achieved through a sophisticated integration of LIDAR and a camera system, utilizing an SSD MobileNet V2 object detection model for accurate and real-time trash identification. The SSD framework, renowned for its efficiency in detecting objects in various scenarios, is coupled with the lightweight and precise highly MobileNet V2 architecture, making it particularly suited for the computational constraints of on-board processing in mobile robotics. Training of the SSD MobileNet V2 model was conducted on Google Colab, harnessing cloud-based GPU resources to facilitate a rapid and cost-effective learning process. The model was refined with an extensive dataset of annotated beach debris, optimizing the parameters using the Adam optimizer and a cross-entropy loss function to achieve high-precision trash detection. This capability allows the machine to intelligently categorize and target waste, leading to more effective cleaning operations. This paper details the design and functionality of the beach cleaning machine, emphasizing its autonomous operational capabilities and the novel application of AI in environmental robotics. The results showcase the potential of such technology to fill existing gaps in beach maintenance, offering a scalable and eco-friendly solution to the growing problem of coastal pollution. The deployment of this machine represents a significant advancement in the field, setting a new standard for the integration of autonomous systems in the service of environmental stewardship.

Keywords: autonomous beach cleaning machine, renewable energy systems, coastal management, environmental robotics

Procedia PDF Downloads 27
242 Enhancing Sell-In and Sell-Out Forecasting Using Ensemble Machine Learning Method

Authors: Vishal Das, Tianyi Mao, Zhicheng Geng, Carmen Flores, Diego Pelloso, Fang Wang

Abstract:

Accurate sell-in and sell-out forecasting is a ubiquitous problem in the retail industry. It is an important element of any demand planning activity. As a global food and beverage company, Nestlé has hundreds of products in each geographical location that they operate in. Each product has its sell-in and sell-out time series data, which are forecasted on a weekly and monthly scale for demand and financial planning. To address this challenge, Nestlé Chilein collaboration with Amazon Machine Learning Solutions Labhas developed their in-house solution of using machine learning models for forecasting. Similar products are combined together such that there is one model for each product category. In this way, the models learn from a larger set of data, and there are fewer models to maintain. The solution is scalable to all product categories and is developed to be flexible enough to include any new product or eliminate any existing product in a product category based on requirements. We show how we can use the machine learning development environment on Amazon Web Services (AWS) to explore a set of forecasting models and create business intelligence dashboards that can be used with the existing demand planning tools in Nestlé. We explored recent deep learning networks (DNN), which show promising results for a variety of time series forecasting problems. Specifically, we used a DeepAR autoregressive model that can group similar time series together and provide robust predictions. To further enhance the accuracy of the predictions and include domain-specific knowledge, we designed an ensemble approach using DeepAR and XGBoost regression model. As part of the ensemble approach, we interlinked the sell-out and sell-in information to ensure that a future sell-out influences the current sell-in predictions. Our approach outperforms the benchmark statistical models by more than 50%. The machine learning (ML) pipeline implemented in the cloud is currently being extended for other product categories and is getting adopted by other geomarkets.

Keywords: sell-in and sell-out forecasting, demand planning, DeepAR, retail, ensemble machine learning, time-series

Procedia PDF Downloads 274
241 Rethinking Classical Concerts in the Digital Era: Transforming Sound, Experience, and Engagement for the New Generation

Authors: Orit Wolf

Abstract:

Classical music confronts a crucial challenge: updating cherished concert traditions for the digital age. This paper is a journey, and a quest to make classical concerts resonate with a new generation. It's not just about asking questions; it's about exploring the future of classical concerts and their potential to captivate and connect with today's audience in an era defined by change. The younger generation, known for their love of diversity, interactive experiences, and multi-sensory immersion, cannot be overlooked. This paper explores innovative strategies that forge deep connections with audiences whose relationship with classical music differs from the past. The urgency of this challenge drives the transformation of classical concerts. Examining classical concerts is necessary to understand how they can harmonize with contemporary sensibilities. New dimensions in audiovisual experiences that enchant the emerging generation are sought. Classical music must embrace the technological era while staying open to fusion and cross-cultural collaboration possibilities. The role of technology and Artificial Intelligence (AI) in reshaping classical concerts is under research. The fusion of classical music with digital experiences and dynamic interdisciplinary collaborations breathes new life into the concert experience. It aligns classical music with the expectations of modern audiences, making it more relevant and engaging. Exploration extends to the structure of classical concerts. Conventions are challenged, and ways to make classical concerts more accessible and captivating are sought. Inspired by innovative artistic collaborations, musical genres and styles are redefined, transforming the relationship between performers and the audience. This paper, therefore, aims to be a catalyst for dialogue and a beacon of innovation. A set of critical inquiries integral to reshaping classical concerts for the digital age is presented. As the world embraces digital transformation, classical music seeks resonance with contemporary audiences, redefining the concert experience while remaining true to its roots and embracing revolutions in the digital age.

Keywords: new concert formats, reception of classical music, interdiscplinary concerts, innovation in the new musical era, mash-up, cross culture, innovative concerts, engaging musical performances

Procedia PDF Downloads 64
240 AI/ML Atmospheric Parameters Retrieval Using the “Atmospheric Retrievals conditional Generative Adversarial Network (ARcGAN)”

Authors: Thomas Monahan, Nicolas Gorius, Thanh Nguyen

Abstract:

Exoplanet atmospheric parameters retrieval is a complex, computationally intensive, inverse modeling problem in which an exoplanet’s atmospheric composition is extracted from an observed spectrum. Traditional Bayesian sampling methods require extensive time and computation, involving algorithms that compare large numbers of known atmospheric models to the input spectral data. Runtimes are directly proportional to the number of parameters under consideration. These increased power and runtime requirements are difficult to accommodate in space missions where model size, speed, and power consumption are of particular importance. The use of traditional Bayesian sampling methods, therefore, compromise model complexity or sampling accuracy. The Atmospheric Retrievals conditional Generative Adversarial Network (ARcGAN) is a deep convolutional generative adversarial network that improves on the previous model’s speed and accuracy. We demonstrate the efficacy of artificial intelligence to quickly and reliably predict atmospheric parameters and present it as a viable alternative to slow and computationally heavy Bayesian methods. In addition to its broad applicability across instruments and planetary types, ARcGAN has been designed to function on low power application-specific integrated circuits. The application of edge computing to atmospheric retrievals allows for real or near-real-time quantification of atmospheric constituents at the instrument level. Additionally, edge computing provides both high-performance and power-efficient computing for AI applications, both of which are critical for space missions. With the edge computing chip implementation, ArcGAN serves as a strong basis for the development of a similar machine-learning algorithm to reduce the downlinked data volume from the Compact Ultraviolet to Visible Imaging Spectrometer (CUVIS) onboard the DAVINCI mission to Venus.

Keywords: deep learning, generative adversarial network, edge computing, atmospheric parameters retrieval

Procedia PDF Downloads 170
239 Flood Simulation and Forecasting for Sustainable Planning of Response in Municipalities

Authors: Mariana Damova, Stanko Stankov, Emil Stoyanov, Hristo Hristov, Hermand Pessek, Plamen Chernev

Abstract:

We will present one of the first use cases on the DestinE platform, a joint initiative of the European Commission, European Space Agency and EUMETSAT, providing access to global earth observation, meteorological and statistical data, and emphasize the good practice of intergovernmental agencies acting in concert. Further, we will discuss the importance of space-bound disruptive solutions for improving the balance between the ever-increasing water-related disasters coming from climate change and minimizing their economic and societal impact. The use case focuses on forecasting floods and estimating the impact of flood events on the urban environment and the ecosystems in the affected areas with the purpose of helping municipal decision-makers to analyze and plan resource needs and to forge human-environment relationships by providing farmers with insightful information for improving their agricultural productivity. For the forecast, we will adopt an EO4AI method of our platform ISME-HYDRO, in which we employ a pipeline of neural networks applied to in-situ measurements and satellite data of meteorological factors influencing the hydrological and hydrodynamic status of rivers and dams, such as precipitations, soil moisture, vegetation index, snow cover to model flood events and their span. ISME-HYDRO platform is an e-infrastructure for water resources management based on linked data, extended with further intelligence that generates forecasts with the method described above, throws alerts, formulates queries, provides superior interactivity and drives communication with the users. It provides synchronized visualization of table views, graphviews and interactive maps. It will be federated with the DestinE platform.

Keywords: flood simulation, AI, Earth observation, e-Infrastructure, flood forecasting, flood areas localization, response planning, resource estimation

Procedia PDF Downloads 21
238 Quality Assurance in Translation Crowdsourcing: The TED Open Translation Project

Authors: Ya-Mei Chen

Abstract:

The participatory culture enabled by Web 2.0 technologies has led to the emergence of online translation crowdsourcing, which mainly relies on the collective intelligence of volunteer translators. Due to the fact that many volunteer translators do not have formal translator training, concerns have been raised about the quality of crowdsourced translations. Some empirical research has been done to examine the translation quality of for-profit crowdsourcing initiatives. However, quality assurance of non-profit translation crowdsourcing has rarely been explored in detail. Using the TED Open Translation Project as a case study, this paper investigates how the translation-review-approval method adopted by TED can (1) direct the volunteer translators’ use of translation strategies as well as the reviewers’ adoption of revising strategies and (2) shape the final translation products. To well examine the actual effect of TED’s translation-review-approval method, this paper will focus on its two major quality assurance mechanisms, that is, TED’s style guidelines and quality review. Based on an anonymous questionnaire, this research will first explore whether the volunteer translators and reviewers are aware of the style guidelines and whether their use of translation strategies is similar to that advised in the guidelines. The questionnaire, which will be posted online, will consist of two parts: demographic information and translation strategies. The invitations to complete it will then be distributed through TED Translator Facebook groups. With an aim to investigate if the style guidelines have any substantial impacts on actual subtitling practices, a comparison will be made between the original English subtitles of 20 TED talks (each around 5 to 7 minutes) and their Chinese subtitle translations to identify regularly adopted strategies. Concerning the function of the reviewing stage, a comparative study will be conducted between the drafts of Chinese subtitles for 10 short English talks and the revised versions of these drafts so as to examine the actual revising strategies and their effect on translation quality. According to the results obtained from the questionnaire and textual comparisons, this paper will provide in-depth analysis of quality assurance of the TED Open Translation Project. It is hoped that this research, through a detailed investigation of non-profit translation crowdsourcing, can enable translation researchers and practitioners to have a better understanding of quality control in translation crowdsourcing in the digital age.

Keywords: quality assurance, TED, translation crowdsourcing, volunteer translators

Procedia PDF Downloads 231
237 Hybrid Method for Smart Suggestions in Conversations for Online Marketplaces

Authors: Yasamin Rahimi, Ali Kamandi, Abbas Hoseini, Hesam Haddad

Abstract:

Online/offline chat is a convenient approach in the electronic markets of second-hand products in which potential customers would like to have more information about the products to fill the information gap between buyers and sellers. Online peer in peer market is trying to create artificial intelligence-based systems that help customers ask more informative questions in an easier way. In this article, we introduce a method for the question/answer system that we have developed for the top-ranked electronic market in Iran called Divar. When it comes to secondhand products, incomplete product information in a purchase will result in loss to the buyer. One way to balance buyer and seller information of a product is to help the buyer ask more informative questions when purchasing. Also, the short time to start and achieve the desired result of the conversation was one of our main goals, which was achieved according to A/B tests results. In this paper, we propose and evaluate a method for suggesting questions and answers in the messaging platform of the e-commerce website Divar. Creating such systems is to help users gather knowledge about the product easier and faster, All from the Divar database. We collected a dataset of around 2 million messages in Persian colloquial language, and for each category of product, we gathered 500K messages, of which only 2K were Tagged, and semi-supervised methods were used. In order to publish the proposed model to production, it is required to be fast enough to process 10 million messages daily on CPU processors. In order to reach that speed, in many subtasks, faster and simplistic models are preferred over deep neural models. The proposed method, which requires only a small amount of labeled data, is currently used in Divar production on CPU processors, and 15% of buyers and seller’s messages in conversations is directly chosen from our model output, and more than 27% of buyers have used this model suggestions in at least one daily conversation.

Keywords: smart reply, spell checker, information retrieval, intent detection, question answering

Procedia PDF Downloads 187
236 Career Guidance System Using Machine Learning

Authors: Mane Darbinyan, Lusine Hayrapetyan, Elen Matevosyan

Abstract:

Artificial Intelligence in Education (AIED) has been created to help students get ready for the workforce, and over the past 25 years, it has grown significantly, offering a variety of technologies to support academic, institutional, and administrative services. However, this is still challenging, especially considering the labor market's rapid change. While choosing a career, people face various obstacles because they do not take into consideration their own preferences, which might lead to many other problems like shifting jobs, work stress, occupational infirmity, reduced productivity, and manual error. Besides preferences, people should properly evaluate their technical and non-technical skills, as well as their personalities. Professional counseling has become a difficult undertaking for counselors due to the wide range of career choices brought on by changing technological trends. It is necessary to close this gap by utilizing technology that makes sophisticated predictions about a person's career goals based on their personality. Hence, there is a need to create an automated model that would help in decision-making based on user inputs. Improving career guidance can be achieved by embedding machine learning into the career consulting ecosystem. There are various systems of career guidance that work based on the same logic, such as the classification of applicants, matching applications with appropriate departments or jobs, making predictions, and providing suitable recommendations. Methodologies like KNN, Neural Networks, K-means clustering, D-Tree, and many other advanced algorithms are applied in the fields of data and compute some data, which is helpful to predict the right careers. Besides helping users with their career choice, these systems provide numerous opportunities which are very useful while making this hard decision. They help the candidate to recognize where he/she specifically lacks sufficient skills so that the candidate can improve those skills. They are also capable to offer an e-learning platform, taking into account the user's lack of knowledge. Furthermore, users can be provided with details on a particular job, such as the abilities required to excel in that industry.

Keywords: career guidance system, machine learning, career prediction, predictive decision, data mining, technical and non-technical skills

Procedia PDF Downloads 80
235 Classification of Forest Types Using Remote Sensing and Self-Organizing Maps

Authors: Wanderson Goncalves e Goncalves, José Alberto Silva de Sá

Abstract:

Human actions are a threat to the balance and conservation of the Amazon forest. Therefore the environmental monitoring services play an important role as the preservation and maintenance of this environment. This study classified forest types using data from a forest inventory provided by the 'Florestal e da Biodiversidade do Estado do Pará' (IDEFLOR-BIO), located between the municipalities of Santarém, Juruti and Aveiro, in the state of Pará, Brazil, covering an area approximately of 600,000 hectares, Bands 3, 4 and 5 of the TM-Landsat satellite image, and Self - Organizing Maps. The information from the satellite images was extracted using QGIS software 2.8.1 Wien and was used as a database for training the neural network. The midpoints of each sample of forest inventory have been linked to images. Later the Digital Numbers of the pixels have been extracted, composing the database that fed the training process and testing of the classifier. The neural network was trained to classify two forest types: Rain Forest of Lowland Emerging Canopy (Dbe) and Rain Forest of Lowland Emerging Canopy plus Open with palm trees (Dbe + Abp) in the Mamuru Arapiuns glebes of Pará State, and the number of examples in the training data set was 400, 200 examples for each class (Dbe and Dbe + Abp), and the size of the test data set was 100, with 50 examples for each class (Dbe and Dbe + Abp). Therefore, total mass of data consisted of 500 examples. The classifier was compiled in Orange Data Mining 2.7 Software and was evaluated in terms of the confusion matrix indicators. The results of the classifier were considered satisfactory, and being obtained values of the global accuracy equal to 89% and Kappa coefficient equal to 78% and F1 score equal to 0,88. It evaluated also the efficiency of the classifier by the ROC plot (receiver operating characteristics), obtaining results close to ideal ratings, showing it to be a very good classifier, and demonstrating the potential of this methodology to provide ecosystem services, particularly in anthropogenic areas in the Amazon.

Keywords: artificial neural network, computational intelligence, pattern recognition, unsupervised learning

Procedia PDF Downloads 361
234 Career Guidance System Using Machine Learning

Authors: Mane Darbinyan, Lusine Hayrapetyan, Elen Matevosyan

Abstract:

Artificial Intelligence in Education (AIED) has been created to help students get ready for the workforce, and over the past 25 years, it has grown significantly, offering a variety of technologies to support academic, institutional, and administrative services. However, this is still challenging, especially considering the labor market's rapid change. While choosing a career, people face various obstacles because they do not take into consideration their own preferences, which might lead to many other problems like shifting jobs, work stress, occupational infirmity, reduced productivity, and manual error. Besides preferences, people should evaluate properly their technical and non-technical skills, as well as their personalities. Professional counseling has become a difficult undertaking for counselors due to the wide range of career choices brought on by changing technological trends. It is necessary to close this gap by utilizing technology that makes sophisticated predictions about a person's career goals based on their personality. Hence, there is a need to create an automated model that would help in decision-making based on user inputs. Improving career guidance can be achieved by embedding machine learning into the career consulting ecosystem. There are various systems of career guidance that work based on the same logic, such as the classification of applicants, matching applications with appropriate departments or jobs, making predictions, and providing suitable recommendations. Methodologies like KNN, neural networks, K-means clustering, D-Tree, and many other advanced algorithms are applied in the fields of data and compute some data, which is helpful to predict the right careers. Besides helping users with their career choice, these systems provide numerous opportunities which are very useful while making this hard decision. They help the candidate to recognize where he/she specifically lacks sufficient skills so that the candidate can improve those skills. They are also capable of offering an e-learning platform, taking into account the user's lack of knowledge. Furthermore, users can be provided with details on a particular job, such as the abilities required to excel in that industry.

Keywords: career guidance system, machine learning, career prediction, predictive decision, data mining, technical and non-technical skills

Procedia PDF Downloads 70
233 Physics Informed Deep Residual Networks Based Type-A Aortic Dissection Prediction

Authors: Joy Cao, Min Zhou

Abstract:

Purpose: Acute Type A aortic dissection is a well-known cause of extremely high mortality rate. A highly accurate and cost-effective non-invasive predictor is critically needed so that the patient can be treated at earlier stage. Although various CFD approaches have been tried to establish some prediction frameworks, they are sensitive to uncertainty in both image segmentation and boundary conditions. Tedious pre-processing and demanding calibration procedures requirement further compound the issue, thus hampering their clinical applicability. Using the latest physics informed deep learning methods to establish an accurate and cost-effective predictor framework are amongst the main goals for a better Type A aortic dissection treatment. Methods: Via training a novel physics-informed deep residual network, with non-invasive 4D MRI displacement vectors as inputs, the trained model can cost-effectively calculate all these biomarkers: aortic blood pressure, WSS, and OSI, which are used to predict potential type A aortic dissection to avoid the high mortality events down the road. Results: The proposed deep learning method has been successfully trained and tested with both synthetic 3D aneurysm dataset and a clinical dataset in the aortic dissection context using Google colab environment. In both cases, the model has generated aortic blood pressure, WSS, and OSI results matching the expected patient’s health status. Conclusion: The proposed novel physics-informed deep residual network shows great potential to create a cost-effective, non-invasive predictor framework. Additional physics-based de-noising algorithm will be added to make the model more robust to clinical data noises. Further studies will be conducted in collaboration with big institutions such as Cleveland Clinic with more clinical samples to further improve the model’s clinical applicability.

Keywords: type-a aortic dissection, deep residual networks, blood flow modeling, data-driven modeling, non-invasive diagnostics, deep learning, artificial intelligence.

Procedia PDF Downloads 89
232 Using Serious Games to Integrate the Potential of Mass Customization into the Fuzzy Front-End of New Product Development

Authors: Michael N. O'Sullivan, Con Sheahan

Abstract:

Mass customization is the idea of offering custom products or services to satisfy the needs of each individual customer while maintaining the efficiency of mass production. Technologies like 3D printing and artificial intelligence have many start-ups hoping to capitalize on this dream of creating personalized products at an affordable price, and well established companies scrambling to innovate and maintain their market share. However, the majority of them are failing as they struggle to understand one key question – where does customization make sense? Customization and personalization only make sense where the value of the perceived benefit outweighs the cost to implement it. In other words, will people pay for it? Looking at the Kano Model makes it clear that it depends on the product. In products where customization is an inherent need, like prosthetics, mass customization technologies can be highly beneficial. However, for products that already sell as a standard, like headphones, offering customization is likely only an added bonus, and so the product development team must figure out if the customers’ perception of the added value of this feature will outweigh its premium price tag. This can be done through the use of a ‘serious game,’ whereby potential customers are given a limited budget to collaboratively buy and bid on potential features of the product before it is developed. If the group choose to buy customization over other features, then the product development team should implement it into their design. If not, the team should prioritize the features on which the customers have spent their budget. The level of customization purchased can also be translated to an appropriate production method, for example, the most expensive type of customization would likely be free-form design and could be achieved through digital fabrication, while a lower level could be achieved through short batch production. Twenty-five teams of final year students from design, engineering, construction and technology tested this methodology when bringing a product from concept through to production specification, and found that it allowed them to confidently decide what level of customization, if any, would be worth offering for their product, and what would be the best method of producing it. They also found that the discussion and negotiations between players during the game led to invaluable insights, and often decided to play a second game where they offered customers the option to buy the various customization ideas that had been discussed during the first game.

Keywords: Kano model, mass customization, new product development, serious game

Procedia PDF Downloads 134
231 The Effect of Artificial Intelligence on Petroleum Industry and Production

Authors: Mina Shokry Hanna Saleh Tadros

Abstract:

The centrality of the Petroleum Industry in the world energy is undoubted. The world economy almost runs and depends on petroleum. Petroleum industry is a multi-trillion industry; it turns otherwise poor and underdeveloped countries into wealthy nations and thrusts them at the center of international diplomacy. Although these developing nations lack the necessary technology to explore and exploit petroleum resources they are not without help as developed nations, represented by their multinational corporations are ready and willing to provide both the technical and managerial expertise necessary for the development of this natural resource. However, the exploration of these petroleum resources comes with, sometimes, grave, concomitant consequences. These consequences are especially pronounced with respect to the environment. From the British Petroleum Oil rig explosion and the resultant oil spillage and pollution in New Mexico, United States to the Mobil Oil spillage along Egyptian coast, the story and consequence is virtually the same. Egypt’s delta Region produces Nigeria’s petroleum which accounts for more than ninety-five percent of Nigeria’s foreign exchange earnings. Between 1999 and 2007, Egypt earned more than $400 billion from petroleum exports. Nevertheless, petroleum exploration and exploitation has devastated the Delta environment. From oil spillage which pollutes the rivers, farms and wetlands to gas flaring by the multi-national corporations; the consequences is similar-a region that has been devastated by petroleum exploitation. This paper thus seeks to examine the consequences and impact of petroleum pollution in the Egypt Delta with particular reference on the right of the people of Niger Delta to a healthy environment. The paper further seeks to examine the relevant international, regional instrument and Nigeria’s municipal laws that are meant to protect the result of the people of the Egypt Delta and their enforcement by the Nigerian State. It is quite worrisome that the Egypt Delta Region and its people have suffered and are still suffering grave violations of their right to a healthy environment as a result of petroleum exploitation in their region. The Egypt effort at best is half-hearted in its protection of the people’s right.

Keywords: crude oil, fire, floating roof tank, lightning protection systemenvironment, exploration, petroleum, pollutionDuvernay petroleum system, oil generation, oil-source correlation, Re-Os

Procedia PDF Downloads 79
230 A Study on Economic Impacts of Entrepreneurial Firms and Self-Employment: Minority Ethnics in Putatan, Penampang, Inanam, Menggatal, Uitm, Tongod, Sabah, Malaysia

Authors: Lizinis Cassendra Frederick Dony, Jirom Jeremy Frederick Dony, Andrew Nicholas, Dewi Binti Tajuddin

Abstract:

Starting and surviving a business is influenced by various entrepreneurship socio-economics activities. The study revealed that some of the entrepreneurs are not registered under SME but running own business as an intermediary with the private organization entrusted as “Self-Employed.” SME is known as “Small Medium Enterprise” contributes growth in Malaysia. Therefore, the entrepreneurialism business interest and entrepreneurial intention enhancing new spurring production, expanding employment opportunities, increasing productivity, promoting exports, stimulating innovation and providing new avenue in the business market place. This study has identified the unique contribution to the full understanding of complex mechanisms through entrepreneurship obstacles and education impacts on happiness and well-being to society. Moreover, “Ethnic” term has defined as a curious meaning refers to a classification of a large group of people customs implies to ancestral, racial, national, tribal, religious, linguistic and cultural origins. It is a social phenomenon.1 According to Sabah data population is amounting to 2,389,494 showed the predominant ethnic group being the Kadazan Dusun (18.4%) followed by Bajau (17.3%) and Malays (15.3%). For the year 2010, data statistic immigrants population report showed the amount to 239,765 people which cover 4% of the Sabahan’s population.2 Sabah has numerous group of talented entrepreneurs. The business environment among the minority ethnics are influenced with the business sentiment competition. The literature on ethnic entrepreneurship recognizes two main type entrepreneurships: the middleman and enclave entrepreneurs. According to Adam Smith,3 there are evidently some principles disposition to admire and maintain the distinction business rank status and cause most universal business sentiments. Due to credit barriers competition, the minority ethnics are losing the business market and since 2014, many illegal immigrants have been found to be using permits of the locals to operate businesses in Malaysia.4 The development of small business entrepreneurship among the minority ethnics in Sabah evidenced based variety of complex perception and differences concepts. The studies also confirmed the effects of heterogeneity on group decision and thinking caused partly by excessive pre-occupation with maintaining cohesiveness and the presence of cultural diversity in groups should reduce its probability.5 The researchers proposed that there are seven success determinants particularly to determine the involvement of minority ethnics comparing to the involvement of the immigrants in Sabah. Although, (SMEs) have always been considered the backbone of the economy development, the minority ethnics are often categorized it as the “second-choice.’ The study showed that illegal immigrants entrepreneur imposed a burden on Sabahan social programs as well as the prison, court and health care systems. The tension between the need for cheap labor and the impulse to protect Malaysian in Sabah workers, entrepreneurs and taxpayers, among the subjects discussed in this study. This is clearly can be advantages and disadvantages to the Sabah economic development.

Keywords: entrepreneurial firms, self-employed, immigrants, minority ethnic, economic impacts

Procedia PDF Downloads 411
229 Price Prediction Line, Investment Signals and Limit Conditions Applied for the German Financial Market

Authors: Cristian Păuna

Abstract:

In the first decades of the 21st century, in the electronic trading environment, algorithmic capital investments became the primary tool to make a profit by speculations in financial markets. A significant number of traders, private or institutional investors are participating in the capital markets every day using automated algorithms. The autonomous trading software is today a considerable part in the business intelligence system of any modern financial activity. The trading decisions and orders are made automatically by computers using different mathematical models. This paper will present one of these models called Price Prediction Line. A mathematical algorithm will be revealed to build a reliable trend line, which is the base for limit conditions and automated investment signals, the core for a computerized investment system. The paper will guide how to apply these tools to generate entry and exit investment signals, limit conditions to build a mathematical filter for the investment opportunities, and the methodology to integrate all of these in automated investment software. The paper will also present trading results obtained for the leading German financial market index with the presented methods to analyze and to compare different automated investment algorithms. It was found that a specific mathematical algorithm can be optimized and integrated into an automated trading system with good and sustained results for the leading German Market. Investment results will be compared in order to qualify the presented model. In conclusion, a 1:6.12 risk was obtained to reward ratio applying the trigonometric method to the DAX Deutscher Aktienindex on 24 months investment. These results are superior to those obtained with other similar models as this paper reveal. The general idea sustained by this paper is that the Price Prediction Line model presented is a reliable capital investment methodology that can be successfully applied to build an automated investment system with excellent results.

Keywords: algorithmic trading, automated trading systems, high-frequency trading, DAX Deutscher Aktienindex

Procedia PDF Downloads 130
228 The Impact of Artificial Intelligence on Legislations and Laws

Authors: Keroles Akram Saed Ghatas

Abstract:

The near future will bring significant changes in modern organizations and management due to the growing role of intangible assets and knowledge workers. The area of copyright, intellectual property, digital (intangible) assets and media redistribution appears to be one of the greatest challenges facing business and society in general and management sciences and organizations in particular. The proposed article examines the views and perceptions of fairness in digital media sharing among Harvard Law School's LL.M.s. Students, based on 50 qualitative interviews and 100 surveys. The researcher took an ethnographic approach to her research and entered the Harvard LL.M. in 2016. at, a Face book group that allows people to connect naturally and attend in-person and private events more easily. After listening to numerous students, the researcher conducted a quantitative survey among 100 respondents to assess respondents' perceptions of fairness in digital file sharing in various contexts (based on media price, its availability, regional licenses, copyright holder status, etc.). to understand better . .). Based on the survey results, the researcher conducted long-term, open-ended and loosely structured ethnographic interviews (50 interviews) to further deepen the understanding of the results. The most important finding of the study is that Harvard lawyers generally support digital piracy in certain contexts, despite having the best possible legal and professional knowledge. Interestingly, they are also more accepting of working for the government than the private sector. The results of this study provide a better understanding of how “fairness” is perceived by the younger generation of lawyers and pave the way for a more rational application of licensing laws.

Keywords: cognitive impairments, communication disorders, death penalty, executive function communication disorders, cognitive disorders, capital murder, executive function death penalty, egyptian law absence, justice, political cases piracy, digital sharing, perception of fairness, legal profession

Procedia PDF Downloads 64
227 Talking Back to Hollywood: Museum Representation in Popular Culture as a Gateway to Understanding Public Perception

Authors: Jessica BrodeFrank, Beka Bryer, Lacey Wilson, Sierra Van Ryck deGroot

Abstract:

Museums are enjoying quite the moment in pop culture. From discussions of labor in Bob’s Burger to introducing cultural repatriation in The Black Panther, discussions of various museum issues are making their way to popular media. “Talking Back to Hollywood” analyzes the impact museums have on movies and television. The paper will highlight a series of cultural cameos and discuss what each reveals about critical themes in museums: repatriation, labor, obfuscated histories, institutional legacies, artificial intelligence, and holograms. Using a mixed methods approach to include surveys, descriptive research, thematic analysis, and context analysis, the authors of this paper will explore how we, as the museum staff, might begin to cite museums and movies together as texts. Drawing from their experience working in museums and public history, this contingent of mid-career professionals will highlight the impact museums have had on movies and television and the didactic lessons these portrayals can provide back to cultural heritage professionals. From tackling critical themes in museums such as repatriation, labor conditions/inequities, obfuscated histories, curatorial choice and control, institutional legacies, and more, this paper is grounded in the cultural zeitgeist of the 2000s and the message these media portrayals send to the public and the cultural heritage sector. In particular, the paper will examine how portrayals of AI, holograms, and more technology can be used as entry points for necessary discussions with the public on mistrust, misinformation, and emerging technologies. This paper will not only expose the legacy and cultural understanding of the museum field within popular culture but also will discuss actionable ways that public historians can use these portrayals as an entry point for discussions with the public, citing literature reviews and quantitative and qualitative analysis of survey results. As Hollywood is talking about museums, museums can use that to better connect to the audiences who feel comfortable at the cinema but are excluded from the museum.

Keywords: museums, public memory, representation, popular culture

Procedia PDF Downloads 83
226 Teachers’ Protective Factors of Resilience Scale: Factorial Structure, Validity and Reliability Issues

Authors: Athena Daniilidou, Maria Platsidou

Abstract:

Recently developed scales addressed -specifically- teachers’ resilience. Although they profited from the field, they do not include some of the critical protective factors of teachers’ resilience identified in the literature. To address this limitation, we aimed at designing a more comprehensive scale for measuring teachers' resilience which encompasses various personal and environmental protective factors. To this end, two studies were carried out. In Study 1, 407 primary school teachers were tested with the new scale, the Teachers’ Protective Factors of Resilience Scale (TPFRS). Similar scales, such as the Multidimensional Teachers’ Resilience Scale and the Teachers’ Resilience Scale), were used to test the convergent validity, while the Maslach Burnout Inventory and the Teachers’ Sense of Efficacy Scale was used to assess the discriminant validity of the new scale. The factorial structure of the TPFRS was checked with confirmatory factor analysis and a good fit of the model to the data was found. Next, item response theory analysis using a two-parameter model (2PL) was applied to check the items within each factor. It revealed that 9 items did not fit the corresponding factors well and they were removed. The final version of the TPFRS includes 29 items, which assess six protective factors of teachers’ resilience: values and beliefs (5 items, α=.88), emotional and behavioral adequacy (6 items, α=.74), physical well-being (3 items, α=.68), relationships within the school environment, (6 items, α=.73) relationships outside the school environment (5 items, α=.84), and the legislative framework of education (4 items, α=.83). Results show that it presents a satisfactory convergent and discriminant validity. Study 2, in which 964 primary and secondary school teachers were tested, confirmed the factorial structure of the TPFRS as well as its discriminant validity, which was tested with the Schutte Emotional Intelligence Scale-Short Form. In conclusion, our results confirmed that the TPFRS is a valid instrument for assessing teachers' protective factors of resilience and it can be safely used in future research and interventions in the teaching profession. In conclusion, our results showed that the TPFRS is a new multi-dimensional instrument valid for assessing teachers' protective factors of resilience and it can be safely used in future research and interventions in the teaching profession.

Keywords: resilience, protective factors, teachers, item response theory

Procedia PDF Downloads 99
225 Corpus-Based Neural Machine Translation: Empirical Study Multilingual Corpus for Machine Translation of Opaque Idioms - Cloud AutoML Platform

Authors: Khadija Refouh

Abstract:

Culture bound-expressions have been a bottleneck for Natural Language Processing (NLP) and comprehension, especially in the case of machine translation (MT). In the last decade, the field of machine translation has greatly advanced. Neural machine translation NMT has recently achieved considerable development in the quality of translation that outperformed previous traditional translation systems in many language pairs. Neural machine translation NMT is an Artificial Intelligence AI and deep neural networks applied to language processing. Despite this development, there remain some serious challenges that face neural machine translation NMT when translating culture bounded-expressions, especially for low resources language pairs such as Arabic-English and Arabic-French, which is not the case with well-established language pairs such as English-French. Machine translation of opaque idioms from English into French are likely to be more accurate than translating them from English into Arabic. For example, Google Translate Application translated the sentence “What a bad weather! It runs cats and dogs.” to “يا له من طقس سيء! تمطر القطط والكلاب” into the target language Arabic which is an inaccurate literal translation. The translation of the same sentence into the target language French was “Quel mauvais temps! Il pleut des cordes.” where Google Translate Application used the accurate French corresponding idioms. This paper aims to perform NMT experiments towards better translation of opaque idioms using high quality clean multilingual corpus. This Corpus will be collected analytically from human generated idiom translation. AutoML translation, a Google Neural Machine Translation Platform, is used as a custom translation model to improve the translation of opaque idioms. The automatic evaluation of the custom model will be compared to the Google NMT using Bilingual Evaluation Understudy Score BLEU. BLEU is an algorithm for evaluating the quality of text which has been machine-translated from one natural language to another. Human evaluation is integrated to test the reliability of the Blue Score. The researcher will examine syntactical, lexical, and semantic features using Halliday's functional theory.

Keywords: multilingual corpora, natural language processing (NLP), neural machine translation (NMT), opaque idioms

Procedia PDF Downloads 149
224 The Underground Ecosystem of Credit Card Frauds

Authors: Abhinav Singh

Abstract:

Point Of Sale (POS) malwares have been stealing the limelight this year. They have been the elemental factor in some of the biggest breaches uncovered in past couple of years. Some of them include • Target: A Retail Giant reported close to 40 million credit card data being stolen • Home Depot : A home product Retailer reported breach of close to 50 million credit records • Kmart: A US retailer recently announced breach of 800 thousand credit card details. Alone in 2014, there have been reports of over 15 major breaches of payment systems around the globe. Memory scrapping malwares infecting the point of sale devices have been the lethal weapon used in these attacks. These malwares are capable of reading the payment information from the payment device memory before they are being encrypted. Later on these malwares send the stolen details to its parent server. These malwares are capable of recording all the critical payment information like the card number, security number, owner etc. All these information are delivered in raw format. This Talk will cover the aspects of what happens after these details have been sent to the malware authors. The entire ecosystem of credit card frauds can be broadly classified into these three steps: • Purchase of raw details and dumps • Converting them to plastic cash/cards • Shop! Shop! Shop! The focus of this talk will be on the above mentioned points and how they form an organized network of cyber-crime. The first step involves buying and selling of the stolen details. The key point to emphasize are : • How is this raw information been sold in the underground market • The buyer and seller anatomy • Building your shopping cart and preferences • The importance of reputation and vouches • Customer support and replace/refunds These are some of the key points that will be discussed. But the story doesn’t end here. As of now the buyer only has the raw card information. How will this raw information be converted to plastic cash? Now comes in picture the second part of this underground economy where-in these raw details are converted into actual cards. There are well organized services running underground that can help you in converting these details into plastic cards. We will discuss about this technique in detail. At last, the final step involves shopping with the stolen cards. The cards generated with the stolen details can be easily used to swipe-and-pay for purchased goods at different retail shops. Usually these purchases are of expensive items that have good resale value. Apart from using the cards at stores, there are underground services that lets you deliver online orders to their dummy addresses. Once the package is received it will be delivered to the original buyer. These services charge based on the value of item that is being delivered. The overall underground ecosystem of credit card fraud works in a bulletproof way and it involves people working in close groups and making heavy profits. This is a brief summary of what I plan to present at the talk. I have done an extensive research and have collected good deal of material to present as samples. Some of them include: • List of underground forums • Credit card dumps • IRC chats among these groups • Personal chat with big card sellers • Inside view of these forum owners. The talk will be concluded by throwing light on how these breaches are being tracked during investigation. How are credit card breaches tracked down and what steps can financial institutions can build an incidence response over it.

Keywords: POS mawalre, credit card frauds, enterprise security, underground ecosystem

Procedia PDF Downloads 439