Search results for: interactive architecture
1354 Environmental Online Campaigns Through Website Interactivity: The Case of Malaysia Environmental NGOs (MENGO)
Authors: Mohd Fadzil Mohd Idris, Aida Nasirah Abdullah, Kalthom Husain, Hanipah Hussin
Abstract:
Online campaigns reflect all the advantages; namely speed, low cost, accessibility, customization, interactivity, and persuasive ability over other media channels. Normally via websites, expensive campaigns could be done not only faster and cheaper, but also successfully. Web interactivity seems to be highly beneficial to ENGOs in advocating environmental campaigns and trigger interaction. This paper looks into the environmental online campaigns through websites of the environmental NGOs in Malaysia (MENGO); particularly on how is web interactivity structured and employed by the selected the MENGO to conduct campaigns on important issues and encourage dialogue among the audience. In this study, a quantitative method for website content analysis was conducted to investigate the availability of the coded units and to determine on which level(s) the units were placed. Twelve (12) interactivity features were coded, including the placement of units of analysis for interactivity category as units of analysis until the fourth level (Level 0-Level 3). The result demonstrates how the MENGO do not effectively structure and employ the web interactivity to conduct campaigns on important issues and encourage dialogue among the audience. It is suggested that the MENGO should redevelop the interactive website in order to effectively advocate environmental campaigns on important issues and encourage dialogue among the audience.Keywords: environmental NGOs (ENGO), Malaysia environmental NGOs (MENGO), internet, website, online campaigns, web interactivity
Procedia PDF Downloads 4401353 Governance and Public Policy: The Perception of Efficiency and Equility in Brazil and South Africa
Authors: Paulino V. Tavares, Ana L. Romao
Abstract:
Public governance represents an articulated arrangement, dynamic and interactive, present in the exercise of authority aimed at strengthening the decision-making procedure in public administration with transparency, accountability, responsiveness and capable of to emerge control and social empowerment, to pursue and achieve the objectives efficiently and with the effectiveness desired by the collective, respecting laws and providing social, institutional and economic equility in society. With this, using a multidimensional approach with the application of three questionnaires to a universe of twenty Counselors of the Courts of Auditors (Brazil), twenty professionals of public administration (Brazil), twenty Government/Provincial Counselors (South Africa), and twenty South African professionals of public administration, the present work aims to capture what is the perception about the efficiency and equility of public policies in Brazil and South Africa. With this, up until now, 30 responses have been obtained, and the results indicate that, in Brazil, 65% affirm due to the inefficiency of public policies, 70% point out that they do not believe in the equility of these same policies. In South Africa, the results indicate that 45% believe in government efficiency, and, with regard to the equility of public policies, 65% do not believe. In Brazil, the research reveals at least three reasons for this result, that is, lack of planning, lack of clear objectives of public policies, and lack of information on the part of society, while in South Africa, so far, research has not identified a specific reason for this result.Keywords: efficiency, equility, governance, public policy
Procedia PDF Downloads 1251352 Facial Emotion Recognition Using Deep Learning
Authors: Ashutosh Mishra, Nikhil Goyal
Abstract:
A 3D facial emotion recognition model based on deep learning is proposed in this paper. Two convolution layers and a pooling layer are employed in the deep learning architecture. After the convolution process, the pooling is finished. The probabilities for various classes of human faces are calculated using the sigmoid activation function. To verify the efficiency of deep learning-based systems, a set of faces. The Kaggle dataset is used to verify the accuracy of a deep learning-based face recognition model. The model's accuracy is about 65 percent, which is lower than that of other facial expression recognition techniques. Despite significant gains in representation precision due to the nonlinearity of profound image representations.Keywords: facial recognition, computational intelligence, convolutional neural network, depth map
Procedia PDF Downloads 2311351 Challenges of Implementing Zero Trust Security Based on NIST SP 800-207
Authors: Mazhar Hamayun
Abstract:
Organizations need to take a holistic approach to their Zero Trust strategic and tactical security needs. This includes using a framework-agnostic model that will ensure all enterprise resources are being accessed securely, regardless of their location. Such can be achieved through the implementation of a security posture, monitoring the posture, and adjusting the posture through the Identify, Detect, Protect, Respond, and Recover Methods, The target audience of this document includes those involved in the management and operational functions of risk, information security, and information technology. This audience consists of the chief information security officer, chief information officer, chief technology officer, and those leading digital transformation initiatives where Zero Trust methods can help protect an organization’s data assets.Keywords: ZTNA, zerotrust architecture, microsegmentation, NIST SP 800-207
Procedia PDF Downloads 861350 Queer Social Realism and Architecture in British Cinema: Tenement Housing, Unions and the Affective Body
Authors: Christopher Pullen
Abstract:
This paper explores the significance of British cinema in the late 1950s and early 1960s as offering a renaissance of realist discourse, in the representation of everyday social issues. Offering a rejection of Hollywood cinema and the superficially of the middle classes, these ‘kitchen sink dramas’ often set within modest and sometimes squalid domestic and social environments, focused on the political struggle of the disenfranchised examining poverty, the oppressed and the outsider. While films like Look Back in Anger and Room at the Top looked primarily at male heterosexual subjectivity, films like A Taste of Honey and Victim focused on female and queer male narratives. Framing the urban landscape as a discursive architectural arena, representing basic living conditions and threatening social worlds, these iconic films established new storytelling processes for the outsider. This paper examines this historical context foregrounding the contemporary films Beautiful Thing (Hettie Macdonald, 1996), Weekend (Andrew Haigh, 2011) and Pride (Marcus Warchus, 2014), while employing the process of textual analysis in relation to theories of affect, defined by writers such as Lisa U. Marks and Sara Ahmed. Considering both romance narratives and public demonstrations of unity, where the queer ‘affective’ body is placed within architectural and social space, Beautiful Thing tells the story of gay male teenagers falling in love despite oppression from family and school, Weekend examines a one-night stand between young gay men and the unlikeliness of commitment, but the drive for sensitivity, and Pride foregrounds an historical relationship between queer youth activists and the miner’s union, who were on strike between 1984-5. These films frame the queer ‘affective’ body within politicized public space, evident in lower class men’s working clubs, tenement housing and brutal modernist tower blocks, focusing on architectural features such as windows, doorways and staircases, relating temporality, desire and change. Through such an examination a hidden history of gay male performativity is revealed, framing the potential of contemporary cinema to focus on the context of the outsider in encouraging social change.Keywords: queer, affect, cinema, architecture, life chances
Procedia PDF Downloads 3581349 E-Service and the Nigerian Banking Sector: A Review of ATM Architecture and Operations
Authors: Bashir Aliyu Yauri, Rufai Aliyu Yauri
Abstract:
With the introduction of cash-less society policy by the Central Bank of Nigeria, the concept of e-banking services has experienced a significant improvement over the years. Today quite a number of people are embracing e-banking activities especially ATM, thereby moving away from the conventional banking system. This paper presents a review of the underlying Architectural Layout of Intra-Bank and Inter-Bank ATM connectivity in Nigeria. The paper further investigates and discusses factors affecting the Intra-Bank and Inter-Bank ATM connectivity in Nigeria. And as well possible solutions to these factors affecting ATM Connectivity and Operations are proposed.Keywords: architectural layout, automated teller machine, e-services, postilion
Procedia PDF Downloads 6331348 Complex Event Processing System Based on the Extended ECA Rule
Authors: Kwan Hee Han, Jun Woo Lee, Sung Moon Bae, Twae Kyung Park
Abstract:
ECA (Event-Condition-Action) languages are largely adopted for event processing since they are an intuitive and powerful paradigm for programming reactive systems. However, there are some limitations about ECA rules for processing of complex events such as coupling of event producer and consumer. The objective of this paper is to propose an ECA rule pattern to improve the current limitations of ECA rule, and to develop a prototype system. In this paper, conventional ECA rule is separated into 3 parts and each part is extended to meet the requirements of CEP. Finally, event processing logic is established by combining the relevant elements of 3 parts. The usability of proposed extended ECA rule is validated by a test scenario in this study.Keywords: complex event processing, ECA rule, Event processing system, event-driven architecture, internet of things
Procedia PDF Downloads 5301347 Security Issues in Long Term Evolution-Based Vehicle-To-Everything Communication Networks
Authors: Mujahid Muhammad, Paul Kearney, Adel Aneiba
Abstract:
The ability for vehicles to communicate with other vehicles (V2V), the physical (V2I) and network (V2N) infrastructures, pedestrians (V2P), etc. – collectively known as V2X (Vehicle to Everything) – will enable a broad and growing set of applications and services within the intelligent transport domain for improving road safety, alleviate traffic congestion and support autonomous driving. The telecommunication research and industry communities and standardization bodies (notably 3GPP) has finally approved in Release 14, cellular communications connectivity to support V2X communication (known as LTE – V2X). LTE – V2X system will combine simultaneous connectivity across existing LTE network infrastructures via LTE-Uu interface and direct device-to-device (D2D) communications. In order for V2X services to function effectively, a robust security mechanism is needed to ensure legal and safe interaction among authenticated V2X entities in the LTE-based V2X architecture. The characteristics of vehicular networks, and the nature of most V2X applications, which involve human safety makes it significant to protect V2X messages from attacks that can result in catastrophically wrong decisions/actions include ones affecting road safety. Attack vectors include impersonation attacks, modification, masquerading, replay, MiM attacks, and Sybil attacks. In this paper, we focus our attention on LTE-based V2X security and access control mechanisms. The current LTE-A security framework provides its own access authentication scheme, the AKA protocol for mutual authentication and other essential cryptographic operations between UEs and the network. V2N systems can leverage this protocol to achieve mutual authentication between vehicles and the mobile core network. However, this protocol experiences technical challenges, such as high signaling overhead, lack of synchronization, handover delay and potential control plane signaling overloads, as well as privacy preservation issues, which cannot satisfy the adequate security requirements for majority of LTE-based V2X services. This paper examines these challenges and points to possible ways by which they can be addressed. One possible solution, is the implementation of the distributed peer-to-peer LTE security mechanism based on the Bitcoin/Namecoin framework, to allow for security operations with minimal overhead cost, which is desirable for V2X services. The proposed architecture can ensure fast, secure and robust V2X services under LTE network while meeting V2X security requirements.Keywords: authentication, long term evolution, security, vehicle-to-everything
Procedia PDF Downloads 1671346 Specification of Requirements to Ensure Proper Implementation of Security Policies in Cloud-Based Multi-Tenant Systems
Authors: Rebecca Zahra, Joseph G. Vella, Ernest Cachia
Abstract:
The notion of cloud computing is rapidly gaining ground in the IT industry and is appealing mostly due to making computing more adaptable and expedient whilst diminishing the total cost of ownership. This paper focuses on the software as a service (SaaS) architecture of cloud computing which is used for the outsourcing of databases with their associated business processes. One approach for offering SaaS is basing the system’s architecture on multi-tenancy. Multi-tenancy allows multiple tenants (users) to make use of the same single application instance. Their requests and configurations might then differ according to specific requirements met through tenant customisation through the software. Despite the known advantages, companies still feel uneasy to opt for the multi-tenancy with data security being a principle concern. The fact that multiple tenants, possibly competitors, would have their data located on the same server process and share the same database tables heighten the fear of unauthorised access. Security is a vital aspect which needs to be considered by application developers, database administrators, data owners and end users. This is further complicated in cloud-based multi-tenant system where boundaries must be established between tenants and additional access control models must be in place to prevent unauthorised cross-tenant access to data. Moreover, when altering the database state, the transactions need to strictly adhere to the tenant’s known business processes. This paper focuses on the fact that security in cloud databases should not be considered as an isolated issue. Rather it should be included in the initial phases of the database design and monitored continuously throughout the whole development process. This paper aims to identify a number of the most common security risks and threats specifically in the area of multi-tenant cloud systems. Issues and bottlenecks relating to security risks in cloud databases are surveyed. Some techniques which might be utilised to overcome them are then listed and evaluated. After a description and evaluation of the main security threats, this paper produces a list of software requirements to ensure that proper security policies are implemented by a software development team when designing and implementing a multi-tenant based SaaS. This would then assist the cloud service providers to define, implement, and manage security policies as per tenant customisation requirements whilst assuring security for the customers’ data.Keywords: cloud computing, data management, multi-tenancy, requirements, security
Procedia PDF Downloads 1561345 Managing Cognitive Load in Accounting: An Analysis of Three Instructional Designs in Financial Accounting
Authors: Seedwell Sithole
Abstract:
One of the persistent problems in accounting education is how to effectively support students’ learning. A promising technique to this issue is to investigate the extent that learning is determined by the design of instructional material. This study examines the academic performance of students using three instructional designs in financial accounting. Student’s performance scores and reported mental effort ratings were used to determine the instructional effectiveness. The findings of this study show that accounting students prefer graph and text designs that are integrated. The results suggest that spatially separated graph and text presentations in accounting should be reorganized to align with the requirements of human cognitive architecture.Keywords: accounting, cognitive load, education, instructional preferences, students
Procedia PDF Downloads 1511344 An Informative Marketing Platform: Methodology and Architecture
Authors: Martina Marinelli, Samanta Vellante, Francesco Pilotti, Daniele Di Valerio, Gaetanino Paolone
Abstract:
Any development in web marketing technology requires changes in information engineering to identify instruments and techniques suitable for the production of software applications for informative marketing. Moreover, for large web solutions, designing an interface that enables human interactions is a complex process that must bridge between informative marketing requirements and the developed solution. A user-friendly interface in web marketing applications is crucial for a successful business. The paper introduces mkInfo - a software platform that implements informative marketing. Informative marketing is a new interpretation of marketing which places the information at the center of every marketing action. The creative team includes software engineering researchers who have recently authored an article on automatic code generation. The authors have created the mkInfo software platform to generate informative marketing web applications. For each web application, it is possible to automatically implement an opt in page, a landing page, a sales page, and a thank you page: one only needs to insert the content. mkInfo implements an autoresponder to send mail according to a predetermined schedule. The mkInfo platform also includes e-commerce for a product or service. The stakeholder can access any opt-in page and get basic information about a product or service. If he wants to know more, he will need to provide an e-mail address to access a landing page that will generate an e-mail sequence. It will provide him with complete information about the product or the service. From this point on, the stakeholder becomes a user and is now able to purchase the product or related services through the mkInfo platform. This paper suggests a possible definition for Informative Marketing, illustrates its basic principles, and finally details the mkInfo platform that implements it. This paper also offers some Informative Marketing models, which are implemented in the mkInfo platform. Informative marketing can be applied to products or services. It is necessary to realize a web application for each product or service. The mkInfo platform enables the product or the service producer to send information concerning a specific product or service to all stakeholders. In conclusion, the technical contributions of this paper are: a different interpretation of marketing based on information; a modular architecture for web applications, particularly for one with standard features such as information storage, exchange, and delivery; multiple models to implement informative marketing; a software platform enabling the implementation of such models in a web application. Future research aims to enable stakeholders to provide information about a product or a service so that the information gathered about a product or a service includes both the producer’s and the stakeholders' point of view. The purpose is to create an all-inclusive management system of the knowledge regarding a specific product or service: a system that includes everything about the product or service and is able to address even unexpected questions.Keywords: informative marketing, opt in page, software platform, web application
Procedia PDF Downloads 1271343 The Use of Indicators to Evaluate Minor Heritage Areas in a City
Authors: J. L. Oliver, T. Agryzkov, L. Tortosa, J. F. Vicent, J. Santacruz
Abstract:
This paper aims to demonstrate how a system of indicators can be used in order to evaluate some heritage areas which can be understood as minor ones. We mean by that those urban areas with high heritage interest from an academical point of view, but never properly valued. The reasons for this situation may be diverse, either they are not old enough, or they may show the modest architecture, the fact is these areas have not been considered deserving of protection, as the historical ones. As a result of this reality, they usually show now a very degraded urban space, which in addition contribute to accelerate a process of deterioration. Using a technic well known in urban design, we propose here a system of indicators for patrimonial purposes, as a tool to identify and quantify the heritage value of these kinds of areas. As a case study, we apply this system in some part of the City of Quito (El Ecuador).Keywords: heritage cities, indicators, spatial analysis, historic sites
Procedia PDF Downloads 2431342 Investigating the Relationship between Bank and Cloud Provider
Authors: Hatim Elhag
Abstract:
Banking and Financial Service Institutions are possibly the most advanced in terms of technology adoption and use it as a key differentiator. With high levels of business process automation, maturity in the functional portfolio, straight through processing and proven technology outsourcing benefits, Banking sector stand to benefit significantly from Cloud computing capabilities. Additionally, with complex Compliance and Regulatory policies, combined with expansive products and geography coverage, the business impact is even greater. While the benefits are exponential, there are also significant challenges in adopting this model– including Legal, Security, Performance, Reliability, Transformation complexity, Operating control and Governance and most importantly proof for the promised cost benefits. However, new architecture designed should be implemented to align this approach.Keywords: security, cloud, banking sector, cloud computing
Procedia PDF Downloads 4991341 The Unique Journeys from Different Pasts to Multiple Presents in the Work of the Pritzker Prize Laureates of 2010-2020
Authors: Christakis Chatzichristou Kyriakos Miltiadou, Konstantinos Gounaridis
Abstract:
The paper discusses how the Pritzker Prize Laureates of the last decade themselves identify the various ways different aspects or interpretations of the past have influenced their design methodologies. As the recipients of what is considered to be the most prestigious award in architecture, these architects are worth examining not only because of their exemplary work but also because of the strong influence they have on architectural culture in general. Rather than attempting to interpret their projects, the methodology chosen focuses on what the architects themselves have to say on the subject. The research aims at, and, as the tabular form of the findings shows, also succeeds in revealing the numerous and diverse ways different aspects of what is termed as the Past can potentially enrich contemporary design practices.Keywords: design methodology, Pritzker Prize Laureates, past, culture, tradition
Procedia PDF Downloads 441340 Research on the Evolution of Public Space in Tourism-Oriented Traditional Rural Settlements
Authors: Yu Zhang, Mingxue Lang, Li Dong
Abstract:
The hundreds of years of slow succession of living environment in rural area is a crucial carrier of China’s long history of culture and national wisdom. In recent years, the space evolution of traditional rural settlements has been promoted by the intervention of tourism development, among which the public architecture and outdoor activity areas together served as the major places for villagers, and tourists’ social activities are an important characterization for settlement spatial evolution. Traditional public space upgrade and layout study of new public space can effectively promote the tourism industry development of traditional rural settlements. This article takes Qi County, one China Traditional Culture Village as the exemplification and uses the technology of Remote Sensing (RS), Geographic Information System (GIS) and Space Syntax, studies the evolution features of public space of tourism-oriented traditional rural settlements in four steps. First, acquire the 2003 and 2016 image data of Qi County, using the remote sensing application EDRAS8.6. Second, vectorize the basic maps of Qi County including its land use map with the application of ArcGIS 9.3 meanwhile, associating with architectural and site information concluded from field research. Third, analyze the accessibility and connectivity of the inner space of settlements using space syntax; run cross-correlation with the public space data of 2003 and 2016. Finally, summarize the evolution law of the public space of settlements; study the upgrade pattern of traditional public space and location plan for new public space. Major findings of this paper including: first, location layout of traditional public space has a larger association with the calculation results of space syntax and further confirmed the objective value of space syntax in expressing the space and social relations. Second, the intervention of tourism development generates remarkable impact on public space location of tradition rural settlements. Third, traditional public space produces the symbols of both strengthening and decline and forms a diversified upgrade pattern for the purpose of meeting the different tourism functional needs. Finally, space syntax provides an objective basis for location plan of new public space that meets the needs of tourism service. Tourism development has a significant impact on the evolution of public space of traditional rural settlements. Two types of public space, architecture, and site are both with changes seen from the perspective of quantity, location, dimension and function after the intervention of tourism development. Function upgrade of traditional public space and scientific layout of new public space are two important ways in achieving the goal of sustainable development of tourism-oriented traditional rural settlements.Keywords: public space evolution, Qi county, space syntax, tourism oriented, traditional rural settlements
Procedia PDF Downloads 3411339 Hydrogen: Contention-Aware Hybrid Memory Management for Heterogeneous CPU-GPU Architectures
Authors: Yiwei Li, Mingyu Gao
Abstract:
Integrating hybrid memories with heterogeneous processors could leverage heterogeneity in both compute and memory domains for better system efficiency. To ensure performance isolation, we introduce Hydrogen, a hardware architecture to optimize the allocation of hybrid memory resources to heterogeneous CPU-GPU systems. Hydrogen supports efficient capacity and bandwidth partitioning between CPUs and GPUs in both memory tiers. We propose decoupled memory channel mapping and token-based data migration throttling to enable flexible partitioning. We also support epoch-based online search for optimized configurations and lightweight reconfiguration with reduced data movements. Hydrogen significantly outperforms existing designs by 1.21x on average and up to 1.31x.Keywords: hybrid memory, heterogeneous systems, dram cache, graphics processing units
Procedia PDF Downloads 961338 On-Chip Aging Sensor Circuit Based on Phase Locked Loop Circuit
Authors: Ararat Khachatryan, Davit Mirzoyan
Abstract:
In sub micrometer technology, the aging phenomenon starts to have a significant impact on the reliability of integrated circuits by bringing performance degradation. For that reason, it is important to have a capability to evaluate the aging effects accurately. This paper presents an accurate aging measurement approach based on phase-locked loop (PLL) and voltage-controlled oscillator (VCO) circuit. The architecture is rejecting the circuit self-aging effect from the characteristics of PLL, which is generating the frequency without any aging phenomena affects. The aging monitor is implemented in low power 32 nm CMOS technology, and occupies a pretty small area. Aging simulation results show that the proposed aging measurement circuit improves accuracy by about 2.8% at high temperature and 19.6% at high voltage.Keywords: aging effect, HCI, NBTI, nanoscale
Procedia PDF Downloads 3591337 Spatial Optimization of Riverfront Street Based on Inclusive Design
Authors: Lianxue Shi
Abstract:
Riverfront street has the dual characteristics of street space and waterfront space, which is not only a vital place for residents to travel and communicate but also a high-frequency space for people's leisure and entertainment. However, under the development of cities and towns pursuing efficiency, riverfront streets appear to have a variety of problems, such as a lack of multifunctionality, insufficient facilities, and loss of characteristics, which fail to meet the needs of various groups of people, and their inclusiveness is facing a great challenge. It is, therefore, evident that the optimization of riverfront street space from an inclusivity perspective is important to the establishment of a human-centered, high-quality urban space. Therefore, this article starts by exploring the interactive relationship between inclusive design and street space. Based on the analysis of the characteristics of the riverfront street space and people's needs, it proposes the four inclusive design orientations of natural inclusion, group inclusion, spatial inclusion, and social inclusion. It then constructs a design framework for the inclusive optimization of riverfront street space, aiming to create streets that are “safe and accessible, diverse and shared, distinctive and friendly, green and sustainable”. Riverfront streets in Wansheng District, Chongqing, are selected as a practice case, and specific strategies are put forward in four aspects: the creation of an accessible slow-traffic system, the provision of diversified functional services, the reshaping of emotional bonds and the integration of ecological spaces.Keywords: inclusiveness design, riverfront street, spatial optimization, street spaces
Procedia PDF Downloads 331336 Interactive of Calcium, Potassium, and Dynamic Unequal Salt Distribution on the Growth of Tomato in Hydroponic System
Authors: Mohammad Koushafar, Amir Hossein Khoshgoftarmanesh
Abstract:
Due to water shortage, application of saline water for irrigation is an urgent requirement in agriculture. Thus, this study, the effect of calcium and potassium application as additive in saline root media for reduce salinity adverse effects was investigated on tomato growth in a hydroponic system with unequal distribution of salts in the root media, which was divided into two equal parts containing full Johnson nutrient solution and 40 mM NaCl solution, alone or in combination with KCl (6 mM), CaCl2 (4 mM), K+Ca (3+2 mM) or half-strength Johnson nutrient solution. The root splits were exchanged every 7 days. Results showed that addition of calcium, calcium-potassium and nutrition elements equivalent to half the concentration of Johnson formula to the saline-half of culture media minimized the reduction in plant growth caused by NaCl, although the addition of potassium to culture media was not effective. The greatest concentration of sodium was observed at the shoot of treatments which had the smallest growth. According to the results of this study, in the case of dynamic and non-uniform distribution of salts in the root media, by the addition of additive to the saline solution, it would be possible to use of saline water with no significant growth reduction.Keywords: calcium, hydroponic, local salinity, potassium, salin water, tomato
Procedia PDF Downloads 4431335 Enhance Engineering Learning Using Cognitive Simulator
Authors: Lior Davidovitch
Abstract:
Traditional training based on static models and case studies is the backbone of most teaching and training programs of engineering education. However, project management learning is characterized by dynamics models that requires new and enhanced learning method. The results of empirical experiments evaluating the effectiveness and efficiency of using cognitive simulator as a new training technique are reported. The empirical findings are focused on the impact of keeping and reviewing learning history in a dynamic and interactive simulation environment of engineering education. The cognitive simulator for engineering project management learning had two learning history keeping modes: manual (student-controlled), automatic (simulator-controlled) and a version with no history keeping. A group of industrial engineering students performed four simulation-runs divided into three identical simple scenarios and one complicated scenario. The performances of participants running the simulation with the manual history mode were significantly better than users running the simulation with the automatic history mode. Moreover, the effects of using the undo enhanced further the learning process. The findings indicate an enhancement of engineering students’ learning and decision making when they use the record functionality of the history during their engineering training process. Furthermore, the cognitive simulator as educational innovation improves students learning and training. The practical implications of using simulators in the field of engineering education are discussed.Keywords: cognitive simulator, decision making, engineering learning, project management
Procedia PDF Downloads 2491334 A Psychophysiological Evaluation of an Effective Recognition Technique Using Interactive Dynamic Virtual Environments
Authors: Mohammadhossein Moghimi, Robert Stone, Pia Rotshtein
Abstract:
Recording psychological and physiological correlates of human performance within virtual environments and interpreting their impacts on human engagement, ‘immersion’ and related emotional or ‘effective’ states is both academically and technologically challenging. By exposing participants to an effective, real-time (game-like) virtual environment, designed and evaluated in an earlier study, a psychophysiological database containing the EEG, GSR and Heart Rate of 30 male and female gamers, exposed to 10 games, was constructed. Some 174 features were subsequently identified and extracted from a number of windows, with 28 different timing lengths (e.g. 2, 3, 5, etc. seconds). After reducing the number of features to 30, using a feature selection technique, K-Nearest Neighbour (KNN) and Support Vector Machine (SVM) methods were subsequently employed for the classification process. The classifiers categorised the psychophysiological database into four effective clusters (defined based on a 3-dimensional space – valence, arousal and dominance) and eight emotion labels (relaxed, content, happy, excited, angry, afraid, sad, and bored). The KNN and SVM classifiers achieved average cross-validation accuracies of 97.01% (±1.3%) and 92.84% (±3.67%), respectively. However, no significant differences were found in the classification process based on effective clusters or emotion labels.Keywords: virtual reality, effective computing, effective VR, emotion-based effective physiological database
Procedia PDF Downloads 2331333 FisherONE: Employing Distinct Pedagogy through Technology Integration in Senior Secondary Education
Authors: J. Kontoleon, D.Gall, M.Pidskalny
Abstract:
FisherONE offers a distinct pedagogic model for senior secondary education that integrates advanced technology to meet the learning needs of Year 11 and 12 students across Catholic schools in Queensland. As a fully online platform, FisherONE employs pedagogy that combines flexibility with personalized, data-driven learning. The model leverages tools like the MaxHub hybrid interactive system and AI-powered learning assistants to create tailored learning pathways that promote student autonomy and engagement. This paper examines FisherONE’s success in employing pedagogic strategies through technology. Initial findings suggest that students benefit from the blended approach of virtual assessments and real-time support, even as AI-assisted tools remain in the proof-of-concept phase. The study outlines how FisherONE plans to continue refining its educational methods to better serve students in distance learning environments, specifically in challenging subjects like physics. The integration of technology in FisherONE enhances the effectiveness of teaching and learning, addressing common challenges in online education by offering scalable, individualized learning experiences. This approach demonstrates the future potential of technology in education and the role it can play in fostering meaningful student outcomes.Keywords: AI-assisted learning, innovative pedagogy, personalized learning, senior education, technology in education
Procedia PDF Downloads 181332 The Voyage of Adolfo Caminha to the USA: The Discourse about Americanism in Brazil in the Late Nineteenth Century
Authors: Maxwel F. Silva, José Geraldo Pedrosa
Abstract:
This paper is a study about the voyage of Adolfo Caminha to the USA in the late nineteenth century described in “No país dos ianques”. The hypothesis is that the USA constitutes a civilizing reference that moves away from Europe. The Americanism expression it means that the Yankees have invented a new repertoire through which built a new idea of civilization. The base is European, but your architecture is new. This paper is not concerned with the meanings and uses of the Americanism expression among the Yankees, but with the ways in which the America were understood by otherness, especially in the late nineteenth and early twentieth century. In this way, this study discusses the concept of Americanism in the thought of Adolfo Caminha and it is relation with the Brazil in the late nineteenth century, especially in questions about democracy, liberty and progress.Keywords: Adolfo Caminha, Americanism, discourse, voyage
Procedia PDF Downloads 5521331 Analysis and Design Modeling for Next Generation Network Intrusion Detection and Prevention System
Authors: Nareshkumar Harale, B. B. Meshram
Abstract:
The continued exponential growth of successful cyber intrusions against today’s businesses has made it abundantly clear that traditional perimeter security measures are no longer adequate and effective. We evolved the network trust architecture from trust-untrust to Zero-Trust, With Zero Trust, essential security capabilities are deployed in a way that provides policy enforcement and protection for all users, devices, applications, data resources, and the communications traffic between them, regardless of their location. Information exchange over the Internet, in spite of inclusion of advanced security controls, is always under innovative, inventive and prone to cyberattacks. TCP/IP protocol stack, the adapted standard for communication over network, suffers from inherent design vulnerabilities such as communication and session management protocols, routing protocols and security protocols are the major cause of major attacks. With the explosion of cyber security threats, such as viruses, worms, rootkits, malwares, Denial of Service attacks, accomplishing efficient and effective intrusion detection and prevention is become crucial and challenging too. In this paper, we propose a design and analysis model for next generation network intrusion detection and protection system as part of layered security strategy. The proposed system design provides intrusion detection for wide range of attacks with layered architecture and framework. The proposed network intrusion classification framework deals with cyberattacks on standard TCP/IP protocol, routing protocols and security protocols. It thereby forms the basis for detection of attack classes and applies signature based matching for known cyberattacks and data mining based machine learning approaches for unknown cyberattacks. Our proposed implemented software can effectively detect attacks even when malicious connections are hidden within normal events. The unsupervised learning algorithm applied to network audit data trails results in unknown intrusion detection. Association rule mining algorithms generate new rules from collected audit trail data resulting in increased intrusion prevention though integrated firewall systems. Intrusion response mechanisms can be initiated in real-time thereby minimizing the impact of network intrusions. Finally, we have shown that our approach can be validated and how the analysis results can be used for detecting and protection from the new network anomalies.Keywords: network intrusion detection, network intrusion prevention, association rule mining, system analysis and design
Procedia PDF Downloads 2271330 The Quantitative Analysis of the Traditional Rural Settlement Plane Boundary
Authors: Yifan Dong, Xincheng Pu
Abstract:
Rural settlements originate from the accumulation of residential building elements, and their agglomeration forms the settlement pattern and defines the relationship between the settlement and the inside and outside. The settlement boundary is an important part of the settlement pattern. Compared with the simplification of the urban settlement boundary, the settlement of the country is more complex, fuzzy and uncertain, and then presents a rich and diverse boundary morphological phenomenon. In this paper, China traditional rural settlements plane boundary as the research object, using fractal theory and fractal dimension method, quantitative analysis of planar shape boundary settlement, and expounds the research for the architectural design, ancient architecture protection and renewal and development and the significance of the protection of settlements.Keywords: rural settlement, border, fractal, quantification
Procedia PDF Downloads 2481329 Data Analysis Tool for Predicting Water Scarcity in Industry
Authors: Tassadit Issaadi Hamitouche, Nicolas Gillard, Jean Petit, Valerie Lavaste, Celine Mayousse
Abstract:
Water is a fundamental resource for the industry. It is taken from the environment either from municipal distribution networks or from various natural water sources such as the sea, ocean, rivers, aquifers, etc. Once used, water is discharged into the environment, reprocessed at the plant or treatment plants. These withdrawals and discharges have a direct impact on natural water resources. These impacts can apply to the quantity of water available, the quality of the water used, or to impacts that are more complex to measure and less direct, such as the health of the population downstream from the watercourse, for example. Based on the analysis of data (meteorological, river characteristics, physicochemical substances), we wish to predict water stress episodes and anticipate prefectoral decrees, which can impact the performance of plants and propose improvement solutions, help industrialists in their choice of location for a new plant, visualize possible interactions between companies to optimize exchanges and encourage the pooling of water treatment solutions, and set up circular economies around the issue of water. The development of a system for the collection, processing, and use of data related to water resources requires the functional constraints specific to the latter to be made explicit. Thus the system will have to be able to store a large amount of data from sensors (which is the main type of data in plants and their environment). In addition, manufacturers need to have 'near-real-time' processing of information in order to be able to make the best decisions (to be rapidly notified of an event that would have a significant impact on water resources). Finally, the visualization of data must be adapted to its temporal and geographical dimensions. In this study, we set up an infrastructure centered on the TICK application stack (for Telegraf, InfluxDB, Chronograf, and Kapacitor), which is a set of loosely coupled but tightly integrated open source projects designed to manage huge amounts of time-stamped information. The software architecture is coupled with the cross-industry standard process for data mining (CRISP-DM) data mining methodology. The robust architecture and the methodology used have demonstrated their effectiveness on the study case of learning the level of a river with a 7-day horizon. The management of water and the activities within the plants -which depend on this resource- should be considerably improved thanks, on the one hand, to the learning that allows the anticipation of periods of water stress, and on the other hand, to the information system that is able to warn decision-makers with alerts created from the formalization of prefectoral decrees.Keywords: data mining, industry, machine Learning, shortage, water resources
Procedia PDF Downloads 1211328 Automatic Aggregation and Embedding of Microservices for Optimized Deployments
Authors: Pablo Chico De Guzman, Cesar Sanchez
Abstract:
Microservices are a software development methodology in which applications are built by composing a set of independently deploy-able, small, modular services. Each service runs a unique process and it gets instantiated and deployed in one or more machines (we assume that different microservices are deployed into different machines). Microservices are becoming the de facto standard for developing distributed cloud applications due to their reduced release cycles. In principle, the responsibility of a microservice can be as simple as implementing a single function, which can lead to the following issues: - Resource fragmentation due to the virtual machine boundary. - Poor communication performance between microservices. Two composition techniques can be used to optimize resource fragmentation and communication performance: aggregation and embedding of microservices. Aggregation allows the deployment of a set of microservices on the same machine using a proxy server. Aggregation helps to reduce resource fragmentation, and is particularly useful when the aggregated services have a similar scalability behavior. Embedding deals with communication performance by deploying on the same virtual machine those microservices that require a communication channel (localhost bandwidth is reported to be about 40 times faster than cloud vendor local networks and it offers better reliability). Embedding can also reduce dependencies on load balancer services since the communication takes place on a single virtual machine. For example, assume that microservice A has two instances, a1 and a2, and it communicates with microservice B, which also has two instances, b1 and b2. One embedding can deploy a1 and b1 on machine m1, and a2 and b2 are deployed on a different machine m2. This deployment configuration allows each pair (a1-b1), (a2-b2) to communicate using the localhost interface without the need of a load balancer between microservices A and B. Aggregation and embedding techniques are complex since different microservices might have incompatible runtime dependencies which forbid them from being installed on the same machine. There is also a security concern since the attack surface between microservices can be larger. Luckily, container technology allows to run several processes on the same machine in an isolated manner, solving the incompatibility of running dependencies and the previous security concern, thus greatly simplifying aggregation/embedding implementations by just deploying a microservice container on the same machine as the aggregated/embedded microservice container. Therefore, a wide variety of deployment configurations can be described by combining aggregation and embedding to create an efficient and robust microservice architecture. This paper presents a formal method that receives a declarative definition of a microservice architecture and proposes different optimized deployment configurations by aggregating/embedding microservices. The first prototype is based on i2kit, a deployment tool also submitted to ICWS 2018. The proposed prototype optimizes the following parameters: network/system performance, resource usage, resource costs and failure tolerance.Keywords: aggregation, deployment, embedding, resource allocation
Procedia PDF Downloads 2031327 The Studies of Client Requirements in Home Stay: A Case Study of Thailand
Authors: Kanamon Suwantada
Abstract:
The purpose of this research is to understand customer’s expectations towards homestays and to establish the precise strategies to increase numbers of tourists for homestay business in Amphawa district, Samutsongkram, Thailand. The researcher aims to ensure that each host provides experiences to travelers who are looking for and determining new targets for homestay business in Amphawa as well as creating sustainable homestay using marketing strategies to increase customers. The methods allow interview and questionnaire to gain both overview data from the tourists and qualitative data from the homestay owner’s perspective to create a GAP analysis. The data was collected from 200 tourists, during 15th May - 30th July, 2011 from homestay in Amphawa Community. The questionnaires were divided into three sections: the demographic profile, customer information and influencing on purchasing position, and customer expectation towards homestay. The analysis, in fact, will be divided into two methods which are percentage and correlation analyses. The result of this research revealed that homestay had already provided customers with reasonable prices in good locations. Antithetically, activities that they offered still could not have met the customer’s requirements. Homestay providers should prepare additional activities such as village tour, local attraction tour, village daily life experiences, local ceremony participation, and interactive conversation with local people. Moreover, the results indicated that a price was the most important factor for choosing homestay.Keywords: ecotourism, homestay, marketing, sufficiency economic philosophy
Procedia PDF Downloads 3101326 Enhancing Athlete Training using Real Time Pose Estimation with Neural Networks
Authors: Jeh Patel, Chandrahas Paidi, Ahmed Hambaba
Abstract:
Traditional methods for analyzing athlete movement often lack the detail and immediacy required for optimal training. This project aims to address this limitation by developing a Real-time human pose estimation system specifically designed to enhance athlete training across various sports. This system leverages the power of convolutional neural networks (CNNs) to provide a comprehensive and immediate analysis of an athlete’s movement patterns during training sessions. The core architecture utilizes dilated convolutions to capture crucial long-range dependencies within video frames. Combining this with the robust encoder-decoder architecture to further refine pose estimation accuracy. This capability is essential for precise joint localization across the diverse range of athletic poses encountered in different sports. Furthermore, by quantifying movement efficiency, power output, and range of motion, the system provides data-driven insights that can be used to optimize training programs. Pose estimation data analysis can also be used to develop personalized training plans that target specific weaknesses identified in an athlete’s movement patterns. To overcome the limitations posed by outdoor environments, the project employs strategies such as multi-camera configurations or depth sensing techniques. These approaches can enhance pose estimation accuracy in challenging lighting and occlusion scenarios, where pose estimation accuracy in challenging lighting and occlusion scenarios. A dataset is collected From the labs of Martin Luther King at San Jose State University. The system is evaluated through a series of tests that measure its efficiency and accuracy in real-world scenarios. Results indicate a high level of precision in recognizing different poses, substantiating the potential of this technology in practical applications. Challenges such as enhancing the system’s ability to operate in varied environmental conditions and further expanding the dataset for training were identified and discussed. Future work will refine the model’s adaptability and incorporate haptic feedback to enhance the interactivity and richness of the user experience. This project demonstrates the feasibility of an advanced pose detection model and lays the groundwork for future innovations in assistive enhancement technologies.Keywords: computer vision, deep learning, human pose estimation, U-NET, CNN
Procedia PDF Downloads 551325 Deep Learning for Image Correction in Sparse-View Computed Tomography
Authors: Shubham Gogri, Lucia Florescu
Abstract:
Medical diagnosis and radiotherapy treatment planning using Computed Tomography (CT) rely on the quantitative accuracy and quality of the CT images. At the same time, requirements for CT imaging include reducing the radiation dose exposure to patients and minimizing scanning time. A solution to this is the sparse-view CT technique, based on a reduced number of projection views. This, however, introduces a new problem— the incomplete projection data results in lower quality of the reconstructed images. To tackle this issue, deep learning methods have been applied to enhance the quality of the sparse-view CT images. A first approach involved employing Mir-Net, a dedicated deep neural network designed for image enhancement. This showed promise, utilizing an intricate architecture comprising encoder and decoder networks, along with the incorporation of the Charbonnier Loss. However, this approach was computationally demanding. Subsequently, a specialized Generative Adversarial Network (GAN) architecture, rooted in the Pix2Pix framework, was implemented. This GAN framework involves a U-Net-based Generator and a Discriminator based on Convolutional Neural Networks. To bolster the GAN's performance, both Charbonnier and Wasserstein loss functions were introduced, collectively focusing on capturing minute details while ensuring training stability. The integration of the perceptual loss, calculated based on feature vectors extracted from the VGG16 network pretrained on the ImageNet dataset, further enhanced the network's ability to synthesize relevant images. A series of comprehensive experiments with clinical CT data were conducted, exploring various GAN loss functions, including Wasserstein, Charbonnier, and perceptual loss. The outcomes demonstrated significant image quality improvements, confirmed through pertinent metrics such as Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index (SSIM) between the corrected images and the ground truth. Furthermore, learning curves and qualitative comparisons added evidence of the enhanced image quality and the network's increased stability, while preserving pixel value intensity. The experiments underscored the potential of deep learning frameworks in enhancing the visual interpretation of CT scans, achieving outcomes with SSIM values close to one and PSNR values reaching up to 76.Keywords: generative adversarial networks, sparse view computed tomography, CT image correction, Mir-Net
Procedia PDF Downloads 161