Search results for: computing curricula
821 Exploring Augmented Reality in Graphic Design: A Hybrid Pedagogical Model for Design Education
Authors: Nan Hu, Wujun Wang
Abstract:
In the ever-changing digital arena, augmented reality (AR) applications have transitioned from technological enthusiasm into business endeavors, signaling a near future in which AR applications are integrated into daily life. While practitioners in the design industry continue to explore AR’s potential for innovative communication, educators have taken steps to incorporate AR into the curricula for design, explore its creative potential, and realize early initiatives for teaching AR in design-related disciplines. In alignment with recent advancements, this paper presents a pedagogical model for a hybrid studio course in which students collaborate with AR alongside 3D modeling and graphic design. The course extended students’ digital capacity, fostered their design thinking skills, and immersed them in a multidisciplinary design process. This paper outlines the course and evaluates its effectiveness by discussing challenges encountered and outcomes generated in this particular pedagogical context. By sharing insights from the teaching experience, we aim to empower the community of design educators and offer institutions a valuable reference for advancing their curricular approaches. This paper is a testament to the ever-evolving landscape of design education and its response to the digital age.Keywords: 3D, AR, augmented reality, design thinking, graphic design
Procedia PDF Downloads 71820 Design Thinking and Project-Based Learning: Opportunities, Challenges, and Possibilities
Authors: Shoba Rathilal
Abstract:
High unemployment rates and a shortage of experienced and qualified employees appear to be a paradox that currently plagues most countries worldwide. In a developing country like South Africa, the rate of unemployment is reported to be approximately 35%, the highest recorded globally. At the same time, a countrywide deficit in experienced and qualified potential employees is reported in South Africa, which is causing fierce rivalry among firms. Employers have reported that graduates are very rarely able to meet the demands of the job as there are gaps in their knowledge and conceptual understanding and other 21st-century competencies, attributes, and dispositions required to successfully negotiate the multiple responsibilities of employees in organizations. In addition, the rates of unemployment and suitability of graduates appear to be skewed by race and social class, the continued effects of a legacy of inequitable educational access. Higher Education in the current technologically advanced and dynamic world needs to serve as an agent of transformation, aspiring to develop graduates to be creative, flexible, critical, and with entrepreneurial acumen. This requires that higher education curricula and pedagogy require a re-envisioning of our selection, sequencing, and pacing of the learning, teaching, and assessment. At a particular Higher education Institution in South Africa, Design Thinking and Project Based learning are being adopted as two approaches that aim to enhance the student experience through the provision of a “distinctive education” that brings together disciplinary knowledge, professional engagement, technology, innovation, and entrepreneurship. Using these methodologies forces the students to solve real-time applied problems using various forms of knowledge and finding innovative solutions that can result in new products and services. The intention is to promote the development of skills for self-directed learning, facilitate the development of self-awareness, and contribute to students being active partners in the application and production of knowledge. These approaches emphasize active and collaborative learning, teamwork, conflict resolution, and problem-solving through effective integration of theory and practice. In principle, both these approaches are extremely impactful. However, at the institution in this study, the implementation of the PBL and DT was not as “smooth” as anticipated. This presentation reports on the analysis of the implementation of these two approaches within higher education curricula at a particular university in South Africa. The study adopts a qualitative case study design. Data were generated through the use of surveys, evaluation feedback at workshops, and content analysis of project reports. Data were analyzed using document analysis, content, and thematic analysis. Initial analysis shows that the forces constraining the implementation of PBL and DT range from the capacity to engage with DT and PBL, both from staff and students, educational contextual realities of higher education institutions, administrative processes, and resources. At the same time, the implementation of DT and PBL was enabled through the allocation of strategic funding and capacity development workshops. These factors, however, could not achieve maximum impact. In addition, the presentation will include recommendations on how DT and PBL could be adapted for differing contexts will be explored.Keywords: design thinking, project based learning, innovative higher education pedagogy, student and staff capacity development
Procedia PDF Downloads 77819 An Adiabatic Quantum Optimization Approach for the Mixed Integer Nonlinear Programming Problem
Authors: Maxwell Henderson, Tristan Cook, Justin Chan Jin Le, Mark Hodson, YoungJung Chang, John Novak, Daniel Padilha, Nishan Kulatilaka, Ansu Bagchi, Sanjoy Ray, John Kelly
Abstract:
We present a method of using adiabatic quantum optimization (AQO) to solve a mixed integer nonlinear programming (MINLP) problem instance. The MINLP problem is a general form of a set of NP-hard optimization problems that are critical to many business applications. It requires optimizing a set of discrete and continuous variables with nonlinear and potentially nonconvex constraints. Obtaining an exact, optimal solution for MINLP problem instances of non-trivial size using classical computation methods is currently intractable. Current leading algorithms leverage heuristic and divide-and-conquer methods to determine approximate solutions. Creating more accurate and efficient algorithms is an active area of research. Quantum computing (QC) has several theoretical benefits compared to classical computing, through which QC algorithms could obtain MINLP solutions that are superior to current algorithms. AQO is a particular form of QC that could offer more near-term benefits compared to other forms of QC, as hardware development is in a more mature state and devices are currently commercially available from D-Wave Systems Inc. It is also designed for optimization problems: it uses an effect called quantum tunneling to explore all lowest points of an energy landscape where classical approaches could become stuck in local minima. Our work used a novel algorithm formulated for AQO to solve a special type of MINLP problem. The research focused on determining: 1) if the problem is possible to solve using AQO, 2) if it can be solved by current hardware, 3) what the currently achievable performance is, 4) what the performance will be on projected future hardware, and 5) when AQO is likely to provide a benefit over classical computing methods. Two different methods, integer range and 1-hot encoding, were investigated for transforming the MINLP problem instance constraints into a mathematical structure that can be embedded directly onto the current D-Wave architecture. For testing and validation a D-Wave 2X device was used, as well as QxBranch’s QxLib software library, which includes a QC simulator based on simulated annealing. Our results indicate that it is mathematically possible to formulate the MINLP problem for AQO, but that currently available hardware is unable to solve problems of useful size. Classical general-purpose simulated annealing is currently able to solve larger problem sizes, but does not scale well and such methods would likely be outperformed in the future by improved AQO hardware with higher qubit connectivity and lower temperatures. If larger AQO devices are able to show improvements that trend in this direction, commercially viable solutions to the MINLP for particular applications could be implemented on hardware projected to be available in 5-10 years. Continued investigation into optimal AQO hardware architectures and novel methods for embedding MINLP problem constraints on to those architectures is needed to realize those commercial benefits.Keywords: adiabatic quantum optimization, mixed integer nonlinear programming, quantum computing, NP-hard
Procedia PDF Downloads 526818 Comparative Study and Parallel Implementation of Stochastic Models for Pricing of European Options Portfolios using Monte Carlo Methods
Authors: Vinayak Bassi, Rajpreet Singh
Abstract:
Over the years, with the emergence of sophisticated computers and algorithms, finance has been quantified using computational prowess. Asset valuation has been one of the key components of quantitative finance. In fact, it has become one of the embryonic steps in determining risk related to a portfolio, the main goal of quantitative finance. This study comprises a drawing comparison between valuation output generated by two stochastic dynamic models, namely Black-Scholes and Dupire’s bi-dimensionality model. Both of these models are formulated for computing the valuation function for a portfolio of European options using Monte Carlo simulation methods. Although Monte Carlo algorithms have a slower convergence rate than calculus-based simulation techniques (like FDM), they work quite effectively over high-dimensional dynamic models. A fidelity gap is analyzed between the static (historical) and stochastic inputs for a sample portfolio of underlying assets. In order to enhance the performance efficiency of the model, the study emphasized the use of variable reduction methods and customizing random number generators to implement parallelization. An attempt has been made to further implement the Dupire’s model on a GPU to achieve higher computational performance. Furthermore, ideas have been discussed around the performance enhancement and bottleneck identification related to the implementation of options-pricing models on GPUs.Keywords: monte carlo, stochastic models, computational finance, parallel programming, scientific computing
Procedia PDF Downloads 162817 Enabling Cloud Adoption Based Secured Mobile Banking through Backend as a Service
Authors: P. S. Jagadeesh Kumar, S. Meenakshi Sundaram
Abstract:
With the increase of prevailing non-traditional rivalry, mobile banking experiences an ever changing commercial backdrop. Substantial customer demands have established to be more intricate as customers request more expediency and superintend over their banking services. To enterprise advance and modernization in mobile banking applications, it is gradually obligatory to deeply leapfrog the scuffle using business model transformation. The dramaturgical vicissitudes taking place in mobile banking entail advanced traditions to exploit security. By reforming and transforming older back office into integrated mobile banking applications, banks can engender a supple and nimble banking environment that can rapidly respond to new business requirements over cloud computing. Cloud computing is transfiguring ecosystems in numerous industries, and mobile banking is no exemption providing services innovation, greater flexibility to respond to improved security and enhanced business intelligence with less cost. Cloud technology offer secure deployment possibilities that can provision banks in developing new customer experiences, empower operative relationship and advance speed to efficient banking transaction. Cloud adoption is escalating quickly since it can be made secured for commercial mobile banking transaction through backend as a service in scrutinizing the security strategies of the cloud service provider along with the antiquity of transaction details and their security related practices.Keywords: cloud adoption, backend as a service, business intelligence, secured mobile banking
Procedia PDF Downloads 254816 Introducing Transcending Pedagogies
Authors: Wajeehah Aayeshah, Joy Higgs
Abstract:
The term “transcending pedagogies” has been created to refer to teaching and learning strategies that transcend the mode of student enrolment, the needs of different students, and different learning spaces. The value of such pedagogies in the current arena when learning spaces, technologies and preferences are more volatile than ever before, is a key focus of this paper. The paper will examine current and emerging pedagogies that transcend the learning spaces and enrollment modes of on campus, distance, virtual and workplace learning contexts. A further point of interest is how academics in professional and higher education settings interpret and implement pedagogies in the current global conversation space and re-creation of higher education. This study questioned how the notion and practice of transcending pedagogies enables us to re-imagine and reshape university curricula. It explored the nature of teaching and learning spaces and those professional and higher education (current and emerging) pedagogies that can be implemented across these spaces. We set out to identify how transcending pedagogies can assist students in learning to deal with complexity, uncertainty and change in the practice worlds and better appeal to students who are making decisions on where to enrol. The data for this study was collected through in-depth interviews and focus groups with academics and policy makers within academia.Keywords: Transcending Pedagogies, teaching and learning strategies, learning spaces, pedagogies
Procedia PDF Downloads 538815 Lessons Learned through a Bicultural Approach to Tsunami Education in Aotearoa New Zealand
Authors: Lucy H. Kaiser, Kate Boersen
Abstract:
Kura Kaupapa Māori (kura) and bilingual schools are primary schools in Aotearoa/New Zealand which operate fully or partially under Māori custom and have curricula developed to include Te Reo Māori and Tikanga Māori (Māori language and cultural practices). These schools were established to support Māori children and their families through reinforcing cultural identity by enabling Māori language and culture to flourish in the field of education. Māori kaupapa (values), Mātauranga Māori (Māori knowledge) and Te Reo are crucial considerations for the development of educational resources developed for kura, bilingual and mainstream schools. The inclusion of hazard risk in education has become an important issue in New Zealand due to the vulnerability of communities to a plethora of different hazards. Māori have an extensive knowledge of their local area and the history of hazards which is often not appropriately recognised within mainstream hazard education resources. Researchers from the Joint Centre for Disaster Research, Massey University and East Coast LAB (Life at the Boundary) in Napier were funded to collaboratively develop a toolkit of tsunami risk reduction activities with schools located in Hawke’s Bay’s tsunami evacuation zones. A Māori-led bicultural approach to developing and running the education activities was taken, focusing on creating culturally and locally relevant materials for students and schools as well as giving students a proactive role in making their communities better prepared for a tsunami event. The community-based participatory research is Māori-centred, framed by qualitative and Kaupapa Maori research methodologies and utilizes a range of data collection methods including interviews, focus groups and surveys. Māori participants, stakeholders and the researchers collaborated through the duration of the project to ensure the programme would align with the wider school curricula and kaupapa values. The education programme applied a tuakana/teina, Māori teaching and learning approach in which high school aged students (tuakana) developed tsunami preparedness activities to run with primary school students (teina). At the end of the education programme, high school students were asked to reflect on their participation, what they had learned and what they had enjoyed during the activities. This paper draws on lessons learned throughout this research project. As an exemplar, retaining a bicultural and bilingual perspective resulted in a more inclusive project as there was variability across the students’ levels of confidence using Te Reo and Māori knowledge and cultural frameworks. Providing a range of different learning and experiential activities including waiata (Māori songs), pūrākau (traditional stories) and games was important to ensure students had the opportunity to participate and contribute using a range of different approaches that were appropriate to their individual learning needs. Inclusion of teachers in facilitation also proved beneficial in assisting classroom behavioral management. Lessons were framed by the tikanga and kawa (protocols) of the school to maintain cultural safety for the researchers and the students. Finally, the tuakana/teina component of the education activities became the crux of the programme, demonstrating a path for Rangatahi to support their whānau and communities through facilitating disaster preparedness, risk reduction and resilience.Keywords: school safety, indigenous, disaster preparedness, children, education, tsunami
Procedia PDF Downloads 122814 A Framework of Virtualized Software Controller for Smart Manufacturing
Authors: Pin Xiu Chen, Shang Liang Chen
Abstract:
A virtualized software controller is developed in this research to replace traditional hardware control units. This virtualized software controller transfers motion interpolation calculations from the motion control units of end devices to edge computing platforms, thereby reducing the end devices' computational load and hardware requirements and making maintenance and updates easier. The study also applies the concept of microservices, dividing the control system into several small functional modules and then deploy into a cloud data server. This reduces the interdependency among modules and enhances the overall system's flexibility and scalability. Finally, with containerization technology, the system can be deployed and started in a matter of seconds, which is more efficient than traditional virtual machine deployment methods. Furthermore, this virtualized software controller communicates with end control devices via wireless networks, making the placement of production equipment or the redesign of processes more flexible and no longer limited by physical wiring. To handle the large data flow and maintain low-latency transmission, this study integrates 5G technology, fully utilizing its high speed, wide bandwidth, and low latency features to achieve rapid and stable remote machine control. An experimental setup is designed to verify the feasibility and test the performance of this framework. This study designs a smart manufacturing site with a 5G communication architecture, serving as a field for experimental data collection and performance testing. The smart manufacturing site includes one robotic arm, three Computer Numerical Control machine tools, several Input/Output ports, and an edge computing architecture. All machinery information is uploaded to edge computing servers and cloud servers via 5G communication and the Internet of Things framework. After analysis and computation, this information is converted into motion control commands, which are transmitted back to the relevant machinery for motion control through 5G communication. The communication time intervals at each stage are calculated using the C++ chrono library to measure the time difference for each command transmission. The relevant test results will be organized and displayed in the full-text.Keywords: 5G, MEC, microservices, virtualized software controller, smart manufacturing
Procedia PDF Downloads 82813 Security Design of Root of Trust Based on RISC-V
Authors: Kang Huang, Wanting Zhou, Shiwei Yuan, Lei Li
Abstract:
Since information technology develops rapidly, the security issue has become an increasingly critical for computer system. In particular, as cloud computing and the Internet of Things (IoT) continue to gain widespread adoption, computer systems need to new security threats and attacks. The Root of Trust (RoT) is the foundation for providing basic trusted computing, which is used to verify the security and trustworthiness of other components. Design a reliable Root of Trust and guarantee its own security are essential for improving the overall security and credibility of computer systems. In this paper, we discuss the implementation of self-security technology based on the RISC-V Root of Trust at the hardware level. To effectively safeguard the security of the Root of Trust, researches on security safeguard technology on the Root of Trust have been studied. At first, a lightweight and secure boot framework is proposed as a secure mechanism. Secondly, two kinds of memory protection mechanism are built to against memory attacks. Moreover, hardware implementation of proposed method has been also investigated. A series of experiments and tests have been carried on to verify to effectiveness of the proposed method. The experimental results demonstrated that the proposed approach is effective in verifying the integrity of the Root of Trust’s own boot rom, user instructions, and data, ensuring authenticity and enabling the secure boot of the Root of Trust’s own system. Additionally, our approach provides memory protection against certain types of memory attacks, such as cache leaks and tampering, and ensures the security of root-of-trust sensitive information, including keys.Keywords: root of trust, secure boot, memory protection, hardware security
Procedia PDF Downloads 216812 The Licence, Master, Doctorate in Algeria and Education Quality: Affect and Effect Outcomes
Authors: Farouk A. N. Bouhadiba
Abstract:
This work addresses the issue of the LMD(Licence, Master, Doctorat) in Algeria and the impact it has had on education quality in terms of educational affect and effect. It starts with a brief introduction to the financial means, the educational settings, and the social environment in place when the LMD was institutionalized in Algeria (2003-2004). Some factors for the success or failure of this top-down institutional endeavor are examined and analyzed. These include – among other factors - the teacher/student attitudes, apprehensions, and motivations on the one hand and the institutional euphoria for the LMD in Algeria on the other hand. Some issues at stake are discussed. More specifically, the professional versus the student affect on today’s attitudes, interests, and values is examined as a result of nearly two decades of LMD teaching and learning in Algerian universities. We shall then present some official curricula that, in terms of content, reflect the spirit, principles, and architectures of the LMD but which, in reality, are partially, if not fully, set aside when it comes to teaching practices, learning behaviors, motivation, and evaluation. The discussion on effect highlights attitudinal, developmental, and social markers that are indicative of the extent to which Education Quality in Algeria has been positively or negatively affected by the implementation of the LMD.Keywords: LMD bachelor's masters doctorat, affects and effects, education quality, Algeria
Procedia PDF Downloads 30811 Application of Griddization Management to Construction Hazard Management
Authors: Lingzhi Li, Jiankun Zhang, Tiantian Gu
Abstract:
Hazard management that can prevent fatal accidents and property losses is a fundamental process during the buildings’ construction stage. However, due to lack of safety supervision resources and operational pressures, the conduction of hazard management is poor and ineffective in China. In order to improve the quality of construction safety management, it is critical to explore the use of information technologies to ensure that the process of hazard management is efficient and effective. After exploring the existing problems of construction hazard management in China, this paper develops the griddization management model for construction hazard management. First, following the knowledge grid infrastructure, the griddization computing infrastructure for construction hazards management is designed which includes five layers: resource entity layer, information management layer, task management layer, knowledge transformation layer and application layer. This infrastructure will be as the technical support for realizing grid management. Second, this study divides the construction hazards into grids through city level, district level and construction site level according to grid principles. Last, a griddization management process including hazard identification, assessment and control is developed. Meanwhile, all stakeholders of construction safety management, such as owners, contractors, supervision organizations and government departments, should take the corresponding responsibilities in this process. Finally, a case study based on actual construction hazard identification, assessment and control is used to validate the effectiveness and efficiency of the proposed griddization management model. The advantage of this designed model is to realize information sharing and cooperative management between various safety management departments.Keywords: construction hazard, griddization computing, grid management, process
Procedia PDF Downloads 275810 Optimal Design Solution in "The Small Module" Within the Possibilities of Ecology, Environmental Science/Engineering, and Economics
Authors: Hassan Wajid
Abstract:
We will commend accommodating an environmentally friendly architectural proposal that is extremely common/usual but whose features will make it a sustainable space. In this experiment, the natural and artificial built space is being proposed in such a way that deals with Environmental, Ecological, and Economic Criteria under different climatic conditions. Moreover, the criteria against ecology-environment-economics reflect in the different modules which are being experimented with and analyzed by multiple research groups. The ecological, environmental, and economic services are provided used as units of production side by side, resulting in local job creation and saving resources, for instance, conservation of rainwater, soil formation or protection, less energy consumption to achieve Net Zero, and a stable climate as a whole. The synthesized results from the collected data suggest several aspects to consider when designing buildings for beginning the design process under the supervision of instructors/directors who are responsible for developing curricula and sustainable goals. Hence, the results of the research and the suggestions will benefit the sustainable design through multiple results, heat analysis of different small modules, and comparisons. As a result, it is depleted as the resources are either consumed or the pollution contaminates the resources.Keywords: optimization, ecology, environment, sustainable solution
Procedia PDF Downloads 73809 Cloud Resources Utilization and Science Teacher’s Effectiveness in Secondary Schools in Cross River State, Nigeria
Authors: Michael Udey Udam
Abstract:
Background: This study investigated the impact of cloud resources, a component of cloud computing, on science teachers’ effectiveness in secondary schools in Cross River State. Three (3) research questions and three (3) alternative hypotheses guided the study. Method: The descriptive survey design was adopted for the study. The population of the study comprised 1209 science teachers in public secondary schools of Cross River state. Sample: A sample of 487 teachers was drawn from the population using a stratified random sampling technique. The researcher-made structured questionnaire with 18 was used for data collection for the study. Research question one was answered using the Pearson Product Moment Correlation, while research question two and the hypotheses were answered using the Analysis of Variance (ANOVA) statistics in the Statistical Package for Social Sciences (SPSS) at a 0.05 level of significance. Results: The results of the study revealed that there is a positive correlation between the utilization of cloud resources in teaching and teaching effectiveness among science teachers in secondary schools in Cross River state; there is a negative correlation between gender and utilization of cloud resources among science teachers in secondary schools in Cross River state; and that there is a significant correlation between teaching experience and the utilization of cloud resources among science teachers in secondary schools in Cross River state. Conclusion: The study justifies the effectiveness of the Cross River state government policy of introducing cloud computing into the education sector. The study recommends that the policy should be sustained.Keywords: cloud resources, science teachers, effectiveness, secondary school
Procedia PDF Downloads 75808 The Case of ESPRIT (HigherSchool of Engineering)
Authors: Amira Potter
Abstract:
Since three years, ESPRIT has adopted project-based learning across its curricula. The philosophy behind this reform is to prepare its future engineers to become more operational once they integrate the workplace. It allows them to learn all the required skills expected from them by their future employers. This learner-centered method helps the students take responsibility for their own learning, solve real-world problems and carry out muli-faceted projects. Therefore, the teacher who used to be considered as the detainer of the knowledge has become more of a facilitator and a coach, encouraging their students’ learning process. This innovative way to English teaching has enabled the students to learn the English language differently. The target language is learnt cooperatively through group work, presentations, debating and many other communicative activities. The speaking skill in English language remains by far the most challenging skill for Tunisian students with an educational background based on Arabic as a first language and French as a second language. The student’s initial resistance to speak English in front of their classmates and the way they end up performing their work, shows the real progress they managed to achieve through PBL approach. The article will focus on the positive impact PBL has had on oral fluency among Esprit engineering students and how it has been achieved. It will also describe how speaking skill is taught and assessed at ESPRIT.Keywords: cooperative, engineer, innovative, project-based learning
Procedia PDF Downloads 318807 Head of the Class: A Study of What United States Journalism School Administrators Consider the Most Valuable Educational Tenets for Their Graduates Seeking Careers at U.S. Legacy Newspapers
Authors: Adam Pitluk
Abstract:
In a time period populated by legacy newspaper readers who throw around the term “fake news” as though it has long been a part of the lexicon, journalism schools must convince would-be students that their degree is still viable and that they are not teaching a curriculum of deception. As such, journalism schools’ academic administrators tasked with creating and maintaining conversant curricula must stay ahead of legacy newspaper industry trends – both in the print and online products – and ensure that what is being taught in the classroom is both fresh and appropriate to the demands of the evolving legacy newspaper industry. This study examines the information obtained from the result of interviews of journalism academic administrators in order to identify institutional pedagogy for recent journalism school graduates interested in pursuing careers at legacy newspapers. This research also explores the existing relationship between journalism school academic administrators and legacy newspaper editors. The results indicate the value administrators put on various academy teachings, and they also highlight a perceived disconnect between journalism academic administrators and legacy newspaper hiring editors.Keywords: academic administration, education, journalism, journalism school graduates, media management, newspapers, grounded theory
Procedia PDF Downloads 125806 An Adjoint-Based Method to Compute Derivatives with Respect to Bed Boundary Positions in Resistivity Measurements
Authors: Mostafa Shahriari, Theophile Chaumont-Frelet, David Pardo
Abstract:
Resistivity measurements are used to characterize the Earth’s subsurface. They are categorized into two different groups: (a) those acquired on the Earth’s surface, for instance, controlled source electromagnetic (CSEM) and Magnetotellurics (MT), and (b) those recorded with borehole logging instruments such as Logging-While-Drilling (LWD) devices. LWD instruments are mostly used for geo-steering purposes, i.e., to adjust dip and azimuthal angles of a well trajectory to drill along a particular geological target. Modern LWD tools measure all nine components of the magnetic field corresponding to three orthogonal transmitter and receiver orientations. In order to map the Earth’s subsurface and perform geo-steering, we invert measurements using a gradient-based method that utilizes the derivatives of the recorded measurements with respect to the inversion variables. For resistivity measurements, these inversion variables are usually the constant resistivity value of each layer and the bed boundary positions. It is well-known how to compute derivatives with respect to the constant resistivity value of each layer using semi-analytic or numerical methods. However, similar formulas for computing the derivatives with respect to bed boundary positions are unavailable. The main contribution of this work is to provide an adjoint-based formulation for computing derivatives with respect to the bed boundary positions. The key idea to obtain the aforementioned adjoint state formulations for the derivatives is to separate the tangential and normal components of the field and treat them differently. This formulation allows us to compute the derivatives faster and more accurately than with traditional finite differences approximations. In the presentation, we shall first derive a formula for computing the derivatives with respect to the bed boundary positions for the potential equation. Then, we shall extend our formulation to 3D Maxwell’s equations. Finally, by considering a 1D domain and reducing the dimensionality of the problem, which is a common practice in the inversion of resistivity measurements, we shall derive a formulation to compute the derivatives of the measurements with respect to the bed boundary positions using a 1.5D variational formulation. Then, we shall illustrate the accuracy and convergence properties of our formulations by comparing numerical results with the analytical derivatives for the potential equation. For the 1.5D Maxwell’s system, we shall compare our numerical results based on the proposed adjoint-based formulation vs those obtained with a traditional finite difference approach. Numerical results shall show that our proposed adjoint-based technique produces enhanced accuracy solutions while its cost is negligible, as opposed to the finite difference approach that requires the solution of one additional problem per derivative.Keywords: inverse problem, bed boundary positions, electromagnetism, potential equation
Procedia PDF Downloads 178805 Optimizing Data Transfer and Processing in Multi-Cloud Environments for Big Data Workloads
Authors: Gaurav Kumar Sinha
Abstract:
In an era defined by the proliferation of data and the utilization of cloud computing environments, the efficient transfer and processing of big data workloads across multi-cloud platforms have emerged as critical challenges. This research paper embarks on a comprehensive exploration of the complexities associated with managing and optimizing big data in a multi-cloud ecosystem.The foundation of this study is rooted in the recognition that modern enterprises increasingly rely on multiple cloud providers to meet diverse business needs, enhance redundancy, and reduce vendor lock-in. As a consequence, managing data across these heterogeneous cloud environments has become intricate, necessitating innovative approaches to ensure data integrity, security, and performance.The primary objective of this research is to investigate strategies and techniques for enhancing the efficiency of data transfer and processing in multi-cloud scenarios. It recognizes that big data workloads are characterized by their sheer volume, variety, velocity, and complexity, making traditional data management solutions insufficient for harnessing the full potential of multi-cloud architectures.The study commences by elucidating the challenges posed by multi-cloud environments in the context of big data. These challenges encompass data fragmentation, latency, security concerns, and cost optimization. To address these challenges, the research explores a range of methodologies and solutions. One of the key areas of focus is data transfer optimization. The paper delves into techniques for minimizing data movement latency, optimizing bandwidth utilization, and ensuring secure data transmission between different cloud providers. It evaluates the applicability of dedicated data transfer protocols, intelligent data routing algorithms, and edge computing approaches in reducing transfer times.Furthermore, the study examines strategies for efficient data processing across multi-cloud environments. It acknowledges that big data processing requires distributed and parallel computing capabilities that span across cloud boundaries. The research investigates containerization and orchestration technologies, serverless computing models, and interoperability standards that facilitate seamless data processing workflows.Security and data governance are paramount concerns in multi-cloud environments. The paper explores methods for ensuring data security, access control, and compliance with regulatory frameworks. It considers encryption techniques, identity and access management, and auditing mechanisms as essential components of a robust multi-cloud data security strategy.The research also evaluates cost optimization strategies, recognizing that the dynamic nature of multi-cloud pricing models can impact the overall cost of data transfer and processing. It examines approaches for workload placement, resource allocation, and predictive cost modeling to minimize operational expenses while maximizing performance.Moreover, this study provides insights into real-world case studies and best practices adopted by organizations that have successfully navigated the challenges of multi-cloud big data management. It presents a comparative analysis of various multi-cloud management platforms and tools available in the market.Keywords: multi-cloud environments, big data workloads, data transfer optimization, data processing strategies
Procedia PDF Downloads 68804 Yawning Computing Using Bayesian Networks
Authors: Serge Tshibangu, Turgay Celik, Zenzo Ncube
Abstract:
Road crashes kill nearly over a million people every year, and leave millions more injured or permanently disabled. Various annual reports reveal that the percentage of fatal crashes due to fatigue/driver falling asleep comes directly after the percentage of fatal crashes due to intoxicated drivers. This percentage is higher than the combined percentage of fatal crashes due to illegal/Un-Safe U-turn and illegal/Un-Safe reversing. Although a relatively small percentage of police reports on road accidents highlights drowsiness and fatigue, the importance of these factors is greater than we might think, hidden by the undercounting of their events. Some scenarios show that these factors are significant in accidents with killed and injured people. Thus the need for an automatic drivers fatigue detection system in order to considerably reduce the number of accidents owing to fatigue.This research approaches the drivers fatigue detection problem in an innovative way by combining cues collected from both temporal analysis of drivers’ faces and environment. Monotony in driving environment is inter-related with visual symptoms of fatigue on drivers’ faces to achieve fatigue detection. Optical and infrared (IR) sensors are used to analyse the monotony in driving environment and to detect the visual symptoms of fatigue on human face. Internal cues from drivers faces and external cues from environment are combined together using machine learning algorithms to automatically detect fatigue.Keywords: intelligent transportation systems, bayesian networks, yawning computing, machine learning algorithms
Procedia PDF Downloads 455803 FRATSAN: A New Software for Fractal Analysis of Signals
Authors: Hamidreza Namazi
Abstract:
Fractal analysis is assessing fractal characteristics of data. It consists of several methods to assign fractal characteristics to a dataset which may be a theoretical dataset or a pattern or signal extracted from phenomena including natural geometric objects, sound, market fluctuations, heart rates, digital images, molecular motion, networks, etc. Fractal analysis is now widely used in all areas of science. An important limitation of fractal analysis is that arriving at an empirically determined fractal dimension does not necessarily prove that a pattern is fractal; rather, other essential characteristics have to be considered. For this purpose a Visual C++ based software called FRATSAN (FRActal Time Series ANalyser) was developed which extract information from signals through three measures. These measures are Fractal Dimensions, Jeffrey’s Measure and Hurst Exponent. After computing these measures, the software plots the graphs for each measure. Besides computing three measures the software can classify whether the signal is fractal or no. In fact, the software uses a dynamic method of analysis for all the measures. A sliding window is selected with a value equal to 10% of the total number of data entries. This sliding window is moved one data entry at a time to obtain all the measures. This makes the computation very sensitive to slight changes in data, thereby giving the user an acute analysis of the data. In order to test the performance of this software a set of EEG signals was given as input and the results were computed and plotted. This software is useful not only for fundamental fractal analysis of signals but can be used for other purposes. For instance by analyzing the Hurst exponent plot of a given EEG signal in patients with epilepsy the onset of seizure can be predicted by noticing the sudden changes in the plot.Keywords: EEG signals, fractal analysis, fractal dimension, hurst exponent, Jeffrey’s measure
Procedia PDF Downloads 467802 Discursively Examination of 8th Grade Students’ Geometric Thinking Levels
Authors: Ferdağ Çulhan, Emine Gaye Çontay
Abstract:
Geometric thinking levels created by Van Hiele are used to determine students' progress in geometric thinking. Many studies have been conducted on geometric thinking levels and they have taken their place in teaching curricula over time. It is thought that geometric thinking levels, which have become so important in teaching, can be examined in depth. In order to make an in-depth analysis, it was decided that the most appropriate management was discourse analysis. In this study, the focus is on examining the geometric thinking levels of 8th grade students from a discursive point of view. Sfard (2008)'s "Commognitive" theory will be used to conduct discursive analysis. The "Global Van Hiele Questionnaire" created by Patkin (2014) and translated into Turkish for this research will be used in the research. The "Global Van Hiele Questionnaire" contains questions from the sub-learning domain of triangles and quadrilaterals, circles and geometric objects. It has a wider scope than many "Van Hiele Questionnaires". “Global Van Hiele Questionnaire” will be applied to 8th grade students. Then, the geometric thinking levels of the students will be determined and interviews will be held with two students from each of the 1st, 2nd and 3rd levels. The interviews will be recorded and the students' discourses will be examined. By evaluating the relations between the students' geometric thinking levels and their discourses, it will be examined how much their discourse reflects their level of thinking. In this way, it is thought that students' geometric thinking processes can be better understood.Keywords: mathematical discourses, commognitive framework, geometric thinking levels, van hiele
Procedia PDF Downloads 129801 Pharmaceutical Innovation in Jordan: KAP Analysis
Authors: Abdel Qader Al Bawab, Mohannad Odeh, Rami Amer
Abstract:
Recently, there has been an increasing interest in innovative business development. Nevertheless, in the pharmacy practice field, there seems to be a gap in perceptions, attitudes, and knowledge about innovation between practicing pharmacists and academia. This study explores this gap and aspects of pharmaceutical innovation in Jordan, comparing pharmacists and last-year pharmacy students. A validated (r2 = 0.74) and reliable (Pearson’s r = 0.88) online questionnaire was designed to assess and compare knowledge, attitude, and perceptions about pharmaceutical innovation. A total of 397 participants (215 pharmacy students and 182 pharmaceutical professionals) responded. Compared with 50% of the pharmacists, only 32.1% of the students claimed that they knew the differences between pharmaceutical innovation, discovery, invention, and entrepreneurship [x2 (2) = 14.238, p = 0.001; Cramer’s V = 0.189]. Pharmacists demonstrated a higher level of trust in the innovative website design for their institution compared with students (25.3% vs. 16.3%, p < 0.001, Cramer’s V = 0.327). However, 60% of the students did not know the innovative design standards for websites, while the corresponding percentage was 37% for the pharmacists (p < 0.001; Cramer’s V = 0.327). The majority of the students were interested in pharmaceutical innovation (81.9%). Unfortunately, 76.3% never studied innovation in their pharmacy curricula. Similarly, most pharmacists (76.4%) considered adopting innovation, but only 30% had a concrete plan. For the field where pharmacists aim to innovate in the next 5 years, new pharmaceutical services were the dominant field (34.6%). Despite a positive attitude and perception, pharmacists and pharmacy students expressed poor knowledge about innovation. Policies to enhance awareness about innovation and professional educational tools should be implemented.Keywords: pharmacy, innovation, knowledge, attitude, practice
Procedia PDF Downloads 87800 Gaia (Earth) Education Philosophy – A Journey Back to the Future
Authors: Darius Singh
Abstract:
This study adopts a research, develop, and deploy methodology to create a state-of-the-art forest preschool environment using technology and the Gaia (Earth) Education Philosophy as design support. The new philosophy adopts an ancient Greek terminology, “Gaia,” meaning “Mother Earth”, and it take its principle to model everything with the oldest living and breathing entity that it know – Earth. This includes using nature and biomimicry-based principles in building design, environments, curricula, teaching, learning, values and outcomes for children. The study highlights the potential effectiveness of the Gaia (Earth) Education Philosophy as a means of designing Earth-inspired environments for children’s learning. The discuss the strengths of biomimicry-based design principles and propose a curriculum that emphasizes natural outcomes for early childhood learning. Theoretical implications of the study are that the Gaia (Earth) Education Philosophy could serve as a strong foundation for educating young learners.it present a unique approach that promotes connections with Earth-principles and lessons that can contribute to the development of social and environmental consciousness among children and help educate generations to come into a stable and balanced future.Keywords: earth science, nature education, sustainability, gaia, forest school, nature, inspirational teaching and learning
Procedia PDF Downloads 65799 Managing Data from One Hundred Thousand Internet of Things Devices Globally for Mining Insights
Authors: Julian Wise
Abstract:
Newcrest Mining is one of the world’s top five gold and rare earth mining organizations by production, reserves and market capitalization in the world. This paper elaborates on the data acquisition processes employed by Newcrest in collaboration with Fortune 500 listed organization, Insight Enterprises, to standardize machine learning solutions which process data from over a hundred thousand distributed Internet of Things (IoT) devices located at mine sites globally. Through the utilization of software architecture cloud technologies and edge computing, the technological developments enable for standardized processes of machine learning applications to influence the strategic optimization of mineral processing. Target objectives of the machine learning optimizations include time savings on mineral processing, production efficiencies, risk identification, and increased production throughput. The data acquired and utilized for predictive modelling is processed through edge computing by resources collectively stored within a data lake. Being involved in the digital transformation has necessitated the standardization software architecture to manage the machine learning models submitted by vendors, to ensure effective automation and continuous improvements to the mineral process models. Operating at scale, the system processes hundreds of gigabytes of data per day from distributed mine sites across the globe, for the purposes of increased improved worker safety, and production efficiency through big data applications.Keywords: mineral technology, big data, machine learning operations, data lake
Procedia PDF Downloads 112798 Unveiling the Mystery of Innovation in Higher Education Institutions
Authors: Ana Martins, Isabel Martins
Abstract:
The purpose of this research is to ascertain whether students at HEIs cultivate distributed leadership and higher-level skills to inspire knowledge creation. Critical reflection of extant literature illustrates the need for a culture of innovation in organizational sustainability. New age leadership behaviors harmonize innovation. The leadership self-efficacy construct supports organizational learning. This exploratory study applies the pragmatic paradigm methodology using the survey research method for primary data collection. A questionnaire was distributed to a sample of university students based in the Southern Anatolian region of Turkey, from both under and postgraduate Business degree programs. An analysis of the findings reveals a greater connection in influencing behavior relying more on the task-centered perspective rather than with the people perspective. These results reveal the need for HEIs to instill a humanistic perspective in curricula enabling graduates to be capable leaders with the awareness soft skills to energize creativity and innovation. A limitation of this research is that one university makes it difficult to generalize to a broader population. This study is of added value for scholars and organizations in the current knowledge and innovation economy.Keywords: distributed leadership, exploration, higher education institutions, innovation, knowledge creation, learning, self-efficacy
Procedia PDF Downloads 195797 A Comparative Case Study on Teaching Romanian Language to Foreign Students: Swedes in Lund versus Arabs in Alba Iulia
Authors: Lucian Vasile Bagiu, Paraschiva Bagiu
Abstract:
The study is a contrastive essay on language acquisition and learning and follows the outcomes of teaching Romanian language to foreign students both at Lund University, Sweden (from 2014 to 2017) and at '1 Decembrie 1918' University in Alba Iulia, Romania (2017-2018). Having employed the same teaching methodology (on campus, same curricula) for the same level of study (beginners’ level: A1-A2), the essay focuses on the written exam at the end of the semester. The study argues on grammar exercises concerned with: the indefinite and the definite article; the conjugation of verbs in the present indicative; the possessive; verbs in the past tense; the subjunctive; the degrees of comparison for adjectives. Identifying similar errors when solving identical grammar exercises by different groups of foreign students is an opportunity to emphasize the major challenges any foreigner has to face and overcome when trying to acquire Romanian language. The conclusion draws attention to the complexity of the morphology of Romanian language in several key elements which may be insurmountable for a foreign speaker no matter if the language acquisition takes place in a foreign country or a Romanian university.Keywords: Arab students, morphological errors, Romanian language, Swedish students, written exam
Procedia PDF Downloads 258796 An Evolutionary Approach for QAOA for Max-Cut
Authors: Francesca Schiavello
Abstract:
This work aims to create a hybrid algorithm, combining Quantum Approximate Optimization Algorithm (QAOA) with an Evolutionary Algorithm (EA) in the place of traditional gradient based optimization processes. QAOA’s were first introduced in 2014, where, at the time, their algorithm performed better than the traditional best known classical algorithm for Max-cut graphs. Whilst classical algorithms have improved since then and have returned to being faster and more efficient, this was a huge milestone for quantum computing, and their work is often used as a benchmarking tool and a foundational tool to explore variants of QAOA’s. This, alongside with other famous algorithms like Grover’s or Shor’s, highlights to the world the potential that quantum computing holds. It also presents the reality of a real quantum advantage where, if the hardware continues to improve, this could constitute a revolutionary era. Given that the hardware is not there yet, many scientists are working on the software side of things in the hopes of future progress. Some of the major limitations holding back quantum computing are the quality of qubits and the noisy interference they generate in creating solutions, the barren plateaus that effectively hinder the optimization search in the latent space, and the availability of number of qubits limiting the scale of the problem that can be solved. These three issues are intertwined and are part of the motivation for using EAs in this work. Firstly, EAs are not based on gradient or linear optimization methods for the search in the latent space, and because of their freedom from gradients, they should suffer less from barren plateaus. Secondly, given that this algorithm performs a search in the solution space through a population of solutions, it can also be parallelized to speed up the search and optimization problem. The evaluation of the cost function, like in many other algorithms, is notoriously slow, and the ability to parallelize it can drastically improve the competitiveness of QAOA’s with respect to purely classical algorithms. Thirdly, because of the nature and structure of EA’s, solutions can be carried forward in time, making them more robust to noise and uncertainty. Preliminary results show that the EA algorithm attached to QAOA can perform on par with the traditional QAOA with a Cobyla optimizer, which is a linear based method, and in some instances, it can even create a better Max-Cut. Whilst the final objective of the work is to create an algorithm that can consistently beat the original QAOA, or its variants, due to either speedups or quality of the solution, this initial result is promising and show the potential of EAs in this field. Further tests need to be performed on an array of different graphs with the parallelization aspect of the work commencing in October 2023 and tests on real hardware scheduled for early 2024.Keywords: evolutionary algorithm, max cut, parallel simulation, quantum optimization
Procedia PDF Downloads 60795 An Analytical Metric and Process for Critical Infrastructure Architecture System Availability Determination in Distributed Computing Environments under Infrastructure Attack
Authors: Vincent Andrew Cappellano
Abstract:
In the early phases of critical infrastructure system design, translating distributed computing requirements to an architecture has risk given the multitude of approaches (e.g., cloud, edge, fog). In many systems, a single requirement for system uptime / availability is used to encompass the system’s intended operations. However, when architected systems may perform to those availability requirements only during normal operations and not during component failure, or during outages caused by adversary attacks on critical infrastructure (e.g., physical, cyber). System designers lack a structured method to evaluate availability requirements against candidate system architectures through deep degradation scenarios (i.e., normal ops all the way down to significant damage of communications or physical nodes). This increases risk of poor selection of a candidate architecture due to the absence of insight into true performance for systems that must operate as a piece of critical infrastructure. This research effort proposes a process to analyze critical infrastructure system availability requirements and a candidate set of systems architectures, producing a metric assessing these architectures over a spectrum of degradations to aid in selecting appropriate resilient architectures. To accomplish this effort, a set of simulation and evaluation efforts are undertaken that will process, in an automated way, a set of sample requirements into a set of potential architectures where system functions and capabilities are distributed across nodes. Nodes and links will have specific characteristics and based on sampled requirements, contribute to the overall system functionality, such that as they are impacted/degraded, the impacted functional availability of a system can be determined. A machine learning reinforcement-based agent will structurally impact the nodes, links, and characteristics (e.g., bandwidth, latency) of a given architecture to provide an assessment of system functional uptime/availability under these scenarios. By varying the intensity of the attack and related aspects, we can create a structured method of evaluating the performance of candidate architectures against each other to create a metric rating its resilience to these attack types/strategies. Through multiple simulation iterations, sufficient data will exist to compare this availability metric, and an architectural recommendation against the baseline requirements, in comparison to existing multi-factor computing architectural selection processes. It is intended that this additional data will create an improvement in the matching of resilient critical infrastructure system requirements to the correct architectures and implementations that will support improved operation during times of system degradation due to failures and infrastructure attacks.Keywords: architecture, resiliency, availability, cyber-attack
Procedia PDF Downloads 109794 Exploring Data Stewardship in Fog Networking Using Blockchain Algorithm
Authors: Ruvaitha Banu, Amaladhithyan Krishnamoorthy
Abstract:
IoT networks today solve various consumer problems, from home automation systems to aiding in driving autonomous vehicles with the exploration of multiple devices. For example, in an autonomous vehicle environment, multiple sensors are available on roads to monitor weather and road conditions and interact with each other to aid the vehicle in reaching its destination safely and timely. IoT systems are predominantly dependent on the cloud environment for data storage, and computing needs that result in latency problems. With the advent of Fog networks, some of this storage and computing is pushed to the edge/fog nodes, saving the network bandwidth and reducing the latency proportionally. Managing the data stored in these fog nodes becomes crucial as it might also store sensitive information required for a certain application. Data management in fog nodes is strenuous because Fog networks are dynamic in terms of their availability and hardware capability. It becomes more challenging when the nodes in the network also live a short span, detaching and joining frequently. When an end-user or Fog Node wants to access, read, or write data stored in another Fog Node, then a new protocol becomes necessary to access/manage the data stored in the fog devices as a conventional static way of managing the data doesn’t work in Fog Networks. The proposed solution discusses a protocol that acts by defining sensitivity levels for the data being written and read. Additionally, a distinct data distribution and replication model among the Fog nodes is established to decentralize the access mechanism. In this paper, the proposed model implements stewardship towards the data stored in the Fog node using the application of Reinforcement Learning so that access to the data is determined dynamically based on the requests.Keywords: IoT, fog networks, data stewardship, dynamic access policy
Procedia PDF Downloads 59793 High Performance Computing Enhancement of Agent-Based Economic Models
Authors: Amit Gill, Lalith Wijerathne, Sebastian Poledna
Abstract:
This research presents the details of the implementation of high performance computing (HPC) extension of agent-based economic models (ABEMs) to simulate hundreds of millions of heterogeneous agents. ABEMs offer an alternative approach to study the economy as a dynamic system of interacting heterogeneous agents, and are gaining popularity as an alternative to standard economic models. Over the last decade, ABEMs have been increasingly applied to study various problems related to monetary policy, bank regulations, etc. When it comes to predicting the effects of local economic disruptions, like major disasters, changes in policies, exogenous shocks, etc., on the economy of the country or the region, it is pertinent to study how the disruptions cascade through every single economic entity affecting its decisions and interactions, and eventually affect the economic macro parameters. However, such simulations with hundreds of millions of agents are hindered by the lack of HPC enhanced ABEMs. In order to address this, a scalable Distributed Memory Parallel (DMP) implementation of ABEMs has been developed using message passing interface (MPI). A balanced distribution of computational load among MPI-processes (i.e. CPU cores) of computer clusters while taking all the interactions among agents into account is a major challenge for scalable DMP implementations. Economic agents interact on several random graphs, some of which are centralized (e.g. credit networks, etc.) whereas others are dense with random links (e.g. consumption markets, etc.). The agents are partitioned into mutually-exclusive subsets based on a representative employer-employee interaction graph, while the remaining graphs are made available at a minimum communication cost. To minimize the number of communications among MPI processes, real-life solutions like the introduction of recruitment agencies, sales outlets, local banks, and local branches of government in each MPI-process, are adopted. Efficient communication among MPI-processes is achieved by combining MPI derived data types with the new features of the latest MPI functions. Most of the communications are overlapped with computations, thereby significantly reducing the communication overhead. The current implementation is capable of simulating a small open economy. As an example, a single time step of a 1:1 scale model of Austria (i.e. about 9 million inhabitants and 600,000 businesses) can be simulated in 15 seconds. The implementation is further being enhanced to simulate 1:1 model of Euro-zone (i.e. 322 million agents).Keywords: agent-based economic model, high performance computing, MPI-communication, MPI-process
Procedia PDF Downloads 128792 Integrating Sustainable Construction Principles into Curriculum Design for Built Environment Professional Programs in Nigeria
Authors: M. Yakubu, M. B. Isah, S. Bako
Abstract:
This paper presents the findings of a research which sought to investigate the readiness to integrate sustainable construction principles into curriculum design for built environment professional programs in the Nigerian Universities. Developing the knowledge and understanding that construction professionals acquire of sustainable construction practice leads to considerable improvement in the environmental performance of the construction sector. Integrating sustainable environmental issues within the built environment education curricula provide the basis of this research. An integration of sustainable development principles into the universities built environment professional programmes are carried out with a view of finding solutions to the key issues identified. The perspectives of academia have been assessed and findings tested for validity through the analysis of primary quantitative data that has been collected. The secondary data generated has shown that there are significant differences in the approach to curriculum design within the built environment professional programmes, and this reveals that there is no ‘best practice’ that is clearly identifiable. Sequel to the above, this research reveals that engaging all stakeholders would be a useful component of built environment curriculum development, and that the curriculum be negotiated with interested parties. These parties have been identified as academia, government, construction industry and built environment professionals.Keywords: built environment, curriculum development, sustainable construction, sustainable development
Procedia PDF Downloads 420