Search results for: corpus based approach
32766 Integrating Cooperative Education Experience into Engineering Curriculum: An Approach
Authors: Robin Lok-Wang Ma
Abstract:
The Center/Unit for Industry Engagement and Collaboration, as well as Internship, play a significant role at university. In general, the Center serves as the official interface between the industry and the School or Department to cultivate students’ early exposure to professional experience. The missions of the Center are not limited to provide a communication channel and collaborative platform for the industries and the university but also to assist students to build up their career paths early while still in the university. In recent years, a cooperative education experience (commonly known as a co-op) has been strongly advocated for students to make the school-to-work transition. The nature of the co-op program is not only consistent with the internships/final year design projects, but it is also more industrial-oriented with academic support from faculty at the university. The purpose of this paper is to describe an approach to how cooperative education experience can be integrated into Engineering Curriculum. It provides a mutual understanding and exchange of ideas for the approach between the university and the industry. A suggested format in terms of timeline, duration, selection of candidates, students, and companies’ expectations for the co-op program is described. Also, feedbacks from employers/industries show that a longer-term co-op program is well suited for students compared with a short-term internship. To this end, it provides a new insight into collaboration and/or partnership between the university and the industries to prepare professional work-ready graduates.Keywords: cooperative education, industry, engagement, collaboration
Procedia PDF Downloads 9632765 An Observation of the Information Technology Research and Development Based on Article Data Mining: A Survey Study on Science Direct
Authors: Muhammet Dursun Kaya, Hasan Asil
Abstract:
One of the most important factors of research and development is the deep insight into the evolutions of scientific development. The state-of-the-art tools and instruments can considerably assist the researchers, and many of the world organizations have become aware of the advantages of data mining for the acquisition of the knowledge required for the unstructured data. This paper was an attempt to review the articles on the information technology published in the past five years with the aid of data mining. A clustering approach was used to study these articles, and the research results revealed that three topics, namely health, innovation, and information systems, have captured the special attention of the researchers.Keywords: information technology, data mining, scientific development, clustering
Procedia PDF Downloads 27832764 Selection of Rayleigh Damping Coefficients for Seismic Response Analysis of Soil Layers
Authors: Huai-Feng Wang, Meng-Lin Lou, Ru-Lin Zhang
Abstract:
One good analysis method in seismic response analysis is direct time integration, which widely adopts Rayleigh damping. An approach is presented for selection of Rayleigh damping coefficients to be used in seismic analyses to produce a response that is consistent with Modal damping response. In the presented approach, the expression of the error of peak response, acquired through complete quadratic combination method, and Rayleigh damping coefficients was set up and then the coefficients were produced by minimizing the error. Two finite element modes of soil layers, excited by 28 seismic waves, were used to demonstrate the feasibility and validity.Keywords: Rayleigh damping, modal damping, damping coefficients, seismic response analysis
Procedia PDF Downloads 43832763 Fused Structure and Texture (FST) Features for Improved Pedestrian Detection
Authors: Hussin K. Ragb, Vijayan K. Asari
Abstract:
In this paper, we present a pedestrian detection descriptor called Fused Structure and Texture (FST) features based on the combination of the local phase information with the texture features. Since the phase of the signal conveys more structural information than the magnitude, the phase congruency concept is used to capture the structural features. On the other hand, the Center-Symmetric Local Binary Pattern (CSLBP) approach is used to capture the texture information of the image. The dimension less quantity of the phase congruency and the robustness of the CSLBP operator on the flat images, as well as the blur and illumination changes, lead the proposed descriptor to be more robust and less sensitive to the light variations. The proposed descriptor can be formed by extracting the phase congruency and the CSLBP values of each pixel of the image with respect to its neighborhood. The histogram of the oriented phase and the histogram of the CSLBP values for the local regions in the image are computed and concatenated to construct the FST descriptor. Several experiments were conducted on INRIA and the low resolution DaimlerChrysler datasets to evaluate the detection performance of the pedestrian detection system that is based on the FST descriptor. A linear Support Vector Machine (SVM) is used to train the pedestrian classifier. These experiments showed that the proposed FST descriptor has better detection performance over a set of state of the art feature extraction methodologies.Keywords: pedestrian detection, phase congruency, local phase, LBP features, CSLBP features, FST descriptor
Procedia PDF Downloads 48832762 Data Hiding in Gray Image Using ASCII Value and Scanning Technique
Authors: R. K. Pateriya, Jyoti Bharti
Abstract:
This paper presents an approach for data hiding methods which provides a secret communication between sender and receiver. The data is hidden in gray-scale images and the boundary of gray-scale image is used to store the mapping information. In this an approach data is in ASCII format and the mapping is in between ASCII value of hidden message and pixel value of cover image, since pixel value of an image as well as ASCII value is in range of 0 to 255 and this mapping information is occupying only 1 bit per character of hidden message as compared to 8 bit per character thus maintaining good quality of stego image.Keywords: ASCII value, cover image, PSNR, pixel value, stego image, secret message
Procedia PDF Downloads 41632761 Sustainable Intensification of Agriculture in Victoria’s Food Bowl: Optimizing Productivity with the use of Decision-Support Tools
Authors: M. Johnson, R. Faggian, V. Sposito
Abstract:
A participatory and engaged approach is key in connecting agricultural managers to sustainable agricultural systems to support and optimize production in Victoria’s food bowl. A sustainable intensification (SI) approach is well documented globally, but participation rates amongst Victorian farmers is fragmentary, and key outcomes and implementation strategies are poorly understood. Improvement in decision-support management tools and a greater understanding of the productivity gains available upon implementation of SI is necessary. This paper reviews the current understanding and uptake of SI practices amongst farmers in one of Victoria’s premier food producing regions, the Goulburn Broken; and it spatially analyses the potential for this region to adapt to climate change and optimize food production. A Geographical Information Systems (GIS) approach is taken to develop an interactive decision-support tool that can be accessible to on-ground agricultural managers. The tool encompasses multiple criteria analysis (MCA) that identifies factors during the construction phase of the tool, using expert witnesses and regional knowledge, framed within an Analytical Hierarchy Process. Given the complexities of the interrelations between each of the key outcomes, this participatory approach, in which local realities and factors inform the key outcomes and help to strategies for a particular region, results in a robust strategy for sustainably intensifying production in key food producing regions. The creation of an interactive, locally embedded, decision-support management and education tool can help to close the gap between farmer knowledge and production, increase on-farm adoption of sustainable farming strategies and techniques, and optimize farm productivity.Keywords: agriculture, decision-support management tool, Geographic Information System, GIS, sustainable intensification
Procedia PDF Downloads 16632760 Intellectual Capital as Resource Based Business Strategy
Authors: Vidya Nimkar Tayade
Abstract:
Introduction: Intellectual capital of an organization is a key factor to success. Many companies invest a huge amount in their Research and development activities. Any innovation is helpful not only to that particular company but also to many other companies, industry and mankind as a whole. Companies undertake innovative changes for increasing their capital profitability and indirectly increase in pay packages of their employees. The quality of human capital can also improve due to such positive changes. Employees become more skilled and experienced due to such innovations and inventions. For increasing intangible capital, the author has referred to a couple of books and referred case studies to come to a conclusion. Different charts and tables are also referred to by the author. Case studies are more important because they are proven and established techniques. They enable students to apply theoretical concepts in real-world situations. It gives solutions to an open-ended problem with multiple potential solutions. There are three different strategies for undertaking intellectual capital increase. They are: Research push strategy/ Technology pushed approach, Market pull strategy/ approach and Open innovation strategy/approach. Research push strategy, In this strategy, research is undertaken and innovation is achieved on its own. After invention inventor company protects such invention and finds buyers for such invention. In this way, the invention is pushed into the market. In this method, research and development are undertaken first and the outcome of this research is commercialized. Market pull strategy, In this strategy, commercial opportunities are identified first and our research is concentrated in that particular area. For solving a particular problem, research is undertaken. It becomes easier to commercialize this type of invention. Because what is the problem is identified first and in that direction, research and development activities are carried on. Open invention strategy, In this type of research, more than one company enters into an agreement of research. The benefits of the outcome of this research will be shared by both companies. Internal and external ideas and technologies are involved. These ideas are coordinated and then they are commercialized. Due to globalization, people from the outside company are also invited to undertake research and development activities. Remuneration of employees of both the companies can increase and the benefit of commercialization of such invention is also shared by both the companies. Conclusion: In modern days, not only can tangible assets be commercialized, but also intangible assets can also be commercialized. The benefits of such an invention can be shared by more than one company. Competition can become more meaningful. Pay packages of employees can improve. It Is a need for time to adopt such strategies to benefit employees, competitors, stakeholders.Keywords: innovation, protection, management, commercialization
Procedia PDF Downloads 16832759 A Genetic Algorithm Approach for Multi Constraint Team Orienteering Problem with Time Windows
Authors: Uyanga Sukhbaatar, Ahmed Lbath, Mendamar Majig
Abstract:
The Orienteering Problem is the most known example to start modeling tourist trip design problem. In order to meet tourist’s interest and constraint the OP is becoming more and more complicate to solve. The Multi Constraint Team Orienteering Problem with Time Windows is the last extension of the OP which differentiates from other extensions by including more extra associated constraints. The goal of the MCTOPTW is maximizing tourist’s satisfaction score in same time not to violate any of these constraints. This paper presents a genetic algorithmic approach to tackle the MCTOPTW. The benchmark data from literature is tested by our algorithm and the performance results are compared.Keywords: multi constraint team orienteering problem with time windows, genetic algorithm, tour planning system
Procedia PDF Downloads 62732758 Using Q-Learning to Auto-Tune PID Controller Gains for Online Quadcopter Altitude Stabilization
Authors: Y. Alrubyli
Abstract:
Unmanned Arial Vehicles (UAVs), and more specifically, quadcopters need to be stable during their flights. Altitude stability is usually achieved by using a PID controller that is built into the flight controller software. Furthermore, the PID controller has gains that need to be tuned to reach optimal altitude stabilization during the quadcopter’s flight. For that, control system engineers need to tune those gains by using extensive modeling of the environment, which might change from one environment and condition to another. As quadcopters penetrate more sectors, from the military to the consumer sectors, they have been put into complex and challenging environments more than ever before. Hence, intelligent self-stabilizing quadcopters are needed to maneuver through those complex environments and situations. Here we show that by using online reinforcement learning with minimal background knowledge, the altitude stability of the quadcopter can be achieved using a model-free approach. We found that by using background knowledge instead of letting the online reinforcement learning algorithm wander for a while to tune the PID gains, altitude stabilization can be achieved faster. In addition, using this approach will accelerate development by avoiding extensive simulations before applying the PID gains to the real-world quadcopter. Our results demonstrate the possibility of using the trial and error approach of reinforcement learning combined with background knowledge to achieve faster quadcopter altitude stabilization in different environments and conditions.Keywords: reinforcement learning, Q-leanring, online learning, PID tuning, unmanned aerial vehicle, quadcopter
Procedia PDF Downloads 17432757 Challenges in Employment and Adjustment of Academic Expatriates Based in Higher Education Institutions in the KwaZulu-Natal Province, South Africa
Authors: Thulile Ndou
Abstract:
The purpose of this study was to examine the challenges encountered in the mediation of attracting and recruiting academic expatriates who in turn encounter their own obstacles in adjusting into and settling in their host country, host academic institutions and host communities. The none-existence of literature on attraction, placement and management of academic expatriates in the South African context has been acknowledged. Moreover, Higher Education Institutions in South Africa have voiced concerns relating to delayed and prolonged recruitment and selection processes experienced in the employment process of academic expatriates. Once employed, academic expatriates should be supported and acquainted with the surroundings, the local communities as well as be assisted to establish working relations with colleagues in order to facilitate their adjustment and integration process. Hence, an employer should play a critical role in facilitating the adjustment of academic expatriates. This mixed methods study was located in four Higher Education Institutions based in the KwaZulu-Natal province, in South Africa. The explanatory sequential design approach was deployed in the study. The merits of this approach were chiefly that it employed both the quantitative and qualitative techniques of inquiry. Therefore, the study examined and interrogated its subject from a multiplicity of quantitative and qualitative vantage points, yielding a much more enriched and enriching illumination. Mixing the strengths of both the quantitative and the qualitative techniques delivered much more durable articulation and understanding of the subject. A 5-point Likert scale questionnaire was used to collect quantitative data relating to interaction adjustment, general adjustment and work adjustment from academic expatriates. One hundred and forty two (142) academic expatriates participated in the quantitative study. Qualitative data relating to employment process and support offered to academic expatriates was collected through a structured questionnaire and semi-structured interviews. A total of 48 respondents; including, line managers, human resources practitioners, and academic expatriates participated in the qualitative study. The Independent T-test, ANOVA and Descriptive Statistics were performed to analyse, interpret and make meaning of quantitative data and thematic analysis was used to analyse qualitative data. The qualitative results revealed that academic talent is sourced from outside the borders of the country because of the academic skills shortage in almost all academic disciplines especially in the disciplines associated with Science, Engineering and Accounting. However, delays in work permit application process made it difficult to finalise the recruitment and selection process on time. Furthermore, the quantitative results revealed that academic expatriates experience general and interaction adjustment challenges associated with the use of local language and understanding of local culture. However, female academic expatriates were found to be better adjusted in the two areas as compared to male academic expatriates. Moreover, significant mean differences were found between institutions suggesting that academic expatriates based in rural areas experienced adjustment challenges differently from the academic expatriates based in urban areas. The study gestured to the need for policy revisions in the area of immigration, human resources and academic administration.Keywords: academic expatriates, recruitment and selection, interaction and general adjustment, work adjustment
Procedia PDF Downloads 30632756 Coupled Analysis for Hazard Modelling of Debris Flow Due to Extreme Rainfall
Authors: N. V. Nikhil, S. R. Lee, Do Won Park
Abstract:
Korean peninsula receives about two third of the annual rainfall during summer season. The extreme rainfall pattern due to typhoon and heavy rainfall results in severe mountain disasters among which 55% of them are debris flows, a major natural hazard especially when occurring around major settlement areas. The basic mechanism underlined for this kind of failure is the unsaturated shallow slope failure by reduction of matric suction due to infiltration of water and liquefaction of the failed mass due to generation of positive pore water pressure leading to abrupt loss of strength and commencement of flow. However only an empirical model cannot simulate this complex mechanism. Hence, we have employed an empirical-physical based approach for hazard analysis of debris flow using TRIGRS, a debris flow initiation criteria and DAN3D in mountain Woonmyun, South Korea. Debris flow initiation criteria is required to discern the potential landslides which can transform into debris flow. DAN-3D, being a new model, does not have the calibrated values of rheology parameters for Korean conditions. Thus, in our analysis we have used the recent 2011 debris flow event in mountain Woonmyun san for calibration of both TRIGRS model and DAN-3D, thereafter identifying and predicting the debris flow initiation points, path, run out velocity, and area of spreading for future extreme rainfall based scenarios.Keywords: debris flow, DAN-3D, extreme rainfall, hazard analysis
Procedia PDF Downloads 24732755 Community-based Mapping as a Planning Tool; Examples from Pakistan
Authors: Noman Ahmed, Fariha Tahseen
Abstract:
Since several decades, unplanned urbanization and rapid growth of informal settlements have evolved and increased in size and number. Large cities such as Karachi have been impacted with sprawl and rising share of unplanned settlements where poor communities reside. Threats of eviction, deteriorating law and order situation, lack of essential amenities and infrastructure, extortion and bullying from local and non-local musclemen and feeble response of government agencies towards their development needs are some predicaments. Non-governmental organizations (NGOs) have caused important interventions in such locations. Appraisal of the community-based mapping as a tool in supporting the development work in less privileged areas in Karachi has been the objective of this research. The Orangi Pilot Project (OPP), under the leadership of its slain director Perween Rahman had a significant role to play in developing and extending this approach in low income locations in Karachi and beyond. The paper investigates the application of mapping in the process of peri urban land invasion causing rapid transformation of traditional settlements in Karachi. Mixed methodology components comprising literature review, archival research, and unstructured interviews with key informants and case studies have been used.Keywords: squatters (katchi abadis), land grabbing, community empowerment, housing rights, mapping, infrastructure development
Procedia PDF Downloads 31532754 The Political Philosophy of Sikhism as a Resistance to Autocratic Regimes
Authors: Vishavjeet Singh Goraya, Bashir Ahmed Dar
Abstract:
Sikhism is one of the youngest religions which came into existence only 550 years ago with Guru Nanak Dev. Guru Nanak and his successors not only gave religious sermons but also made bold commentaries on the politico-social situations of their times. Sikhism has consistently stood against autocratic regimes and societal polarization. The Sikh Gurus envisioned a political system based on equality, brotherhood, freedom of conscience, justice, and secularism where there was no discrimination and human rights of all were protected. The Sikh Gurus visualized a society based on an amalgamation of politics and religion without compromising morality and resisting any form of oppression and tyranny. The fifth guru of the Sikhs, Guru Arjan Dev, compiled the verses of his predecessors and many other saints belonging to different religions and castes in the form of Adi Granth, highlighting the egalitarian and secular foundations of Sikhism. Later tenth guru, Guru Gobind Singh, gave a concrete shape to the concepts of the earlier gurus written in Adi Granth and created an openly political and sovereign community known as Khalsa Panth with a distinct identity. This paper will use the qualitative historical approach to highlight Sikhism’s enduring commitment to fight oppression and tyranny, and it also examines how Sikhism offers solutions to contemporary issues of authoritarianism and polarization.Keywords: sikhism, Gurus, political thought, equality, scriptures
Procedia PDF Downloads 1632753 Prediction of Compressive Strength in Geopolymer Composites by Adaptive Neuro Fuzzy Inference System
Authors: Mehrzad Mohabbi Yadollahi, Ramazan Demirboğa, Majid Atashafrazeh
Abstract:
Geopolymers are highly complex materials which involve many variables which makes modeling its properties very difficult. There is no systematic approach in mix design for Geopolymers. Since the amounts of silica modulus, Na2O content, w/b ratios and curing time have a great influence on the compressive strength an ANFIS (Adaptive neuro fuzzy inference system) method has been established for predicting compressive strength of ground pumice based Geopolymers and the possibilities of ANFIS for predicting the compressive strength has been studied. Consequently, ANFIS can be used for geopolymer compressive strength prediction with acceptable accuracy.Keywords: geopolymer, ANFIS, compressive strength, mix design
Procedia PDF Downloads 85332752 A Robust System for Foot Arch Type Classification from Static Foot Pressure Distribution Data Using Linear Discriminant Analysis
Authors: R. Periyasamy, Deepak Joshi, Sneh Anand
Abstract:
Foot posture assessment is important to evaluate foot type, causing gait and postural defects in all age groups. Although different methods are used for classification of foot arch type in clinical/research examination, there is no clear approach for selecting the most appropriate measurement system. Therefore, the aim of this study was to develop a system for evaluation of foot type as clinical decision-making aids for diagnosis of flat and normal arch based on the Arch Index (AI) and foot pressure distribution parameter - Power Ratio (PR) data. The accuracy of the system was evaluated for 27 subjects with age ranging from 24 to 65 years. Foot area measurements (hind foot, mid foot, and forefoot) were acquired simultaneously from foot pressure intensity image using portable PedoPowerGraph system and analysis of the image in frequency domain to obtain foot pressure distribution parameter - PR data. From our results, we obtain 100% classification accuracy of normal and flat foot by using the linear discriminant analysis method. We observe there is no misclassification of foot types because of incorporating foot pressure distribution data instead of only arch index (AI). We found that the mid-foot pressure distribution ratio data and arch index (AI) value are well correlated to foot arch type based on visual analysis. Therefore, this paper suggests that the proposed system is accurate and easy to determine foot arch type from arch index (AI), as well as incorporating mid-foot pressure distribution ratio data instead of physical area of contact. Hence, such computational tool based system can help the clinicians for assessment of foot structure and cross-check their diagnosis of flat foot from mid-foot pressure distribution.Keywords: arch index, computational tool, static foot pressure intensity image, foot pressure distribution, linear discriminant analysis
Procedia PDF Downloads 49932751 Failure Probability Assessment of Concrete Spherical Domes Subjected to Ventilation Controlled Fires Using BIM Tools
Authors: A. T. Kassem
Abstract:
Fires areconsidered a common hazardous action that any building may face. Most buildings’ structural elements are designed, taking into consideration precautions for fire safety, using deterministic design approaches. Public and highly important buildings are commonly designed considering standard fire rating and, in many cases, contain large compartments with central domes. Real fire scenarios are not commonly brought into action in structural design of buildings because of complexities in both scenarios and analysis tools. This paper presents a modern approach towards analysis of spherical domes in real fire condition via implementation of building information modelling, and adopting a probabilistic approach. BIMhas been implemented to bridge the gap between various software packages enabling them to function interactively to model both real fire and corresponding structural response. Ventilation controlled fires scenarios have been modeled using both “Revit” and “Pyrosim”. Monte Carlo simulation has been adopted to engage the probabilistic analysis approach in dealing with various parameters. Conclusions regarding failure probability and fire endurance, in addition to the effects of various parameters, have been extracted.Keywords: concrete, spherical domes, ventilation controlled fires, BIM, monte carlo simulation, pyrosim, revit
Procedia PDF Downloads 9632750 A Semi-supervised Classification Approach for Trend Following Investment Strategy
Authors: Rodrigo Arnaldo Scarpel
Abstract:
Trend following is a widely accepted investment strategy that adopts a rule-based trading mechanism that rather than striving to predict market direction or on information gathering to decide when to buy and when to sell a stock. Thus, in trend following one must respond to market’s movements that has recently happen and what is currently happening, rather than on what will happen. Optimally, in trend following strategy, is to catch a bull market at its early stage, ride the trend, and liquidate the position at the first evidence of the subsequent bear market. For applying the trend following strategy one needs to find the trend and identify trade signals. In order to avoid false signals, i.e., identify fluctuations of short, mid and long terms and to separate noise from real changes in the trend, most academic works rely on moving averages and other technical analysis indicators, such as the moving average convergence divergence (MACD) and the relative strength index (RSI) to uncover intelligible stock trading rules following trend following strategy philosophy. Recently, some works has applied machine learning techniques for trade rules discovery. In those works, the process of rule construction is based on evolutionary learning which aims to adapt the rules to the current environment and searches for the global optimum rules in the search space. In this work, instead of focusing on the usage of machine learning techniques for creating trading rules, a time series trend classification employing a semi-supervised approach was used to early identify both the beginning and the end of upward and downward trends. Such classification model can be employed to identify trade signals and the decision-making procedure is that if an up-trend (down-trend) is identified, a buy (sell) signal is generated. Semi-supervised learning is used for model training when only part of the data is labeled and Semi-supervised classification aims to train a classifier from both the labeled and unlabeled data, such that it is better than the supervised classifier trained only on the labeled data. For illustrating the proposed approach, it was employed daily trade information, including the open, high, low and closing values and volume from January 1, 2000 to December 31, 2022, of the São Paulo Exchange Composite index (IBOVESPA). Through this time period it was visually identified consistent changes in price, upwards or downwards, for assigning labels and leaving the rest of the days (when there is not a consistent change in price) unlabeled. For training the classification model, a pseudo-label semi-supervised learning strategy was used employing different technical analysis indicators. In this learning strategy, the core is to use unlabeled data to generate a pseudo-label for supervised training. For evaluating the achieved results, it was considered the annualized return and excess return, the Sortino and the Sharpe indicators. Through the evaluated time period, the obtained results were very consistent and can be considered promising for generating the intended trading signals.Keywords: evolutionary learning, semi-supervised classification, time series data, trading signals generation
Procedia PDF Downloads 8932749 Effect of Chemical Additive on Fixed Abrasive Polishing of LBO Crystal with Non-Water Based Slurry
Authors: Jun Li, Wenze Wang, Zhanggui Hu, Yongwei Zhu, Dunwen Zuo
Abstract:
Non-water based fixed abrasive polishing was adopted to manufacture LBO crystal for nano precision surface quality because of its deliquescent. Ethyl alcohol was selected as the non-water based slurry solvent and ethanediamine, lactic acid, hydrogen peroxide were add in the slurry as a chemical additive, respectively. Effect of different additives with non-water based slurry on material removal rate, surface topography, microscopic appearances and surface roughness were investigated in fixed abrasive polishing of LBO crystal. The results show the best surface quality of LBO crystal with surface roughness Sa 8.2 nm and small damages was obtained by non-water based slurry with lactic acid. Non-water based fixed abrasive polishing can achieve nano precision surface quality of LBO crystal with high material removal.Keywords: non-water based slurry, LBO crystal, fixed abrasive polishing, surface roughness
Procedia PDF Downloads 47332748 Looking beyond Corporate Social Responsibility to Sustainable Development: Conceptualisation and Theoretical Exploration
Authors: Mercy E. Makpor
Abstract:
Traditional Corporate Social Responsibility (CSR) idea has gone beyond just ensuring safety environments, caring about global warming and ensuring good living standards and conditions for the society at large. The paradigm shift is towards a focus on strategic objectives and the long-term value creation for both businesses and the society at large for a realistic future. As an important approach to solving social and environment issues, CSR has been accepted globally. Yet the approach is expected to go beyond where it is currently. So much is expected from businesses and governments at every level globally and locally. This then leads to the original idea of the concept, that is, how it originated and how it has been perceived over the years. Little wonder there has been a lot of definitions surrounding the concept without a major globally acceptable definition of it. The definition of CSR given by the European Commission will be considered for the purpose of this paper. Sustainable Development (SD), on the other hand, has been viewed in recent years as an ethical concept explained in the UN-Report termed “Our Common Future,” which can also be referred to as the Brundtland report. The report summarises the need for SD to take place in the present without comprising the future. However, the recent 21st-century framework on sustainability known as the “Triple Bottom Line (TBL)” framework, has added its voice to the concepts of CSR and sustainable development. The TBL model is of the opinion that businesses should not only report on their financial performance but also on their social and environmental performances, highlighting that CSR has gone beyond just the “material-impact” approach towards a “Future-Oriented” approach (sustainability). In this paper, the concept of CSR is revisited by exploring the various theories therein. The discourse on the concepts of sustainable development and sustainable development frameworks will also be indicated, thereby inducing these into how CSR can benefit both businesses and their stakeholders as well as the entirety of the society, not just for the present but for the future. It does this by exploring the importance of both concepts (CSR and SD) and concludes by making recommendations for a more empirical research in the near future.Keywords: corporate social responsibility, sustainable development, sustainability, triple bottom line model
Procedia PDF Downloads 24532747 Design and Optimization of a Small Hydraulic Propeller Turbine
Authors: Dario Barsi, Marina Ubaldi, Pietro Zunino, Robert Fink
Abstract:
A design and optimization procedure is proposed and developed to provide the geometry of a high efficiency compact hydraulic propeller turbine for low head. For the preliminary design of the machine, classic design criteria, based on the use of statistical correlations for the definition of the fundamental geometric parameters and the blade shapes are used. These relationships are based on the fundamental design parameters (i.e., specific speed, flow coefficient, work coefficient) in order to provide a simple yet reliable procedure. Particular attention is paid, since from the initial steps, on the correct conformation of the meridional channel and on the correct arrangement of the blade rows. The preliminary geometry thus obtained is used as a starting point for the hydrodynamic optimization procedure, carried out using a CFD calculation software coupled with a genetic algorithm that generates and updates a large database of turbine geometries. The optimization process is performed using a commercial approach that solves the turbulent Navier Stokes equations (RANS) by exploiting the axial-symmetric geometry of the machine. The geometries generated within the database are therefore calculated in order to determine the corresponding overall performance. In order to speed up the optimization calculation, an artificial neural network (ANN) based on the use of an objective function is employed. The procedure was applied for the specific case of a propeller turbine with an innovative design of a modular type, specific for applications characterized by very low heads. The procedure is tested in order to verify its validity and the ability to automatically obtain the targeted net head and the maximum for the total to total internal efficiency.Keywords: renewable energy conversion, hydraulic turbines, low head hydraulic energy, optimization design
Procedia PDF Downloads 15032746 Considering Partially Developed Artifacts in Change Impact Analysis Implementation
Authors: Nazri Kama, Sufyan Basri, Roslina Ibrahim
Abstract:
It is important to manage the changes in the software to meet the evolving needs of the customer. Accepting too many changes causes delay in the completion and it incurs additional cost. One type of information that helps to make the decision is through change impact analysis. Current impact analysis approaches assume that all classes in the class artifact are completely developed and the class artifact is used as a source of analysis. However, these assumptions are impractical for impact analysis in the software development phase as some classes in the class artifact are still under development or partially developed that leads to inaccuracy. This paper presents a novel impact analysis approach to be used in the software development phase. The significant achievements of the approach are demonstrated through an extensive experimental validation using three case studies.Keywords: software development, impact analysis, traceability, static analysis.
Procedia PDF Downloads 60832745 Water Quality Trading with Equitable Total Maximum Daily Loads
Authors: S. Jamshidi, E. Feizi Ashtiani, M. Ardestani, A. Feizi Ashtiani
Abstract:
Waste load allocation (WLA) strategies usually intend to find economical policies for water resource management. Water quality trading (WQT) is an approach that uses discharge permit market to reduce total environmental protection costs. This primarily requires assigning discharge limits known as total maximum daily loads (TMDLs). These are determined by monitoring organizations with respect to the receiving water quality and remediation capabilities. The purpose of this study is to compare two approaches of TMDL assignment for WQT policy in small catchment area of Haraz River, in north of Iran. At first, TMDLs are assigned uniformly for the whole point sources to keep the concentrations of BOD and dissolved oxygen (DO) at the standard level at checkpoint (terminus point). This was simply simulated and controlled by Qual2kw software. In the second scenario, TMDLs are assigned using multi objective particle swarm optimization (MOPSO) method in which the environmental violation at river basin and total treatment costs are minimized simultaneously. In both scenarios, the equity index and the WLA based on trading discharge permits (TDP) are calculated. The comparative results showed that using economically optimized TMDLs (2nd scenario) has slightly more cost savings rather than uniform TMDL approach (1st scenario). The former annually costs about 1 M$ while the latter is 1.15 M$. WQT can decrease these annual costs to 0.9 and 1.1 M$, respectively. In other word, these approaches may save 35 and 45% economically in comparison with command and control policy. It means that using multi objective decision support systems (DSS) may find more economical WLA, however its outcome is not necessarily significant in comparison with uniform TMDLs. This may be due to the similar impact factors of dischargers in small catchments. Conversely, using uniform TMDLs for WQT brings more equity that makes stakeholders not feel that much envious of difference between TMDL and WQT allocation. In addition, for this case, determination of TMDLs uniformly would be much easier for monitoring. Consequently, uniform TMDL for TDP market is recommended as a sustainable approach. However, economical TMDLs can be used for larger watersheds.Keywords: waste load allocation (WLA), water quality trading (WQT), total maximum daily loads (TMDLs), Haraz River, multi objective particle swarm optimization (MOPSO), equity
Procedia PDF Downloads 39432744 Blended Cloud Based Learning Approach in Information Technology Skills Training and Paperless Assessment: Case Study of University of Cape Coast
Authors: David Ofosu-Hamilton, John K. E. Edumadze
Abstract:
Universities have come to recognize the role Information and Communication Technology (ICT) skills plays in the daily activities of tertiary students. The ability to use ICT – essentially, computers and their diverse applications – are important resources that influence an individual’s economic and social participation and human capital development. Our society now increasingly relies on the Internet, and the Cloud as a means to communicate and disseminate information. The educated individual should, therefore, be able to use ICT to create and share knowledge that will improve society. It is, therefore, important that universities require incoming students to demonstrate a level of computer proficiency or trained to do so at a minimal cost by deploying advanced educational technologies. The training and standardized assessment of all in-coming first-year students of the University of Cape Coast in Information Technology Skills (ITS) have become a necessity as students’ most often than not highly overestimate their digital skill and digital ignorance is costly to any economy. The one-semester course is targeted at fresh students and aimed at enhancing the productivity and software skills of students. In this respect, emphasis is placed on skills that will enable students to be proficient in using Microsoft Office and Google Apps for Education for their academic work and future professional work whiles using emerging digital multimedia technologies in a safe, ethical, responsible, and legal manner. The course is delivered in blended mode - online and self-paced (student centered) using Alison’s free cloud-based tutorial (Moodle) of Microsoft Office videos. Online support is provided via discussion forums on the University’s Moodle platform and tutor-directed and assisted at the ICT Centre and Google E-learning laboratory. All students are required to register for the ITS course during either the first or second semester of the first year and must participate and complete it within a semester. Assessment focuses on Alison online assessment on Microsoft Office, Alison online assessment on ALISON ABC IT, Peer assessment on e-portfolio created using Google Apps/Office 365 and an End of Semester’s online assessment at the ICT Centre whenever the student was ready in the cause of the semester. This paper, therefore, focuses on the digital culture approach of hybrid teaching, learning and paperless examinations and the possible adoption by other courses or programs at the University of Cape Coast.Keywords: assessment, blended, cloud, paperless
Procedia PDF Downloads 24832743 Understanding John H. Johnson and Ebony Magazine Financial Responsiveness to Rise of Black Power in the U.S, 1966
Authors: Sid Ahmed Ziane
Abstract:
This paper argues for Johnson's financial responsiveness to the rise of Black Power and its advocate, 'Stokely Carmichael' in 1966. John H. Johnson was a Black businessman and the owner of Ebony magazine, one of the widely read Black magazines in the U.S. His magazine, however, was designed only to promoting Black fashion, aesthetic, marketing, and consumerism. In mid-1966, the mainstream of the Civil Rights movement was heading into two distinct camps when some of its advocates, led by Stokely Carmichael, began to question the slow pace of the Civil Rights and sought to pursue a more radical approach to bring about upheaval to the Black community. Their new approach, however, propelled the national media into paying close attention to their activities, their new methods, and their radical orientations. In fact, the major White-oriented media discredited Carmichael and distorted his public image via sensational stories and race-mongering reports. However, the Black owned outlets such as The Liberator advocated his agendas, whereas other magazines such as The Crisis rejected them. Based on such oral sources and Ebony’s online issues, this paper adds and argues that Johnson had also responded to the rise of Black Power and Carmichael. This reaction had, in fact, aimed at scooping and selling Carmichael and his new orientation as well as advertising him in his magazine to attract the readers who showed a strong tendency to hear and read about the heyday and even the ferment of Black Power. This paper is part of an ongoing project which aims at framing our understanding of how the Black print media and the modern Black liberation struggle were correlated and could shape each other by appraising their agendas, milestones, and their pivotal figures.Keywords: Black power, Ebony magazine, John Johson, Stokely Carmichael
Procedia PDF Downloads 17732742 Geovisualisation for Defense Based on a Deep Learning Monocular Depth Reconstruction Approach
Authors: Daniel R. dos Santos, Mateus S. Maldonado, Estevão J. R. Batista
Abstract:
The military commanders increasingly dependent on spatial awareness, as knowing where enemy are, understanding how war battle scenarios change over time, and visualizing these trends in ways that offer insights for decision-making. Thanks to advancements in geospatial technologies and artificial intelligence algorithms, the commanders are now able to modernize military operations on a universal scale. Thus, geovisualisation has become an essential asset in the defense sector. It has become indispensable for better decisionmaking in dynamic/temporal scenarios, operation planning and management for the war field, situational awareness, effective planning, monitoring, and others. For example, a 3D visualization of war field data contributes to intelligence analysis, evaluation of postmission outcomes, and creation of predictive models to enhance decision-making and strategic planning capabilities. However, old-school visualization methods are slow, expensive, and unscalable. Despite modern technologies in generating 3D point clouds, such as LIDAR and stereo sensors, monocular depth values based on deep learning can offer a faster and more detailed view of the environment, transforming single images into visual information for valuable insights. We propose a dedicated monocular depth reconstruction approach via deep learning techniques for 3D geovisualisation of satellite images. It introduces scalability in terrain reconstruction and data visualization. First, a dataset with more than 7,000 satellite images and associated digital elevation model (DEM) is created. It is based on high resolution optical and radar imageries collected from Planet and Copernicus, on which we fuse highresolution topographic data obtained using technologies such as LiDAR and the associated geographic coordinates. Second, we developed an imagery-DEM fusion strategy that combine feature maps from two encoder-decoder networks. One network is trained with radar and optical bands, while the other is trained with DEM features to compute dense 3D depth. Finally, we constructed a benchmark with sparse depth annotations to facilitate future research. To demonstrate the proposed method's versatility, we evaluated its performance on no annotated satellite images and implemented an enclosed environment useful for Geovisualisation applications. The algorithms were developed in Python 3.0, employing open-source computing libraries, i.e., Open3D, TensorFlow, and Pythorch3D. The proposed method provides fast and accurate decision-making with GIS for localization of troops, position of the enemy, terrain and climate conditions. This analysis enhances situational consciousness, enabling commanders to fine-tune the strategies and distribute the resources proficiently.Keywords: depth, deep learning, geovisualisation, satellite images
Procedia PDF Downloads 1032741 The Effect of Cross-Curriculum of L1 and L2 on Elementary School Students’ Linguistic Proficiency: To Sympathize with Others
Authors: Reiko Yamamoto
Abstract:
This paper reports on a project to integrate Japanese (as a first language) and English (as a second language) education. This study focuses on the mutual effects of the two languages on the linguistic proficiency of elementary school students. The research team consisted of elementary school teachers and researchers at a university. The participants of the experiment were students between 3rd and 6th grades at an elementary school. The research process consisted of seven steps: 1) specifying linguistic proficiency; 2) developing the cross-curriculum of L1 and L2; 3) forming can-do statements; 4) creating a self-evaluation questionnaire; 5) executing the self-evaluation questionnaire at the beginning of the school year; 6) instructing L1 and L2 based on the curriculum; and 7) executing the self-evaluation questionnaire at the beginning of the next school year. In Step 1, the members of the research team brainstormed ways to specify elementary school students’ linguistic proficiency that can be observed in various scenes. It was revealed that the teachers evaluate their students’ linguistic proficiency on the basis of the students’ utterances, but also informed by their non-verbal communication abilities. This led to the idea that competency for understanding others’ minds through the use of physical movement or bodily senses in communication in L1 – to sympathize with others – can be transferred to that same competency in communication in L2. Based on the specification of linguistic proficiency that L1 and L2 have in common, a cross-curriculum of L1 and L2 was developed in Step 2. In Step 3, can-do statements based on the curriculum were also formed, building off of the action-oriented approach from the Common European Framework of Reference for Languages (CEFR) used in Europe. A self-evaluation questionnaire consisting of the main can-do statements was given to the students between 3rd grade and 6th grade at the beginning of the school year (Step 4 and Step 5), and all teachers gave L1 and L2 instruction based on the curriculum to the students for one year (Step 6). The same questionnaire was given to the students at the beginning of the next school year (Step 7). The results of statistical analysis proved the enhancement of the students’ linguistic proficiency. This verified the validity of developing the cross-curriculum of L1 and L2 and adapting it in elementary school. It was concluded that elementary school students do not distinguish between L1 and L2, and that they just try to understand others’ minds through physical movement or senses in any language.Keywords: cross curriculum of L1 and L2, elementary school education, language proficiency, sympathy with others
Procedia PDF Downloads 43832740 Neural Network Approach For Clustering Host Community: Based on Perceptions Toward Tourism, Their Satisfaction Level and Demographic Attributes in Iran (Lahijan)
Authors: Nasibeh Mohammadpour, Ali Rajabzadeh, Adel Azar, Hamid Zargham Borujeni,
Abstract:
Generally, various industries development depends on their stakeholders and beneficiaries supports. One of the most important stakeholders in tourism industry ( which has become one of the most important lucrative and employment-generating activities at the international level these days) are host communities in tourist destination which are affected and effect on this industry development. Recognizing host community and its segmentations can be important to get their support for future decisions and policy making. In order to identify these segments, in this study, clustering of the residents has been done by using some tools that are designed to encounter human complexities and have ability to model and generalize complex systems without any needs for the initial clusters’ seeds like classic methods. Neural networks can help to meet these expectations. The research have been planned to design neural networks-based mathematical model for clustering the host community effectively according to multi criteria, and identifies differences among segments. In order to achieve this goal, the residents’ segmentation has been done by demographic characteristics, their attitude towards the tourism development, the level of satisfaction and the type of their support in this field. The applied method is self-organized neural networks and the results have compared with K-means. As the results show, the use of Self- Organized Map (SOM) method provides much better results by considering the Cophenetic correlation and between clusters variance coefficients. Based on these criteria, the host community is divided into five sections with unique and distinctive features, which are in the best condition (in comparison other modes) according to Cophenetic correlation coefficient of 0.8769 and between clusters variance of 0.1412.Keywords: Artificial Nural Network, Clustering , Resident, SOM, Tourism
Procedia PDF Downloads 18332739 A Long-Standing Methodology Quest Regarding Commentary of the Qur’an: Modern Debates on Function of Hermeneutics in the Quran Scholarship in Turkey
Authors: Merve Palanci
Abstract:
This paper aims to reveal and analyze methodology debates on Qur’an Commentary in Turkish Scholarship and to make sound inductions on the current situation, with reference to the literature evolving around the credibility of Hermeneutics when the case is Qur’an commentary and methodological connotations related to it, together with the other modern approaches to the Qur’an. It is fair to say that Tafseer, constituting one of the main parts of basic Islamic sciences, has drawn great attention from both Muslim and non-Muslim scholars for a long time. And with the emplacement of an acute junction between natural sciences and social sciences in the post-enlightenment period, this interest seems to pave the way for methodology discussions that are conducted by theology spheres, occupying a noticeable slot in Tafseer literature, as well. A panoramic glance at the classical treatise in relation to the methodology of Tafseer, namely Usul al-Tafseer, leads the reader to the conclusion that these classics are intrinsically aimed at introducing the Qur’an and its early history of formation as a corpus and providing a better understanding of its content. To illustrate, the earliest methodology work extant for Qur’an commentary, al- Aql wa’l Fahm al- Qur’an by Harith al-Muhasibi covers content that deals with Qur’an’s rhetoric, its muhkam and mutashabih, and abrogation, etc. And most of the themes in question are evident to share a common ground: understanding the Scripture and producing an accurate commentary to be built on this preliminary phenomenon of understanding. The content of other renowned works in an overtone of Tafseer methodology, such as Funun al Afnan, al- Iqsir fi Ilm al- Tafseer, and other succeeding ones al- Itqan and al- Burhan is also rich in hints related to preliminary phenomena of understanding. However, these works are not eligible for being classified as full-fledged methodology manuals assuring a true understanding of the Qur’an. And Hermeneutics is believed to supply substantial data applicable to Qur’an commentary as it deals with the nature of understanding itself. Referring to the latest tendencies in Tafseer methodology, this paper envisages to centralize hermeneutical debates in modern scholarship of Qur’an commentary and the incentives that lead scholars to apply for Hermeneutics in Tafseer literature. Inspired from these incentives, the study involves three parts. In the introduction part, this paper introduces key features of classical methodology works in general terms and traces back the main methodological shifts of modern times in Qur’an commentary. To this end, revisionist Ecole, scientific Qur’an commentary ventures, and thematic Qur’an commentary are included and analysed briefly. However, historical-critical commentary on the Quran, as it bears a close relationship with hermeneutics, is handled predominantly. The second part is based on the hermeneutical nature of understanding the Scripture, revealing a timeline for the beginning of hermeneutics debates in Tafseer, and Fazlur Rahman’s(d.1988) influence will be manifested for establishing a theoretical bridge. In the following part, reactions against the application of Hermeneutics in Tafseer activity and pro-hermeneutics works will be revealed through cross-references to the prominent figures of both, and the literature in question in theology scholarship in Turkey will be explored critically.Keywords: hermeneutics, Tafseer, methodology, Ulum al- Qur’an, modernity
Procedia PDF Downloads 7532738 Corporate Governance Reforms in a Developing Economy: Making a Case for Upstream and Downstream Interventions
Authors: Franklin Nakpodia, Femi Olan
Abstract:
A blend of internal factors (firm performance, internal stakeholders) and external pressures (globalisation, technology, corporate scandals) have intensified calls for corporate governance reforms. While several countries and their governments have responded to these calls, the effect of such reforms on corporate governance systems across countries remains mixed. In particular, the literature reports that the effectiveness of corporate governance interventions in many developing economies is limited. Relying on the corporate governance system in Africa’s largest economy (Nigeria), this research addresses two issues. First, this study explores why previous corporate governance reforms have failed and second, the article investigates what reforms could improve corporate governance practices in the country. In addressing the above objectives, this study adopts a qualitative approach that permits data collection via semi-structured interviews with 21 corporate executives. The data supports the articulation of two sequential levels of reforms (i.e., the upstream and downstream reforms). The upstream reforms focus on two crucial but often overlooked areas that undermine reform effectiveness, i.e., the extent of government commitment and an enabling environment. The downstream reforms combine awareness and regulatory elements to proffer a path to robust corporate governance in the country. Furthermore, findings from this study stress the need to consider the use of a bottom-up approach to corporate governance practice and policymaking in place of the dominant top-down strategy.Keywords: bottom-up approach, corporate governance, reforms, regulation
Procedia PDF Downloads 20232737 GIS-Based Topographical Network for Minimum “Exertion” Routing
Authors: Katherine Carl Payne, Moshe Dror
Abstract:
The problem of minimum cost routing has been extensively explored in a variety of contexts. While there is a prevalence of routing applications based on least distance, time, and related attributes, exertion-based routing has remained relatively unexplored. In particular, the network structures traditionally used to construct minimum cost paths are not suited to representing exertion or finding paths of least exertion based on road gradient. In this paper, we introduce a topographical network or “topograph” that enables minimum cost routing based on the exertion metric on each arc in a given road network as it is related to changes in road gradient. We describe an algorithm for topograph construction and present the implementation of the topograph on a road network of the state of California with ~22 million nodes.Keywords: topograph, RPE, routing, GIS
Procedia PDF Downloads 546