Search results for: dynamic capability approach
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 17886

Search results for: dynamic capability approach

15126 Self-Marketing on Line Person-to-Person Social Media

Authors: Chih-Ping Chen

Abstract:

Today, technology does not necessitate change; rather, social media has afforded a new arena and digital tools for users/individuals to be symbolized and marketed in meaningful exchanges of digital identities. We argue that these symbolic interactions may afford individuals the ability to create and present less restricted Line person-to-person (P2P) chats than would be possible in face-to-face communications. Individuals can select flexible influence strategies to market themselves, which enables them to create and present their digital identities and impressions in alternative ways within a dynamic sociocultural context. Therefore, this paper aims to explore the novel phenomenon of how individuals market themselves to manage their digital identities and impressions to connect with other users through the symbolic interactions created by new digital tools (e.g., stickers). A netnographic approach was developed by applying a triangulated methodology consisting of user self-diary reports, in-depth interviews, and observations. Totally, 20 participants (10 females and 10 males) were of Taiwanese origin, and their ages ranged from 20–47 years old. The findings of this research showed that individuals on Line P2P social media where traditional cultural gender norms have shifted. Both male and female participants market their modern digital identities by adopting a combination of flexible influence tactics/strategies when using digital stickers. Some findings showed that their influence tactics/strategies often flouted Taiwanese cultural gender norms or skirted traditional rules to fit individual or P2P needs. Finally, these findings potentially contributed to the literature regarding the consumer culture theory and symbolic interaction theory in digital marketing and social media fields.

Keywords: Consumer culture theory, Digital sticker, Self-marketing, Impression, Symbolic interaciton

Procedia PDF Downloads 81
15125 Efficient Filtering of Graph Based Data Using Graph Partitioning

Authors: Nileshkumar Vaishnav, Aditya Tatu

Abstract:

An algebraic framework for processing graph signals axiomatically designates the graph adjacency matrix as the shift operator. In this setup, we often encounter a problem wherein we know the filtered output and the filter coefficients, and need to find out the input graph signal. Solution to this problem using direct approach requires O(N3) operations, where N is the number of vertices in graph. In this paper, we adapt the spectral graph partitioning method for partitioning of graphs and use it to reduce the computational cost of the filtering problem. We use the example of denoising of the temperature data to illustrate the efficacy of the approach.

Keywords: graph signal processing, graph partitioning, inverse filtering on graphs, algebraic signal processing

Procedia PDF Downloads 311
15124 The Processing of Context-Dependent and Context-Independent Scalar Implicatures

Authors: Liu Jia’nan

Abstract:

The default accounts hold the view that there exists a kind of scalar implicature which can be processed without context and own a psychological privilege over other scalar implicatures which depend on context. In contrast, the Relevance Theorist regards context as a must because all the scalar implicatures have to meet the need of relevance in discourse. However, in Katsos, the experimental results showed: Although quantitatively the adults rejected under-informative utterance with lexical scales (context-independent) and the ad hoc scales (context-dependent) at almost the same rate, adults still regarded the violation of utterance with lexical scales much more severe than with ad hoc scales. Neither default account nor Relevance Theory can fully explain this result. Thus, there are two questionable points to this result: (1) Is it possible that the strange discrepancy is due to other factors instead of the generation of scalar implicature? (2) Are the ad hoc scales truly formed under the possible influence from mental context? Do the participants generate scalar implicatures with ad hoc scales instead of just comparing semantic difference among target objects in the under- informative utterance? In my Experiment 1, the question (1) will be answered by repetition of Experiment 1 by Katsos. Test materials will be showed by PowerPoint in the form of pictures, and each procedure will be done under the guidance of a tester in a quiet room. Our Experiment 2 is intended to answer question (2). The test material of picture will be transformed into the literal words in DMDX and the target sentence will be showed word-by-word to participants in the soundproof room in our lab. Reading time of target parts, i.e. words containing scalar implicatures, will be recorded. We presume that in the group with lexical scale, standardized pragmatically mental context would help generate scalar implicature once the scalar word occurs, which will make the participants hope the upcoming words to be informative. Thus if the new input after scalar word is under-informative, more time will be cost for the extra semantic processing. However, in the group with ad hoc scale, scalar implicature may hardly be generated without the support from fixed mental context of scale. Thus, whether the new input is informative or not does not matter at all, and the reading time of target parts will be the same in informative and under-informative utterances. People’s mind may be a dynamic system, in which lots of factors would co-occur. If Katsos’ experimental result is reliable, will it shed light on the interplay of default accounts and context factors in scalar implicature processing? We might be able to assume, based on our experiments, that one single dominant processing paradigm may not be plausible. Furthermore, in the processing of scalar implicature, the semantic interpretation and the pragmatic interpretation may be made in a dynamic interplay in the mind. As to the lexical scale, the pragmatic reading may prevail over the semantic reading because of its greater exposure in daily language use, which may also lead the possible default or standardized paradigm override the role of context. However, those objects in ad hoc scale are not usually treated as scalar membership in mental context, and thus lexical-semantic association of the objects may prevent their pragmatic reading from generating scalar implicature. Only when the sufficient contextual factors are highlighted, can the pragmatic reading get privilege and generate scalar implicature.

Keywords: scalar implicature, ad hoc scale, dynamic interplay, default account, Mandarin Chinese processing

Procedia PDF Downloads 323
15123 A Qualitative Study Examining the Process of EFL Course Design from the Perspectives of Teachers

Authors: Iman Al Khalidi

Abstract:

Recently, English has become the language of globalization and technology. In turn, this has resulted in a seemingly bewildering array of influences and trends in the domain of TESOL curriculum. In light of these changes, higher education has to provide a new and more powerful kind of education. It should prepare students to be more engaged citizens, more capable to solve complex problems at work, and well prepared to lead meaningful life. In response to this, universities, colleges, schools, and departments have to work out in light of the requirements and challenges of the global and technological era. Consequently they have to focus on the adoption of contemporary curriculum which goes in line with the pedagogical shifts from teaching –centered approach to learning centered approach. Ideally, there has been noticeable emphasis on the crucial importance of developing and professionalizing teachers in order to engage them in the process of curriculum development and action research. This is a qualitative study that aims at understanding and exploring the process of designing EFL courses by teachers at the tertiary level from the perspectives of the participants in a professional context in TESOL, Department of English, a private college in Oman. It is a case study that stands on the philosophy of the qualitative approach. It employs multi methods for collecting qualitative data: semi-structured interviews with teachers, focus group discussions with students, and document analysis. The collected data have been analyzed qualitatively by adopting Miles and Huberman's Approach using procedures of reduction, coding, displaying and conclusion drawing and verification.

Keywords: course design, components of course design, case study, data analysis

Procedia PDF Downloads 545
15122 A Qualitative Study Examining the Process of Course Design from the Perspectives of Teachers

Authors: Iman Al Khalidi

Abstract:

Recently, English has become the language of globalization and technology. In turn, this has resulted in a seemingly bewildering array of influences and trends in the domain of TESOL curriculum. In light of these changes, higher education has to provide a new and more powerful kind of education. It should prepare students to be more engaged citizens, more capable to solve complex problems at work, and well prepared to lead a meaningful life. In response to this, universities, colleges, schools, and departments have to work out in light of the requirements and challenges of the global and technological era. Consequently, they have to focus on the adoption of contemporary curriculum which goes in line with the pedagogical shifts from teaching –centered approach to learning centered approach. Ideally, there has been noticeable emphasis on the crucial importance of developing and professionalizing teachers in order to engage them in the process of curriculum development and action research. This is a qualitative study that aims at understanding and exploring the process of designing EFL courses by teachers at the tertiary level from the perspectives of the participants in a professional context in TESOL, Department of English, a private college in Oman. It is a case study that stands on the philosophy of the qualitative approach. It employs multi-methods for collecting qualitative data: semi-structured interviews with teachers, focus group discussions with students, and document analysis. The collected data have been analyzed qualitatively by adopting Miles and Huberman's Approach using procedures of reduction, coding, displaying, and conclusion drawing and verification.

Keywords: course design, components of course design, case study, data analysis

Procedia PDF Downloads 442
15121 Design and Biomechanical Analysis of a Transtibial Prosthesis for Cyclists of the Colombian Team Paralympic

Authors: Jhonnatan Eduardo Zamudio Palacios, Oscar Leonardo Mosquera Dussan, Daniel Guzman Perez, Daniel Alfonso Botero Rosas, Oscar Fabian Rubiano Espinosa, Jose Antonio Garcia Torres, Ivan Dario Chavarro, Ivan Ramiro Rodriguez Camacho, Jaime Orlando Rodriguez

Abstract:

The training of cilsitas with some type of disability finds in the technological development an indispensable ally, generating every day advances to contribute to the quality of life allowing to maximize the capacities of the athletes. The performance of a cyclist depends on physiological and biomechanical factors, such as aerodynamic profile, bicycle measurements, connecting rod length, pedaling systems, type of competition, among others. This study particularly focuses on the description of the dynamic model of a transtibial prosthesis for Paralympic cyclists. To make the model, two points are chosen: in the radius centers of rotation of the plate and pinion of the track bicycle. The parametric scheme of the track bike represents a model of 6 degrees of freedom due to the displacement in X - Y of each of the reference points of the angles of the curve profile β, cant of the velodrome α and the angle of rotation of the connecting rod φ. The force exerted on the crank of the bicycle varies according to the angles of the curve profile β, the velodrome cant of α and the angle of rotation of the crank φ. The behavior is analyzed through the Matlab R2015a software. The average strength that a cyclist exerts on the cranks of a bicycle is 1,607.1 N, the Paralympic cyclist must perform a force on each crank about 803.6 N. Once the maximum force associated with the movement has been determined, it is continued to the dynamic modeling of the transtibial prosthesis that represents a model of 6 degrees of freedom with displacement in X - Y in relation to the angles of rotation of the hip π, knee γ and ankle λ. Subsequently, an analysis of the kinematic behavior of the prosthesis was carried out by means of SolidWorks 2017 and Matlab R2015a, which was used to model and analyze the variation of the hip angles π, knee γ and ankle of the λ prosthesis. The reaction forces generated in the prosthesis were performed on the ankle of the prosthesis, performing the summation of forces on the X and Y axes. The same analysis was then applied to the tibia of the prosthesis and the socket. The reaction force of the parts of the prosthesis varies according to the hip angles π, knee γ and ankle of the prosthesis λ. Therefore, it can be deduced that the maximum forces experienced by the ankle of the prosthesis is 933.6 N on the X axis and 2.160.5 N on the Y axis. Finally, it is calculated that the maximum forces experienced by the tibia and the socket of the transtibial prosthesis in high performance competitions is 3.266 N on the X axis and 1.357 N on the Y axis. In conclusion, it can be said that the performance of the cyclist depends on several physiological factors, linked to biomechanics of training. The influence of biomechanical factors such as aerodynamics, bicycle measurements, connecting rod length, or non-circular pedaling systems on the cyclist performance.

Keywords: biomechanics, dynamic model, paralympic cyclist, transtibial prosthesis

Procedia PDF Downloads 341
15120 Simulation of Multistage Extraction Process of Co-Ni Separation Using Ionic Liquids

Authors: Hongyan Chen, Megan Jobson, Andrew J. Masters, Maria Gonzalez-Miquel, Simon Halstead, Mayri Diaz de Rienzo

Abstract:

Ionic liquids offer excellent advantages over conventional solvents for industrial extraction of metals from aqueous solutions, where such extraction processes bring opportunities for recovery, reuse, and recycling of valuable resources and more sustainable production pathways. Recent research on the use of ionic liquids for extraction confirms their high selectivity and low volatility, but there is relatively little focus on how their properties can be best exploited in practice. This work addresses gaps in research on process modelling and simulation, to support development, design, and optimisation of these processes, focusing on the separation of the highly similar transition metals, cobalt, and nickel. The study exploits published experimental results, as well as new experimental results, relating to the separation of Co and Ni using trihexyl (tetradecyl) phosphonium chloride. This extraction agent is attractive because it is cheaper, more stable and less toxic than fluorinated hydrophobic ionic liquids. This process modelling work concerns selection and/or development of suitable models for the physical properties, distribution coefficients, for mass transfer phenomena, of the extractor unit and of the multi-stage extraction flowsheet. The distribution coefficient model for cobalt and HCl represents an anion exchange mechanism, supported by the literature and COSMO-RS calculations. Parameters of the distribution coefficient models are estimated by fitting the model to published experimental extraction equilibrium results. The mass transfer model applies Newman’s hard sphere model. Diffusion coefficients in the aqueous phase are obtained from the literature, while diffusion coefficients in the ionic liquid phase are fitted to dynamic experimental results. The mass transfer area is calculated from the surface to mean diameter of liquid droplets of the dispersed phase, estimated from the Weber number inside the extractor. New experiments measure the interfacial tension between the aqueous and ionic phases. The empirical models for predicting the density and viscosity of solutions under different metal loadings are also fitted to new experimental data. The extractor is modelled as a continuous stirred tank reactor with mass transfer between the two phases and perfect phase separation of the outlet flows. A multistage separation flowsheet simulation is set up to replicate a published experiment and compare model predictions with the experimental results. This simulation model is implemented in gPROMS software for dynamic process simulation. The results of single stage and multi-stage flowsheet simulations are shown to be in good agreement with the published experimental results. The estimated diffusion coefficient of cobalt in the ionic liquid phase is in reasonable agreement with published data for the diffusion coefficients of various metals in this ionic liquid. A sensitivity study with this simulation model demonstrates the usefulness of the models for process design. The simulation approach has potential to be extended to account for other metals, acids, and solvents for process development, design, and optimisation of extraction processes applying ionic liquids for metals separations, although a lack of experimental data is currently limiting the accuracy of models within the whole framework. Future work will focus on process development more generally and on extractive separation of rare earths using ionic liquids.

Keywords: distribution coefficient, mass transfer, COSMO-RS, flowsheet simulation, phosphonium

Procedia PDF Downloads 190
15119 A Hybrid Artificial Intelligence and Two Dimensional Depth Averaged Numerical Model for Solving Shallow Water and Exner Equations Simultaneously

Authors: S. Mehrab Amiri, Nasser Talebbeydokhti

Abstract:

Modeling sediment transport processes by means of numerical approach often poses severe challenges. In this way, a number of techniques have been suggested to solve flow and sediment equations in decoupled, semi-coupled or fully coupled forms. Furthermore, in order to capture flow discontinuities, a number of techniques, like artificial viscosity and shock fitting, have been proposed for solving these equations which are mostly required careful calibration processes. In this research, a numerical scheme for solving shallow water and Exner equations in fully coupled form is presented. First-Order Centered scheme is applied for producing required numerical fluxes and the reconstruction process is carried out toward using Monotonic Upstream Scheme for Conservation Laws to achieve a high order scheme.  In order to satisfy C-property of the scheme in presence of bed topography, Surface Gradient Method is proposed. Combining the presented scheme with fourth order Runge-Kutta algorithm for time integration yields a competent numerical scheme. In addition, to handle non-prismatic channels problems, Cartesian Cut Cell Method is employed. A trained Multi-Layer Perceptron Artificial Neural Network which is of Feed Forward Back Propagation (FFBP) type estimates sediment flow discharge in the model rather than usual empirical formulas. Hydrodynamic part of the model is tested for showing its capability in simulation of flow discontinuities, transcritical flows, wetting/drying conditions and non-prismatic channel flows. In this end, dam-break flow onto a locally non-prismatic converging-diverging channel with initially dry bed conditions is modeled. The morphodynamic part of the model is verified simulating dam break on a dry movable bed and bed level variations in an alluvial junction. The results show that the model is capable in capturing the flow discontinuities, solving wetting/drying problems even in non-prismatic channels and presenting proper results for movable bed situations. It can also be deducted that applying Artificial Neural Network, instead of common empirical formulas for estimating sediment flow discharge, leads to more accurate results.

Keywords: artificial neural network, morphodynamic model, sediment continuity equation, shallow water equations

Procedia PDF Downloads 187
15118 Collaborative Energy Optimization for Multi-Microgrid Distribution System Based on Two-Stage Game Approach

Authors: Hanmei Peng, Yiqun Wang, Mao Tan, Zhuocen Dai, Yongxin Su

Abstract:

Efficient energy management in multi-microgrid distribution systems holds significant importance for enhancing the economic benefits of regional power grids. To better balance conflicts among various stakeholders, a two-stage game-based collaborative optimization approach is proposed in this paper, effectively addressing the realistic scenario involving both competition and collaboration among stakeholders. The first stage, aimed at maximizing individual benefits, involves constructing a non-cooperative tariff game model for the distribution network and surplus microgrid. In the second stage, considering power flow and physical line capacity constraints we establish a cooperative P2P game model for the multi-microgrid distribution system, and the optimization involves employing the Lagrange method of multipliers to handle complex constraints. Simulation results demonstrate that the proposed approach can effectively improve the system economics while harmonizing individual and collective rationality.

Keywords: cooperative game, collaborative optimization, multi-microgrid distribution system, non-cooperative game

Procedia PDF Downloads 71
15117 A Data-Driven Agent Based Model for the Italian Economy

Authors: Michele Catalano, Jacopo Di Domenico, Luca Riccetti, Andrea Teglio

Abstract:

We develop a data-driven agent based model (ABM) for the Italian economy. We calibrate the model for the initial condition and parameters. As a preliminary step, we replicate the Monte-Carlo simulation for the Austrian economy. Then, we evaluate the dynamic properties of the model: the long-run equilibrium and the allocative efficiency in terms of disequilibrium patterns arising in the search and matching process for final goods, capital, intermediate goods, and credit markets. In this perspective, we use a randomized initial condition approach. We perform a robustness analysis perturbing the system for different parameter setups. We explore the empirical properties of the model using a rolling window forecast exercise from 2010 to 2022 to observe the model’s forecasting ability in the wake of the COVID-19 pandemic. We perform an analysis of the properties of the model with a different number of agents, that is, with different scales of the model compared to the real economy. The model generally displays transient dynamics that properly fit macroeconomic data regarding forecasting ability. We stress the model with a large set of shocks, namely interest policy, fiscal policy, and exogenous factors, such as external foreign demand for export. In this way, we can explore the most exposed sectors of the economy. Finally, we modify the technology mix of the various sectors and, consequently, the underlying input-output sectoral interdependence to stress the economy and observe the long-run projections. In this way, we can include in the model the generation of endogenous crisis due to the implied structural change, technological unemployment, and potential lack of aggregate demand creating the condition for cyclical endogenous crises reproduced in this artificial economy.

Keywords: agent-based models, behavioral macro, macroeconomic forecasting, micro data

Procedia PDF Downloads 69
15116 A Multi-Level Approach to Improve Sustainability Performances of Industrial Agglomerations

Authors: Patrick Innocenti, Elias Montini, Silvia Menato, Marzio Sorlini

Abstract:

Documented experiences of industrial symbiosis are always triggered and driven only by economic goals: environmental and (even rarely) social results are sometimes assessed and declared as effects of virtuous behaviours, but are merely casual and un-pursued side externalities. Even worse: all the symbiotic project candidates entailing economic loss for just one of the (also dozen) partners are simply stopped without considering the overall benefit for the whole partnership. The here-presented approach aims at providing methodologies and tools to effectively manage these situations and fostering the implementation of virtuous symbiotic investments in manufacturing aggregations for a more sustainable production.

Keywords: business model, industrial symbiosis, industrial agglomerations, sustainability

Procedia PDF Downloads 290
15115 A Heuristic for the Integrated Production and Distribution Scheduling Problem

Authors: Christian Meinecke, Bernd Scholz-Reiter

Abstract:

The integrated problem of production and distribution scheduling is relevant in many industrial applications. Thus, many heuristics to solve this integrated problem have been developed in the last decade. Most of these heuristics use a sequential working principal or a single decomposition and integration approach to separate and solve sub-problems. A heuristic using a multi-step decomposition and integration approach is presented in this paper and evaluated in a case study. The result show significant improved results compared with sequential scheduling heuristics.

Keywords: production and outbound distribution, integrated planning, heuristic, decomposition, integration

Procedia PDF Downloads 429
15114 Exploring Visual Methodologies for Measuring Public Perception of Sex Offenders

Authors: Sasha Goodwin

Abstract:

Sex offenders are often viewed as a homogenous group, but they encompass a diverse range of individuals with varying characteristics and offenses. The principal aim of this study was to ascertain how members of the Australian public perceive and define a sex offender while also investigating the emotional underpinnings associated with these attitudes and definitions. To assess public attitude, this study used the innovative utilization of visual methodologies to assess the public's perception of sex offenders. The study employed the iSquare approach, a visual methodology framework that offers unique viewpoints and insights into public attitudes toward sex offenders. Through the utilization of this approach, this study established an academic foundation for a deeper understanding of the public's perception of sex offenders. The data analysis revealed that most participants associated sex offenders with strong negative emotions, primarily disgust and anger. The findings of this research point towards the potential for fostering a social environment characterized by evidence-based discussions instead of reactionary punitive responses. Promoting a comprehensive understanding of the diverse nature of sexual offenders aims to broaden perceptions, fostering constructive attitudes.

Keywords: visual methodologies, public perception, sex offenders, offender characteristics, emotional attitudes, isquare approach, attitudes

Procedia PDF Downloads 63
15113 Triple Diffusive Convection in a Vertically Oscillating Oldroyd-B Liquid

Authors: Sameena Tarannum, S. Pranesh

Abstract:

The effect of linear stability analysis of triple diffusive convection in a vertically oscillating viscoelastic liquid of Oldroyd-B type is studied. The correction Rayleigh number is obtained by using perturbation method which gives prospect to control the convection. The eigenvalue is obtained by using perturbation method by adopting Venezian approach. From the study, it is observed that gravity modulation advances the onset of triple diffusive convection.

Keywords: gravity modulation, Oldroyd-b liquid, triple diffusive convection, venezian approach

Procedia PDF Downloads 176
15112 Non-Destructive Test of Bar for Determination of Critical Compression Force Directed towards the Pole

Authors: Boris Blostotsky, Elia Efraim

Abstract:

The phenomenon of buckling of structural elements under compression is revealed in many cases of loading and found consideration in many structures and mechanisms. In the present work the method and results of dynamic test for buckling of bar loaded by a compression force directed towards the pole are considered. Experimental determination of critical force for such system has not been made previously. The tested object is a bar with semi-rigid connection to the base at one of its ends, and with a hinge moving along a circle at the other. The test includes measuring the natural frequency of the bar at different values of compression load. The lateral stiffness is calculated based on natural frequency and reduced mass on the bar's movable end. The critical load is determined by extrapolation the values of lateral stiffness up to zero value. For the experimental investigation the special test-bed was created that allows the stability testing at positive and negative curvature of the movable end's trajectory, as well as varying the rotational stiffness of the other end connection. Decreasing a friction at the movable end allows extend the diapason of applied compression force. The testing method includes: - Methodology of the experiment planning, that allows determine the required number of tests under various loads values in the defined range and the type of extrapolating function; - Methodology of experimental determination of reduced mass at the bar's movable end including its own mass; - Methodology of experimental determination of lateral stiffness of uncompressed bar rotational semi-rigid connection at the base. For planning the experiment and for comparison of the experimental results with the theoretical values of critical load, the analytical dependencies of lateral stiffness of the bar with defined end conditions on compression load. In the particular case of perfectly rigid connection of the bar to the base, the critical load value corresponds to solution by S.P. Timoshenko. Correspondence of the calculated and experimental values was obtained.

Keywords: non-destructive test, buckling, dynamic method, semi-rigid connections

Procedia PDF Downloads 355
15111 Developing Oral Communication Competence in a Second Language: The Communicative Approach

Authors: Ikechi Gilbert

Abstract:

Oral communication is the transmission of ideas or messages through the speech process. Acquiring competence in this area which, by its volatile nature, is prone to errors and inaccuracies would require the adoption of a well-suited teaching methodology. Efficient oral communication facilitates exchange of ideas and easy accomplishment of day-to-day tasks, by means of a demonstrated mastery of oral expression and the making of fine presentations to audiences or individuals while recognizing verbal signals and body language of others and interpreting them correctly. In Anglophone states such as Nigeria, Ghana, etc., the French language, for instance, is studied as a foreign language, being used majorly in teaching learners who have their own mother tongue different from French. The same applies to Francophone states where English is studied as a foreign language by people whose official language or mother tongue is different from English. The ideal approach would be to teach these languages in these environments through a pedagogical approach that properly takes care of the oral perspective for effective understanding and application by the learners. In this article, we are examining the communicative approach as a methodology for teaching oral communication in a foreign language. This method is a direct response to the communicative needs of the learner involving the use of appropriate materials and teaching techniques that meet those needs. It is also a vivid improvement to the traditional grammatical and audio-visual adaptations. Our contribution will focus on the pedagogical component of oral communication improvement, highlighting its merits and also proposing diverse techniques including aspects of information and communication technology that would assist the second language learner communicate better orally.

Keywords: communication, competence, methodology, pedagogical component

Procedia PDF Downloads 266
15110 Linguistic Trend in the Qur'anic Tafsir of 'Al Tahreer Wa Al Tanveer' by Sheikh Tahir Bin A'shur

Authors: Numan Hasan

Abstract:

We have tried to highlight the linguistic trend in the Qur’anic Tafsir of ‘Al Tahreer wa Al Tanveer’ by Sheikh Tahir Bin A’shur, the brightest linguistic commentator in the modern era. We have started studying the life of Bin A’shur and his contributions to the field of Qur’anic knowledge. We have also studied to focus on the linguistic approach of ‘Al Tahreer wa Al Tanveer’ and emphasized the importance of linguistic interpretations. We have tried to have a clear understanding about the features and characteristics of his Tafsir. We have also reflected on the methodological approach and linguistic reference of his interpretation. In the conclusion we presented the main results of a research.

Keywords: Sheikh Tahir Bin A’shur, tafsir, linguistics, interpretation, Islamic studies

Procedia PDF Downloads 376
15109 Thermal Method for Testing Small Chemisorbent Samples on the Base of Potassium Superoxide

Authors: Pavel V. Balabanov, Daria A. Liubimova, Aleksandr P. Savenkov

Abstract:

The increase of technogenic and natural accidents, accompanied by air pollution, for example, by combustion products, leads to the necessity of respiratory protection. This work is devoted to the development of a calorimetric method and a device which allow investigating quickly the kinetics of carbon dioxide sorption by chemo-sorbents on the base of potassium superoxide in order to assess the protective properties of respiratory protective closed-circuit apparatus. The features of the traditional approach for determining the sorption properties in a thin layer of chemo-sorbent are described, as well as methods and devices, which can be used for the sorption kinetics study. The authors of the paper developed an approach (as opposed to the traditional approach) based on the power measurement of internal heat sources in the chemo-sorbent layer. The emergence of the heat sources is a result of the exothermic reaction of carbon dioxide sorption. This approach eliminates the necessity of chemical analysis of samples and can significantly reduce the time and material expenses during chemo-sorbents testing. The error of determining the volume fraction of adsorbed carbon dioxide by the developed method does not exceed 12%. Taking into account the efficiency of the method, we consider that it is a good alternative to traditional methods of chemical analysis under the assessment of the protection sorbents quality.

Keywords: carbon dioxide chemisorption, exothermic reaction, internal heat sources, respiratory protective apparatus

Procedia PDF Downloads 408
15108 Biocarbon for High-Performance Supercapacitors Derived from the Wastewater Treatment of Sewage Sludge

Authors: Santhosh Ravichandran, F. J. Rodríguez-Varela

Abstract:

In this study, a biocarbon (BC) was made from sewage sludge from the water treatment plant (PTAR) in Saltillo, Coahuila, Mexico. The sludge was carbonized in water and then chemically activated by pyrolysis. The biocarbon was evaluated physicochemically using XRD, SEM-EDS, and FESEM. A broad (002) peak attributable to graphitic structures indicates that the material is amorphous. The resultant biocarbon has a high specific surface area (412 m2 g-1), a large pore volume (0.39 cm3 g-1), interconnected hierarchical porosity, and outstanding electrochemical performance. It is appropriate for high-performance supercapacitor electrode materials due to its high specific capacitance of 358 F g-1, great rate capability, and outstanding cycling stability (around 87% capacitance retention after 10,000 cycles, even at a high current density of 19 A g-1). In an aqueous solution, the constructed BC/BC symmetric supercapacitor exhibits increased super capacitor behavior with a high energy density of 29.5 Whkg-1. The concept provides an efficient method for producing high-performance electrode materials for supercapacitors from conventional water treatment biomass wastes.

Keywords: supercapacitors, carbon, material science, batteries

Procedia PDF Downloads 84
15107 Dynamic Wind Effects in Tall Buildings: A Comparative Study of Synthetic Wind and Brazilian Wind Standard

Authors: Byl Farney Cunha Junior

Abstract:

In this work the dynamic three-dimensional analysis of a 47-story building located in Goiania city when subjected to wind loads generated using both the Wind Brazilian code, NBR6123 (ABNT, 1988) and the Synthetic-Wind method is realized. To model the frames three different methodologies are used: the shear building model and both bi and three-dimensional finite element models. To start the analysis, a plane frame is initially studied to validate the shear building model and, in order to compare the results of natural frequencies and displacements at the top of the structure the same plane frame was modeled using the finite element method through the SAP2000 V10 software. The same steps were applied to an idealized 20-story spacial frame that helps in the presentation of the stiffness correction process applied to columns. Based on these models the two methods used to generate the Wind loads are presented: a discrete model proposed in the Wind Brazilian code, NBR6123 (ABNT, 1988) and the Synthetic-Wind method. The method uses the Davenport spectrum which is divided into a variety of frequencies to generate the temporal series of loads. Finally, the 47- story building was analyzed using both the three-dimensional finite element method through the SAP2000 V10 software and the shear building model. The models were loaded with Wind load generated by the Wind code NBR6123 (ABNT, 1988) and by the Synthetic-Wind method considering different wind directions. The displacements and internal forces in columns and beams were compared and a comparative study considering a situation of a full elevated reservoir is realized. As can be observed the displacements obtained by the SAP2000 V10 model are greater when loaded with NBR6123 (ABNT, 1988) wind load related to the permanent phase of the structure’s response.

Keywords: finite element method, synthetic wind, tall buildings, shear building

Procedia PDF Downloads 273
15106 Mine Project Evaluations in the Rising of Uncertainty: Real Options Analysis

Authors: I. Inthanongsone, C. Drebenstedt, J. C. Bongaerts, P. Sontamino

Abstract:

The major concern in evaluating the value of mining projects related to the deficiency of the traditional discounted cash flow (DCF) method. This method does not take uncertainties into account and, hence it does not allow for an economic assessment of managerial flexibility and operational adaptability, which are increasingly determining long-term corporate success. Such an assessment can be performed with the real options valuation (ROV) approach, since it allows for a comparative evaluation of unforeseen uncertainties in a project life cycle. This paper presents an economic evaluation model for open pit mining projects based on real options valuation approach. Uncertainties in the model are caused by metal prices and cost uncertainties and the system dynamics (SD) modeling method is used to structure and solve the real options model. The model is applied to a case study. It can be shown that that managerial flexibility reacting to uncertainties may create additional value to a mining project in comparison to the outcomes of a DCF method. One important insight for management dealing with uncertainty is seen in choosing the optimal time to exercise strategic options.

Keywords: DCF methods, ROV approach, system dynamics modeling methods, uncertainty

Procedia PDF Downloads 501
15105 Unified Assessment of Power System Reserve-based Reliability Levels

Authors: B. M. Alshammari, M. A. El-Kady

Abstract:

This paper presents a unified framework for assessment of reserve-based reliability levels in electric power systems. The unified approach is based on reserve-based analysis and assessment of the relationship between available generation capacities and required demand levels. The developed approach takes into account the load variations as well as contingencies which occur randomly causing some generation and/or transmission capacities to be lost (become unavailable). The calculated reserve based indices, which are important to assess the reserve capabilities of the power system for various operating scenarios are therefore probabilistic in nature. They reflect the fact that neither the load levels nor the generation or transmission capacities are known with absolute certainty. They are rather subjects to random variations and consequently. The calculated reserve-based reliability indices are all subjects to random variations where only expected values of these indices can be evaluated. This paper presents a unified approach to reserve-based reliability assessment of power systems using various reserve assessment criteria. Practical applications are also presented for demonstration purposes to the Saudi electricity power grid.

Keywords: assessment, power system, reserve, reliability

Procedia PDF Downloads 617
15104 Geovisualisation for Defense Based on a Deep Learning Monocular Depth Reconstruction Approach

Authors: Daniel R. dos Santos, Mateus S. Maldonado, Estevão J. R. Batista

Abstract:

The military commanders increasingly dependent on spatial awareness, as knowing where enemy are, understanding how war battle scenarios change over time, and visualizing these trends in ways that offer insights for decision-making. Thanks to advancements in geospatial technologies and artificial intelligence algorithms, the commanders are now able to modernize military operations on a universal scale. Thus, geovisualisation has become an essential asset in the defense sector. It has become indispensable for better decisionmaking in dynamic/temporal scenarios, operation planning and management for the war field, situational awareness, effective planning, monitoring, and others. For example, a 3D visualization of war field data contributes to intelligence analysis, evaluation of postmission outcomes, and creation of predictive models to enhance decision-making and strategic planning capabilities. However, old-school visualization methods are slow, expensive, and unscalable. Despite modern technologies in generating 3D point clouds, such as LIDAR and stereo sensors, monocular depth values based on deep learning can offer a faster and more detailed view of the environment, transforming single images into visual information for valuable insights. We propose a dedicated monocular depth reconstruction approach via deep learning techniques for 3D geovisualisation of satellite images. It introduces scalability in terrain reconstruction and data visualization. First, a dataset with more than 7,000 satellite images and associated digital elevation model (DEM) is created. It is based on high resolution optical and radar imageries collected from Planet and Copernicus, on which we fuse highresolution topographic data obtained using technologies such as LiDAR and the associated geographic coordinates. Second, we developed an imagery-DEM fusion strategy that combine feature maps from two encoder-decoder networks. One network is trained with radar and optical bands, while the other is trained with DEM features to compute dense 3D depth. Finally, we constructed a benchmark with sparse depth annotations to facilitate future research. To demonstrate the proposed method's versatility, we evaluated its performance on no annotated satellite images and implemented an enclosed environment useful for Geovisualisation applications. The algorithms were developed in Python 3.0, employing open-source computing libraries, i.e., Open3D, TensorFlow, and Pythorch3D. The proposed method provides fast and accurate decision-making with GIS for localization of troops, position of the enemy, terrain and climate conditions. This analysis enhances situational consciousness, enabling commanders to fine-tune the strategies and distribute the resources proficiently.

Keywords: depth, deep learning, geovisualisation, satellite images

Procedia PDF Downloads 10
15103 Oil Reservoir Asphalting Precipitation Estimating during CO2 Injection

Authors: I. Alhajri, G. Zahedi, R. Alazmi, A. Akbari

Abstract:

In this paper, an Artificial Neural Network (ANN) was developed to predict Asphaltene Precipitation (AP) during the injection of carbon dioxide into crude oil reservoirs. In this study, the experimental data from six different oil fields were collected. Seventy percent of the data was used to develop the ANN model, and different ANN architectures were examined. A network with the Trainlm training algorithm was found to be the best network to estimate the AP. To check the validity of the proposed model, the model was used to predict the AP for the thirty percent of the data that was unevaluated. The Mean Square Error (MSE) of the prediction was 0.0018, which confirms the excellent prediction capability of the proposed model. In the second part of this study, the ANN model predictions were compared with modified Hirschberg model predictions. The ANN was found to provide more accurate estimates compared to the modified Hirschberg model. Finally, the proposed model was employed to examine the effect of different operating parameters during gas injection on the AP. It was found that the AP is mostly sensitive to the reservoir temperature. Furthermore, the carbon dioxide concentration in liquid phase increases the AP.

Keywords: artificial neural network, asphaltene, CO2 injection, Hirschberg model, oil reservoirs

Procedia PDF Downloads 365
15102 Commoning as an Approach to Community Planning: An Inquiry into the Role of Urban Local Bodies and Commoners

Authors: Pruthvi Nath Palleti, Sarmada Madhulika Kone

Abstract:

Communities are formed based on the commonalities that exist in a set of individuals and when the group comes together on identifying those commonalities and to achieve their common goals. Thus, community planning with its vision to strengthen the community mostly involves with making or remaking of commons, which results in making or remaking of communities. This paper looks into different practices of planning around the world and tried to establish a link between commoning (the act of exercising the rights over commons by commoners) and participatory approach to community planning.

Keywords: commoners, commoning, community, participatory planning, urban local bodies

Procedia PDF Downloads 373
15101 Constructing the Joint Mean-Variance Regions for Univariate and Bivariate Normal Distributions: Approach Based on the Measure of Cumulative Distribution Functions

Authors: Valerii Dashuk

Abstract:

The usage of the confidence intervals in economics and econometrics is widespread. To be able to investigate a random variable more thoroughly, joint tests are applied. One of such examples is joint mean-variance test. A new approach for testing such hypotheses and constructing confidence sets is introduced. Exploring both the value of the random variable and its deviation with the help of this technique allows checking simultaneously the shift and the probability of that shift (i.e., portfolio risks). Another application is based on the normal distribution, which is fully defined by mean and variance, therefore could be tested using the introduced approach. This method is based on the difference of probability density functions. The starting point is two sets of normal distribution parameters that should be compared (whether they may be considered as identical with given significance level). Then the absolute difference in probabilities at each 'point' of the domain of these distributions is calculated. This measure is transformed to a function of cumulative distribution functions and compared to the critical values. Critical values table was designed from the simulations. The approach was compared with the other techniques for the univariate case. It differs qualitatively and quantitatively in easiness of implementation, computation speed, accuracy of the critical region (theoretical vs. real significance level). Stable results when working with outliers and non-normal distributions, as well as scaling possibilities, are also strong sides of the method. The main advantage of this approach is the possibility to extend it to infinite-dimension case, which was not possible in the most of the previous works. At the moment expansion to 2-dimensional state is done and it allows to test jointly up to 5 parameters. Therefore the derived technique is equivalent to classic tests in standard situations but gives more efficient alternatives in nonstandard problems and on big amounts of data.

Keywords: confidence set, cumulative distribution function, hypotheses testing, normal distribution, probability density function

Procedia PDF Downloads 176
15100 Savinglife®: An Educational Technology for Basic and Advanced Cardiovascular Life Support

Authors: Naz Najma, Grace T. M. Dal Sasso, Maria de Lourdes de Souza

Abstract:

The development of information and communication technologies and the accessibility of mobile devices has increased the possibilities of the teaching and learning process anywhere and anytime. Mobile and web application allows the production of constructive teaching and learning models in various educational settings, showing the potential for active learning in nursing. The objective of this study was to present the development of an educational technology (Savinglife®, an app) for learning cardiopulmonary resuscitation and advanced cardiovascular life support training. Savinglife® is a technological production, based on the concept of virtual learning and problem-based learning approach. The study was developed from January 2016 to November 2016, using five phases (analyze, design, develop, implement, evaluate) of the instructional systems development process. The technology presented 10 scenarios and 12 simulations, covering different aspects of basic and advanced cardiac life support. The contents can be accessed in a non-linear way leaving the students free to build their knowledge based on their previous experience. Each scenario is presented through interactive tools such as scenario description, assessment, diagnose, intervention and reevaluation. Animated ECG rhythms, text documents, images and videos are provided to support procedural and active learning considering real life situation. Accessible equally on small to large devices with or without an internet connection, Savinglife® offers a dynamic, interactive and flexible tool, placing students at the center of the learning process. Savinglife® can contribute to the student’s learning in the assessment and management of basic and advanced cardiac life support in a safe and ethical way.

Keywords: problem-based learning, cardiopulmonary resuscitation, nursing education, advanced cardiac life support, educational technology

Procedia PDF Downloads 305
15099 Seismic Performance of a Framed Structure Retrofitted with Damped Cable Systems

Authors: Asad Naeem, Minsung Kim, Jinkoo Kim

Abstract:

In this work, the effectiveness of damped cable systems (DCS) on the mitigation of earthquake-induced response of a framed structure is investigated. The seismic performance of DCS is investigated using fragility analysis and life cycle cost evaluation of an existing building retrofitted with DCS, and the results are compared with those of the structure retrofitted with viscous dampers. The comparison of the analysis results reveals that, due to the self-centering capability of the DCS, residual displacement becomes nearly zero in the structure retrofitted with the DCS. According to the fragility analysis, the structure retrofitted with the DCS has smaller probability of reaching a limit states compared to the structure with viscous dampers. It is also observed that both the initial and life cycle costs of the DCS method required for the seismic retrofit is smaller than those of the structure retrofitted with viscous dampers. Acknowledgment: This research was supported by a grant (17CTAP-C132889-01) from Technology Advancement Research Program (TARP) funded by Ministry of Land, Infrastructure, and Transport of Korean government.

Keywords: damped cable system, seismic retrofit, self centering, fragility analysis

Procedia PDF Downloads 453
15098 End-to-End Control and Management of Multi-AS Virtual Service Networks Using SDN and Autonomic Computing Architecture

Authors: Yong Xue, Daniel A. Menascé

Abstract:

Automated and end-to-end network resource management and provisioning for virtual service networks in a multiple autonomous systems (a.k.a multi-AS) environment is a challenging and open problem. This paper proposes a novel, scalable and interoperable high-level architecture that incorporates a number of emerging enabling technologies including Software Defined Network (SDN), Network Function Virtualization (NFV), Service Oriented Architecture (SOA), and Autonomic Computing. The proposed architecture can be used to not only automate network resource management and provisioning for virtual service networks across multiple autonomous substrate networks, but also provide an adaptive capability for achieving optimal network resource management and maintaining network-level end-to-end network performance as well. The paper argues that this SDN and autonomic computing based architecture lays a solid foundation that can facilitate the development of the future Internet based on the pluralistic paradigm.

Keywords: virtual network, software defined network, virtual service network, adaptive resource management, SOA, multi-AS, inter-domain

Procedia PDF Downloads 531
15097 Quality of Today's Teachers: Post-Certified Teachers' Competence in Alleviating Poverties towards a Sustainable Development

Authors: Sudirman

Abstract:

Competence is a term describing capability that correlates with a person’s occupation. The competence of a teacher consists of four, i.e., pedagogical, professional, personality and social competence. These four components are implemented during interacting with students to motivate the students and improve their achievement. The objective of this qualitative study is to explore the roles and contributions of certified teachers in alleviating the issue of poverty to promote a sustainable development. The data comprise primary and secondary data which were generated from observation, interview, documentation and library research. Furthermore, this study offers in-depth information regarding the performance of the teachers in coping with poverty and sustaining development. The result shows that the teacher’s competence positively contributes to the improvement of students’ achievement. This helps the students to prepare for the real work experience by which it results in a better income and, therefore, alleviate poverty. All in all, the quality of today’s teachers can be measured by their contribution in enhancing the students’ competence prior to entering real work, resulting in a wealthy society. This is to deal with poverty and conceptualizing a sustainable development.

Keywords: competence, development, poverty, teachers

Procedia PDF Downloads 151