Search results for: technology complexity
9205 Team Cognitive Heterogeneity and Strategic Decision-Making Flexibility: The Role of Transactive Memory System and Task Complexity
Authors: Rui Xing, Baolin Ye, Nan Zhou, Guohong Wang
Abstract:
Drawing upon a perspective of cognitive interaction, this study explores the relationship between team cognitive heterogeneity and team strategic decision-making flexibility, treating the transactive memory system as a mediator and task complexity as a moderator. The hypotheses were tested in linear regression models by using data gathered from 67 strategic decision-making teams in the new-energy vehicle industry. It is found that team cognitive heterogeneity has a positive impact on strategic decision-making flexibility through the mediation of specialization and coordination of the transactive memory system, which is positively moderated by task complexity.Keywords: strategic decision-making flexibility, team cognitive heterogeneity, transactive memory system, task complexity
Procedia PDF Downloads 809204 The Role Played by Awareness and Complexity through the Use of a Logistic Regression Analysis
Authors: Yari Vecchio, Margherita Masi, Jorgelina Di Pasquale
Abstract:
Adoption of Precision Agriculture (PA) is involved in a multidimensional and complex scenario. The process of adopting innovations is complex and social inherently, influenced by other producers, change agents, social norms and organizational pressure. Complexity depends on factors that interact and influence the decision to adopt. Farm and operator characteristics, as well as organizational, informational and agro-ecological context directly affect adoption. This influence has been studied to measure drivers and to clarify 'bottlenecks' of the adoption of agricultural innovation. Making decision process involves a multistage procedure, in which individual passes from first hearing about the technology to final adoption. Awareness is the initial stage and represents the moment in which an individual learns about the existence of the technology. 'Static' concept of adoption has been overcome. Awareness is a precondition to adoption. This condition leads to not encountering some erroneous evaluations, arose from having carried out analysis on a population that is only in part aware of technologies. In support of this, the present study puts forward an empirical analysis among Italian farmers, considering awareness as a prerequisite for adoption. The purpose of the present work is to analyze both factors that affect the probability to adopt and determinants that drive an aware individual to not adopt. Data were collected through a questionnaire submitted in November 2017. A preliminary descriptive analysis has shown that high levels of adoption have been found among younger farmers, better educated, with high intensity of information, with large farm size and high labor-intensive, and whose perception of the complexity of adoption process is lower. The use of a logit model permits to appreciate the weight played by the intensity of labor and complexity perceived by the potential adopter in PA adoption process. All these findings suggest important policy implications: measures dedicated to promoting innovation will need to be more specific for each phase of this adoption process. Specifically, they should increase awareness of PA tools and foster dissemination of information to reduce the degree of perceived complexity of the adoption process. These implications are particularly important in Europe where is pre-announced the reform of Common Agricultural Policy, oriented to innovation. In this context, these implications suggest to the measures supporting innovation to consider the relationship between various organizational and structural dimensions of European agriculture and innovation approaches.Keywords: adoption, awareness, complexity, precision agriculture
Procedia PDF Downloads 1389203 Examining the Changes in Complexity, Accuracy, and Fluency in Japanese L2 Writing Over an Academic Semester
Authors: Robert Long
Abstract:
The results of a one-year study on the evolution of complexity, accuracy, and fluency (CAF) in the compositions of Japanese L2 university students throughout a semester are presented in this study. One goal was to determine if any improvement in writing abilities over this academic term had occurred, while another was to examine methods of editing. Participants had 30 minutes to write each essay with an additional 10 minutes allotted for editing. As for editing, participants were divided into two groups, one of which utilized an online grammar checker, while the other half self-edited their initial manuscripts. From the three different institutions, there was a total of 159 students. Research questions focused on determining if the CAF had evolved over the previous year, identifying potential variations in editing techniques, and describing the connections between the CAF dimensions. According to the findings, there was some improvement in accuracy (fewer errors) in all three of the measures), whereas there was a marked decline in complexity and fluency. As for the second research aim relating to the interaction among the three dimensions (CAF) and of possible increases in fluency being offset by decreases in grammatical accuracy, results showed (there is a logical high correlation with clauses and word counts, and mean length of T-unit (MLT) and (coordinate phrase of T-unit (CP/T) as well as MLT and clause per T-unit (C/T); furthermore, word counts and error/100 ratio correlated highly with error-free clause totals (EFCT). Issues of syntactical complexity had a negative correlation with EFCT, indicating that more syntactical complexity relates to decreased accuracy. Concerning a difference in error correction between those who self-edited and those who used an online grammar correction tool, results indicated that the variable of errors-free clause ratios (EFCR) had the greatest difference regarding accuracy, with fewer errors noted with writers using an online grammar checker. As for possible differences between the first and second (edited) drafts regarding CAF, results indicated there were positive changes in accuracy, the most significant change seen in complexity (CP/T and MLT), while there were relatively insignificant changes in fluency. Results also indicated significant differences among the three institutions, with Fujian University of Technology having the most fluency and accuracy. These findings suggest that to raise students' awareness of their overall writing development, teachers should support them in developing more complex syntactic structures, improving their fluency, and making more effective use of online grammar checkers.Keywords: complexity, accuracy, fluency, writing
Procedia PDF Downloads 429202 Variable Tree Structure QR Decomposition-M Algorithm (QRD-M) in Multiple Input Multiple Output-Orthogonal Frequency Division Multiplexing (MIMO-OFDM) Systems
Authors: Jae-Hyun Ro, Jong-Kwang Kim, Chang-Hee Kang, Hyoung-Kyu Song
Abstract:
In multiple input multiple output-orthogonal frequency division multiplexing (MIMO-OFDM) systems, QR decomposition-M algorithm (QRD-M) has suboptimal error performance. However, the QRD-M has still high complexity due to many calculations at each layer in tree structure. To reduce the complexity of the QRD-M, proposed QRD-M modifies existing tree structure by eliminating unnecessary candidates at almost whole layers. The method of the elimination is discarding the candidates which have accumulated squared Euclidean distances larger than calculated threshold. The simulation results show that the proposed QRD-M has same bit error rate (BER) performance with lower complexity than the conventional QRD-M.Keywords: complexity, MIMO-OFDM, QRD-M, squared Euclidean distance
Procedia PDF Downloads 3359201 Taleb's Complexity Theory Concept of 'Antifragility' Has a Significant Contribution to Make to Positive Psychology as Applied to Wellbeing
Authors: Claudius Peter Van Wyk
Abstract:
Given the increasingly manifest phenomena, as described in complexity theory, of volatility, uncertainty, complexity and ambiguity (VUCA), Taleb's notion of 'antifragility, has a significant contribution to make to positive psychology applied to wellbeing. Antifragility is argued to be fundamentally different from the concepts of resiliency; as the ability to recover from failure, and robustness; as the ability to resist failure. Rather it describes the capacity to reorganise in the face of stress in such a way as to cope more effectively with systemic challenges. The concept, which has been applied in disciplines ranging from physics, molecular biology, planning, engineering, and computer science, can now be considered for its application in individual human and social wellbeing. There are strong correlations to Antonovsky's model of 'salutogenesis' in which an attitude and competencies are developed of transforming burdening factors into greater resourcefulness. We demonstrate, from the perspective of neuroscience, how technology measuring nervous system coherence can be coupled to acquired psychodynamic approaches to not only identify contextual stressors, utilise biofeedback instruments for facilitating greater coherence, but apply these insights to specific life stressors that compromise well-being. Employing an on-going case study with BMW South Africa, the neurological mapping is demonstrated together with 'reframing' and emotional anchoring techniques from neurolinguistic programming. The argument is contextualised in the discipline of psychoneuroimmunology which describes the stress pathways from the CNS and endocrine systems and their impact on immune function and the capacity to restore homeostasis.Keywords: antifragility, complexity, neuroscience, psychoneuroimmunology, salutogenesis, volatility
Procedia PDF Downloads 3779200 The Interplay between Technology and Culture in Inbound Call Center Industry
Authors: Joseph Reylan Viray, Kriztine R. Viray
Abstract:
Call center conversations, more than the business dimensions that they normally manifest, are interactions between human beings. These are communication exchanges that are packed with psychological, cultural and social dimensions that affect the specific experience of the parties. The increasing development of information and communication technology over the past decades brought about important advantages and corresponding disadvantages in the process of communicational transactions in call center industry. It has been established that the technology is so powerful that it strongly affects, among others, call center business. In the present study, the author explores the interplay between the technology being utilized by the industry and the cultural orientations of both the call center agents and their customers in the process of communication exchanges. Specifically, the paper seeks to (1) describe the interplay between culture and technology in inbound call center industry as it affects the communication exchange of the agents and customers; (2) understand the nature and the dynamics of the call center industry as regards the cultural dimensions of Hofstede; and (3) come up with a simple study where the cross-cultural aspect of the call center industry could be highlighted and could provide necessary knowledge to the stakeholders. Cognizant of the complexity of the topic, the researchers employed Hofstede's cultural dimensions. Likewise, another theory that was used in this study is the Computer Mediated Communication Theory.Keywords: call center industry, culture, Hofstede, CMT, technology
Procedia PDF Downloads 3549199 Risk-Based Computer Auditing and Measures of Prevention
Authors: Mohammad Hadi Khorashadi Zadeh, Amin Karkon, Seyd Mohammad Reza Mashhoori
Abstract:
the technology of Computer audit played a major role in the progress and prospects of a proper application to improve the quality and efficiency of audit work. But due to the technical complexity and the specific risks of computer audit, it should be shown effective in audit and preventive action. Mainly through research in this paper, we proposes the causes of audit risk in a computer environment and the risk of further proposals for measures to control, to some extent reduce the risk of computer audit and improve the audit quality.Keywords: computer auditing, risk, measures to prevent, information technology
Procedia PDF Downloads 4909198 Leadership's Controlling via Complexity Investigation in Crisis Scenarios
Authors: Jiří Barta, Oldřich Svoboda, Jiří F. Urbánek
Abstract:
In this paper will be discussed two coin´s sides of crisis scenarios dynamics. On the one's side is negative role of subsidiary scenario branches in its compactness weakening by means unduly chaotic atomizing, having many interactive feedbacks cases, increasing a value of a complexity here. This negative role reflects the complexity of use cases, weakening leader compliancy, which brings something as a ´readiness for controlling capabilities provision´. Leader´s dissatisfaction has zero compliancy, but factual it is a ´crossbar´ (interface in fact) between planning and executing use cases. On the other side of this coin, an advantage of rich scenarios embranchment is possible to see in a support of response awareness, readiness, preparedness, adaptability, creativity and flexibility. Here rich scenarios embranchment contributes to the steadiness and resistance of scenario mission actors. These all will be presented in live power-points ´Blazons´, modelled via DYVELOP (Dynamic Vector Logistics of Processes) on the Conference.Keywords: leadership, controlling, complexity, DYVELOP, scenarios
Procedia PDF Downloads 4049197 Handling Complexity of a Complex System Design: Paradigm, Formalism and Transformations
Authors: Hycham Aboutaleb, Bruno Monsuez
Abstract:
Current systems' complexity has reached a degree that requires addressing conception and design issues while taking into account environmental, operational, social, legal, and financial aspects. Therefore, one of the main challenges is the way complex systems are specified and designed. The exponentially growing effort, cost, and time investment of complex systems in modeling phase emphasize the need for a paradigm, a framework, and an environment to handle the system model complexity. For that, it is necessary to understand the expectations of the human user of the model and his limits. This paper presents a generic framework for designing complex systems, highlights the requirements a system model needs to fulfill to meet human user expectations, and suggests a graph-based formalism for modeling complex systems. Finally, a set of transformations are defined to handle the model complexity.Keywords: higraph-based, formalism, system engineering paradigm, modeling requirements, graph-based transformations
Procedia PDF Downloads 4039196 Effects of Topic Familiarity on Linguistic Aspects in EFL Learners’ Writing Performance
Authors: Jeong-Won Lee, Kyeong-Ok Yoon
Abstract:
The current study aimed to investigate the effects of topic familiarity and language proficiency on linguistic aspects (lexical complexity, syntactic complexity, accuracy, and fluency) in EFL learners’ argumentative essays. For the study 64 college students were asked to write an argumentative essay for the two different topics (Driving and Smoking) chosen by the consideration of topic familiarity. The students were divided into two language proficiency groups (high-level and intermediate) according to their English writing proficiency. The findings of the study are as follows: 1) the participants of this study exhibited lower levels of lexical and syntactic complexity as well as accuracy when performing writing tasks with unfamiliar topics; and 2) they demonstrated the use of a wider range of vocabulary, and longer and more complex structures, and produced accurate and lengthier texts compared to their intermediate peers. Discussion and pedagogical implications for instruction of writing classes in EFL contexts were addressed.Keywords: topic familiarity, complexity, accuracy, fluency
Procedia PDF Downloads 509195 A Fast Convergence Subband BSS Structure
Authors: Salah Al-Din I. Badran, Samad Ahmadi, Ismail Shahin
Abstract:
A blind source separation method is proposed; in this method we use a non-uniform filter bank and a novel normalisation. This method provides a reduced computational complexity and increased convergence speed comparing to the full-band algorithm. Recently, adaptive sub-band scheme has been recommended to solve two problems: reduction of computational complexity and increase the convergence speed of the adaptive algorithm for correlated input signals. In this work the reduction in computational complexity is achieved with the use of adaptive filters of orders less than the full-band adaptive filters, which operate at a sampling rate lower than the sampling rate of the input signal. The decomposed signals by analysis bank filter are less correlated in each sub-band than the input signal at full bandwidth, and can promote better rates of convergence.Keywords: blind source separation, computational complexity, subband, convergence speed, mixture
Procedia PDF Downloads 5559194 Studying Relationship between Local Geometry of Decision Boundary with Network Complexity for Robustness Analysis with Adversarial Perturbations
Authors: Tushar K. Routh
Abstract:
If inputs are engineered in certain manners, they can influence deep neural networks’ (DNN) performances by facilitating misclassifications, a phenomenon well-known as adversarial attacks that question networks’ vulnerability. Recent studies have unfolded the relationship between vulnerability of such networks with their complexity. In this paper, the distinctive influence of additional convolutional layers at the decision boundaries of several DNN architectures was investigated. Here, to engineer inputs from widely known image datasets like MNIST, Fashion MNIST, and Cifar 10, we have exercised One Step Spectral Attack (OSSA) and Fast Gradient Method (FGM) techniques. The aftermaths of adding layers to the robustness of the architectures have been analyzed. For reasoning, separation width from linear class partitions and local geometry (curvature) near the decision boundary have been examined. The result reveals that model complexity has significant roles in adjusting relative distances from margins, as well as the local features of decision boundaries, which impact robustness.Keywords: DNN robustness, decision boundary, local curvature, network complexity
Procedia PDF Downloads 769193 Uncovering the Complex Structure of Building Design Process Based on Royal Institute of British Architects Plan of Work
Authors: Fawaz A. Binsarra, Halim Boussabaine
Abstract:
The notion of complexity science has been attracting the interest of researchers and professionals due to the need of enhancing the efficiency of understanding complex systems dynamic and structure of interactions. In addition, complexity analysis has been used as an approach to investigate complex systems that contains a large number of components interacts with each other to accomplish specific outcomes and emerges specific behavior. The design process is considered as a complex action that involves large number interacted components, which are ranked as design tasks, design team, and the components of the design process. Those three main aspects of the building design process consist of several components that interact with each other as a dynamic system with complex information flow. In this paper, the goal is to uncover the complex structure of information interactions in building design process. The Investigating of Royal Institute of British Architects Plan Of Work 2013 information interactions as a case study to uncover the structure and building design process complexity using network analysis software to model the information interaction will significantly enhance the efficiency of the building design process outcomes.Keywords: complexity, process, building desgin, Riba, design complexity, network, network analysis
Procedia PDF Downloads 5299192 Implementation of Iterative Algorithm for Earthquake Location
Authors: Hussain K. Chaiel
Abstract:
The development in the field of the digital signal processing (DSP) and the microelectronics technology reduces the complexity of the iterative algorithms that need large number of arithmetic operations. Virtex-Field Programmable Gate Arrays (FPGAs) are programmable silicon foundations which offer an important solution for addressing the needs of high performance DSP designer. In this work, Virtex-7 FPGA technology is used to implement an iterative algorithm to estimate the earthquake location. Simulation results show that an implementation based on block RAMB36E1 and DSP48E1 slices of Virtex-7 type reduces the number of cycles of the clock frequency. This enables the algorithm to be used for earthquake prediction.Keywords: DSP, earthquake, FPGA, iterative algorithm
Procedia PDF Downloads 3899191 Relevance of Technology on Education
Authors: Felicia K. Oluwalola
Abstract:
This paper examines the relevance of technology on education. It identified the concept of technology on education, bringing real-world learning to the classroom situation, examples of where technology can be used. This study established the fact that technology facilitates students learning compared with traditional method of teaching. It was recommended that the teachers should use technology to supplement, not replace, other instructional modes. It should be used in conjunction with hands-on labs and activities that also address the concepts targeted by the technology. Also, technology should be students centered and not teachers centered.Keywords: computer, simulation, classroom teaching, education
Procedia PDF Downloads 4519190 Approaches to Ethical Hacking: A Conceptual Framework for Research
Authors: Lauren Provost
Abstract:
The digital world remains increasingly vulnerable, making the development of effective cybersecurity approaches even more critical in supporting the success of the digital economy and national security. Although approaches to cybersecurity have shifted and improved in the last decade with new models, especially with cloud computing and mobility, a record number of high severity vulnerabilities were recorded in the National Institute of Standards and Technology (NIST), and its National Vulnerability Database (NVD) in 2020. This is due, in part, to the increasing complexity of cyber ecosystems. Security must be approached with a more comprehensive, multi-tool strategy that addresses the complexity of cyber ecosystems, including the human factor. Ethical hacking has emerged as such an approach: a more effective, multi-strategy, comprehensive approach to cyber security's most pressing needs, especially understanding the human factor. Research on ethical hacking, however, is limited in scope. The two main objectives of this work are to (1) provide highlights of case studies in ethical hacking, (2) provide a conceptual framework for research in ethical hacking that embraces and addresses both technical and nontechnical security measures. Recommendations include an improved conceptual framework for research centered on ethical hacking that addresses many factors and attributes of significant attacks that threaten computer security; a more robust, integrative multi-layered framework embracing the complexity of cybersecurity ecosystems.Keywords: ethical hacking, literature review, penetration testing, social engineering
Procedia PDF Downloads 2209189 Determination of Complexity Level in Merged Irregular Transposition Cipher
Authors: Okike Benjamin, Garba Ejd
Abstract:
Today, it has been observed security of information along the superhighway is often compromised by those who are not authorized to have access to such information. In order to ensure the security of information along the superhighway, such information should be encrypted by some means to conceal the real meaning of the information. There are many encryption techniques out there in the market. However, some of these encryption techniques are often easily decrypted by adversaries. The researcher has decided to develop an encryption technique that may be more difficult to decrypt. This may be achieved by splitting the message to be encrypted into parts and encrypting each part separately and swapping the positions before transmitting the message along the superhighway. The method is termed Merged Irregular Transposition Cipher. Also, the research would determine the complexity level in respect to the number of splits of the message.Keywords: transposition cipher, merged irregular cipher, encryption, complexity level
Procedia PDF Downloads 3459188 Low-Complexity Multiplication Using Complement and Signed-Digit Recoding Methods
Authors: Te-Jen Chang, I-Hui Pan, Ping-Sheng Huang, Shan-Jen Cheng
Abstract:
In this paper, a fast multiplication computing method utilizing the complement representation method and canonical recoding technique is proposed. By performing complements and canonical recoding technique, the number of partial products can be reduced. Based on these techniques, we propose an algorithm that provides an efficient multiplication method. On average, our proposed algorithm is to reduce the number of k-bit additions from (0.25k+logk/k+2.5) to (k/6 +logk/k+2.5), where k is the bit-length of the multiplicand A and multiplier B. We can therefore efficiently speed up the overall performance of the multiplication. Moreover, if we use the new proposes to compute common-multiplicand multiplication, the computational complexity can be reduced from (0.5 k+2 logk/k+5) to (k/3+2 logk/k+5) k-bit additions.Keywords: algorithm design, complexity analysis, canonical recoding, public key cryptography, common-multiplicand multiplication
Procedia PDF Downloads 4379187 Automated, Short Cycle Production of Polymer Composite Applications with Special Regards to the Complexity and Recyclability of Composite Elements
Authors: Peter Pomlenyi, Orsolya Semperger, Gergely Hegedus
Abstract:
The purpose of the project is to develop a complex composite component with visible class ‘A’ surface. It is going to integrate more functions, including continuous fiber reinforcement, foam core, injection molded ribs, and metal inserts. Therefore we are going to produce recyclable structural composite part from thermoplastic polymer in serial production with short cycle time for automotive applications. Our design of the process line is determined by the principles of Industry 4.0. Accordingly, our goal is to map in details the properties of the final product including the mechanical properties in order to replace metal elements used in automotive industry, with special regard to the effect of each manufacturing process step on the afore mentioned properties. Period of the project is 3 years, which lasts from the 1st of December 2016 to the 30th November 2019. There are four consortium members in the R&D project evopro systems engineering Ltd., Department of Polymer Engineering of the Budapest University of Technology and Economics, Research Centre for Natural Sciences of Hungarian Academy of Sciences and eCon Engineering Ltd. One of the most important result that we can obtain short cycle time (up to 2-3 min) with in-situ polymerization method, which is an innovation in the field of thermoplastic composite production. Because of the mentioned method, our fully automated production line is able to manufacture complex thermoplastic composite parts and satisfies the short cycle time required by the automotive industry. In addition to the innovative technology, we are able to design, analyze complex composite parts with finite element method, and validate our results. We are continuously collecting all the information, knowledge and experience to improve our technology and obtain even more accurate results with respect to the quality and complexity of the composite parts, the cycle time of the production, and the design and analyzing method of the composite parts.Keywords: T-RTM technology, composite, automotive, class A surface
Procedia PDF Downloads 1399186 Modal Density Influence on Modal Complexity Quantification in Dynamic Systems
Authors: Fabrizio Iezzi, Claudio Valente
Abstract:
The viscous damping in dynamic systems can be proportional or non-proportional. In the first case, the mode shapes are real whereas in the second case they are complex. From an engineering point of view, the complexity of the mode shapes is important in order to quantify the non-proportional damping. Different indices exist to provide estimates of the modal complexity. These indices are or not zero, depending whether the mode shapes are not or are complex. The modal density problem arises in the experimental identification when the dynamic systems have close modal frequencies. Depending on the entity of this closeness, the mode shapes can hold fictitious imaginary quantities that affect the values of the modal complexity indices. The results are the failing in the identification of the real or complex mode shapes and then of the proportional or non-proportional damping. The paper aims to show the influence of the modal density on the values of these indices in case of both proportional and non-proportional damping. Theoretical and pseudo-experimental solutions are compared to analyze the problem according to an appropriate mechanical system.Keywords: complex mode shapes, dynamic systems identification, modal density, non-proportional damping
Procedia PDF Downloads 3899185 Green Energy, Fiscal Incentives and Conflicting Signals: Analysing the Challenges Faced in Promoting on Farm Waste to Energy Projects
Authors: Hafez Abdo, Rob Ackrill
Abstract:
Renewable energy (RE) promotion in the UK relies on multiple policy instruments, which are required to overcome the path dependency pressures favouring fossil fuels. These instruments include targeted funding schemes and economy-wide instruments embedded in the tax code. The resulting complexity of incentives raises important questions around the coherence and effectiveness of these instruments for RE generation. This complexity is exacerbated by UK RE policy being nested within EU policy in a multi-level governance (MLG) setting. To gain analytical traction on such complexity, this study will analyse policies promoting the on-farm generation of energy for heat and power, from farm and food waste, via anaerobic digestion. Utilising both primary and secondary data, it seeks to address a particular lacuna in the academic literature. Via a localised, in-depth investigation into the complexity of policy instruments promoting RE, this study will help our theoretical understanding of the challenges that MLG and path dependency pressures present to policymakers of multi-dimensional policies.Keywords: anaerobic digestion, energy, green, policy, renewable, tax, UK
Procedia PDF Downloads 3719184 Determination of Complexity Level in Okike's Merged Irregular Transposition Cipher
Authors: Okike Benjami, Garba Ejd
Abstract:
Today, it has been observed security of information along the superhighway is often compromised by those who are not authorized to have access to such information. In other to ensure the security of information along the superhighway, such information should be encrypted by some means to conceal the real meaning of the information. There are many encryption techniques out there in the market. However, some of these encryption techniques are often decrypted by adversaries with ease. The researcher has decided to develop an encryption technique that may be more difficult to decrypt. This may be achieved by splitting the message to be encrypted into parts and encrypting each part separately and swapping the positions before transmitting the message along the superhighway. The method is termed Okike’s Merged Irregular Transposition Cipher. Also, the research would determine the complexity level in respect to the number of splits of the message.Keywords: transposition cipher, merged irregular cipher, encryption, complexity level
Procedia PDF Downloads 2909183 Efficient Signal Detection Using QRD-M Based on Channel Condition in MIMO-OFDM System
Authors: Jae-Jeong Kim, Ki-Ro Kim, Hyoung-Kyu Song
Abstract:
In this paper, we propose an efficient signal detector that switches M parameter of QRD-M detection scheme is proposed for MIMO-OFDM system. The proposed detection scheme calculates the threshold by 1-norm condition number and then switches M parameter of QRD-M detection scheme according to channel information. If channel condition is bad, the parameter M is set to high value to increase the accuracy of detection. If channel condition is good, the parameter M is set to low value to reduce complexity of detection. Therefore, the proposed detection scheme has better trade off between BER performance and complexity than the conventional detection scheme. The simulation result shows that the complexity of proposed detection scheme is lower than QRD-M detection scheme with similar BER performance.Keywords: MIMO-OFDM, QRD-M, channel condition, BER
Procedia PDF Downloads 3719182 Parents' Perception on the Use of Mobile Technology by Young Children
Authors: Jason Gan
Abstract:
Technology has been around for many years, those that play a crucial part of our lives have evolved quick and fast, from televisions to computers and now mobile technology has also become a part of our lives. However in Singapore a nation with a strong reliance and following in technology, how are young children (0 to 8 years old) coping with these high technology equipment especially mobile technology and why and what are parents doing to manage this trend in the nation. This study aims to uncover some of the parents’ perception behind the use of mobile technology by their children.Keywords: technology, preschoolers, ICT and Singapore, early childhood
Procedia PDF Downloads 4359181 A Research Analysis on the Source Technology and Convergence Types
Authors: Kwounghee Choi
Abstract:
Technological convergence between the various sectors is expected to have a very large impact on future industrial and economy. This study attempts to do empirical approach between specific technologies’ classification. For technological convergence classification, it is necessary to set the target technology to be analyzed. This study selected target technology from national research and development plan. At first we found a source technology for analysis. Depending on the weight of source technology, NT-based, BT-based, IT-based, ET-based, CS-based convergence types were classified. This study aims to empirically show the concept of convergence technology and convergence types. If we use the source technology to classify convergence type, it will be useful to make practical strategies of convergence technology.Keywords: technology convergence, source technology, convergence type, R&D strategy, technology classification
Procedia PDF Downloads 4879180 Investigating the Relationship between Bank and Cloud Provider
Authors: Hatim Elhag
Abstract:
Banking and Financial Service Institutions are possibly the most advanced in terms of technology adoption and use it as a key differentiator. With high levels of business process automation, maturity in the functional portfolio, straight through processing and proven technology outsourcing benefits, Banking sector stand to benefit significantly from Cloud computing capabilities. Additionally, with complex Compliance and Regulatory policies, combined with expansive products and geography coverage, the business impact is even greater. While the benefits are exponential, there are also significant challenges in adopting this model– including Legal, Security, Performance, Reliability, Transformation complexity, Operating control and Governance and most importantly proof for the promised cost benefits. However, new architecture designed should be implemented to align this approach.Keywords: security, cloud, banking sector, cloud computing
Procedia PDF Downloads 5059179 Effect of Phonological Complexity in Children with Specific Language Impairment
Authors: Irfana M., Priyandi Kabasi
Abstract:
Children with specific language impairment (SLI) have difficulty acquiring and using language despite having all the requirements of cognitive skills to support language acquisition. These children have normal non-verbal intelligence, hearing, and oral-motor skills, with no history of social/emotional problems or significant neurological impairment. Nevertheless, their language acquisition lags behind their peers. Phonological complexity can be considered to be the major factor that causes the inaccurate production of speech in this population. However, the implementation of various ranges of complex phonological stimuli in the treatment session of SLI should be followed for a better prognosis of speech accuracy. Hence there is a need to study the levels of phonological complexity. The present study consisted of 7 individuals who were diagnosed with SLI and 10 developmentally normal children. All of them were Hindi speakers with both genders and their age ranged from 4 to 5 years. There were 4 sets of stimuli; among them were minimal contrast vs maximal contrast nonwords, minimal coarticulation vs maximal coarticulation nonwords, minimal contrast vs maximal contrast words and minimal coarticulation vs maximal coarticulation words. Each set contained 10 stimuli and participants were asked to repeat each stimulus. Results showed that production of maximal contrast was significantly accurate, followed by minimal coarticulation, minimal contrast and maximal coarticulation. A similar trend was shown for both word and non-word categories of stimuli. The phonological complexity effect was evident in the study for each participant group. Moreover, present study findings can be implemented for the management of SLI, specifically for the selection of stimuli.Keywords: coarticulation, minimal contrast, phonological complexity, specific language impairment
Procedia PDF Downloads 1449178 Using Wearable Technology to Monitor Workers’ Stress for Construction Safety: A Conceptual Framework
Authors: Namhun Lee, Seong Jin Kim
Abstract:
The construction industry represents one of the largest industries in the United States, yet it continues to face several occupational health and safety challenges. Many workers on construction sites are suffering from extended exposure to stressful situations such as poor and hazardous work environments and task complexity. Stress can be commonly defined as a feeling of emotional or physical tension, which can easily impact construction safety and result in a higher rate of job-related injuries in the construction industry. Physiological signals transmitted from wearable biosensors can be used to detect excessive stress. Therefore, workers’ stress should be detected and mitigated to prevent any type of serious incident or accident proactively. By doing this, construction productivity, as well as job satisfaction, would also be improved in the construction industry. To establish a foundation in this field of research, a conceptual framework for using wearable technology for construction safety has been developed for continuous and automatic monitoring of worker’s stress. The conceptual framework will serve as a foothold in future studies on the application of wearable technology for construction safety.Keywords: construction safety, occupational stress, stress monitoring, wearable biosensors
Procedia PDF Downloads 1619177 Theoretical Paradigms for Total Quality Environmental Management (TQEM)
Authors: Mohammad Hossein Khasmafkan Nezam, Nader Chavoshi Boroujeni, Mohamad Reza Veshaghi
Abstract:
Quality management is dominated by rational paradigms for the measurement and management of quality, but these paradigms start to ‘break down’, when faced with the inherent complexity of managing quality in intensely competitive changing environments. In this article, the various theoretical paradigms employed to manage quality are reviewed and the advantages and limitations of these paradigms are highlighted. A major implication of this review is that when faced with complexity, an ideological stance to any single strategy paradigm for total quality environmental management is ineffective. We suggest that as complexity increases and we envisage intensely competitive changing environments there will be a greater need to consider a multi-paradigm integrationist view of strategy for TQEM.Keywords: total quality management (TQM), total quality environmental management (TQEM), ideologies (philosophy), theoretical paradigms
Procedia PDF Downloads 3219176 Factors Influencing University Student's Acceptance of New Technology
Authors: Fatma Khadra
Abstract:
The objective of this research is to identify the acceptance of new technology in a sample of 150 Participants from Qatar University. Based on the Technology Acceptance Model (TAM), we used the Davis’s scale (1989) which contains two item scales for Perceived Usefulness and Perceived Ease of Use. The TAM represents an important theoretical contribution toward understanding how users come to accept and use technology. This model suggests that when people are presented with a new technology, a number of variables influence their decision about how and when they will use it. The results showed that participants accept more technology because flexibility, clarity, enhancing the experience, enjoying, facility, and useful. Also, results showed that younger participants accept more technology than others.Keywords: new technology, perceived usefulness, perceived ease of use, technology acceptance model
Procedia PDF Downloads 322