Search results for: forward error correction
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3367

Search results for: forward error correction

607 Time Estimation of Return to Sports Based on Classification of Health Levels of Anterior Cruciate Ligament Using a Convolutional Neural Network after Reconstruction Surgery

Authors: Zeinab Jafari A., Ali Sharifnezhad B., Mohammad Razi C., Mohammad Haghpanahi D., Arash Maghsoudi

Abstract:

Background and Objective: Sports-related rupture of the anterior cruciate ligament (ACL) and following injuries have been associated with various disorders, such as long-lasting changes in muscle activation patterns in athletes, which might last after ACL reconstruction (ACLR). The rupture of the ACL might result in abnormal patterns of movement execution, extending the treatment period and delaying athletes’ return to sports (RTS). As ACL injury is especially prevalent among athletes, the lengthy treatment process and athletes’ absence from sports are of great concern to athletes and coaches. Thus, estimating safe time of RTS is of crucial importance. Therefore, using a deep neural network (DNN) to classify the health levels of ACL in injured athletes, this study aimed to estimate the safe time for athletes to return to competitions. Methods: Ten athletes with ACLR and fourteen healthy controls participated in this study. Three health levels of ACL were defined: healthy, six-month post-ACLR surgery and nine-month post-ACLR surgery. Athletes with ACLR were tested six and nine months after the ACLR surgery. During the course of this study, surface electromyography (sEMG) signals were recorded from five knee muscles, namely Rectus Femoris (RF), Vastus Lateralis (VL), Vastus Medialis (VM), Biceps Femoris (BF), Semitendinosus (ST), during single-leg drop landing (SLDL) and forward hopping (SLFH) tasks. The Pseudo-Wigner-Ville distribution (PWVD) was used to produce three-dimensional (3-D) images of the energy distribution patterns of sEMG signals. Then, these 3-D images were converted to two-dimensional (2-D) images implementing the heat mapping technique, which were then fed to a deep convolutional neural network (DCNN). Results: In this study, we estimated the safe time of RTS by designing a DCNN classifier with an accuracy of 90 %, which could classify ACL into three health levels. Discussion: The findings of this study demonstrate the potential of the DCNN classification technique using sEMG signals in estimating RTS time, which will assist in evaluating the recovery process of ACLR in athletes.

Keywords: anterior cruciate ligament reconstruction, return to sports, surface electromyography, deep convolutional neural network

Procedia PDF Downloads 74
606 Risk Mapping of Road Traffic Incidents in Greater Kampala Metropolitan Area for Planning of Emergency Medical Services

Authors: Joseph Kimuli Balikuddembe

Abstract:

Road traffic incidents (RTIs) continue to be a serious public health and development burden around the globe. Compared to high-income countries (HICs), the low and middle-income countries (LMICs) bear the heaviest brunt of RTIs. Like other LMICs, Uganda, a country located in Eastern Africa, has been experiencing a worryingly high burden of RTIs and their associated impacts. Over the years, the highest number of all the total registered RTIs in Uganda has taken place in the Greater Kampala Metropolitan Area (GKMA). This places a tremendous demand on the few existing emergency medical services (EMS) to adequately respond to those affected. In this regard, the overall objective of the study was to risk map RTIs in the GKMA so as to help in the better planning of EMS for the victims of RTIs. Other objectives included: (i) identifying the factors affecting the exposure, vulnerability and EMS capacity for the victims of RTIs; (ii) identifying the RTI prone-areas and estimating their associated risk factors; (iii) identifying the weaknesses and capacities which affect the EMS systems for RTIs; and (iv) determining the strategies and priority actions that can help to improve the EMS response for RTI victims in the GKMA. To achieve these objectives, a mixed methodological approach was used in four phrases for approximately 15 months. It employed a systematic review based on the preferred reporting items for systematic reviews and meta-data analysis guidelines; a Delphi panel technique; retrospective data analysis; and a cross-sectional method. With Uganda progressing forward as envisaged in its 'Vision 2040', the GKMA, which is the country’s political and socioeconomic epicenter, is experiencing significant changes in terms of population growth, urbanization, infrastructure development, rapid motorization and other factors. Unless appropriate actions are taken, these changes are likely to worsen the already alarming rate of RTIs in Uganda, and in turn also to put pressure on the few existing EMS and facilities to render care for those affected. Therefore, road safety vis-à-vis injury prevention measures, which are needed to reduce the burden of RTIs, should be multifaceted in nature so that they closely correlate with the ongoing dynamics that contribute to RTIs, particularly in the GKMA and Uganda as a whole.

Keywords: emergency medical services, Kampala, risk mapping, road traffic incidents

Procedia PDF Downloads 114
605 In vivo Mechanical Characterization of Facial Skin Combining Digital Image Correlation and Finite Element

Authors: Huixin Wei, Shibin Wang, Linan Li, Lei Zhou, Xinhao Tu

Abstract:

Facial skin is a biomedical material with complex mechanical properties of anisotropy, viscoelasticity, and hyperelasticity. The mechanical properties of facial skin are crucial for a number of applications including facial plastic surgery, animation, dermatology, cosmetic industry, and impact biomechanics. Skin is a complex multi-layered material which can be broadly divided into three main layers, the epidermis, the dermis, and the hypodermis. Collagen fibers account for 75% of the dry weight of dermal tissue, and it is these fibers which are responsible for the mechanical properties of skin. Many research on the anisotropic mechanical properties are mainly concentrated on in vitro, but there is a great difference between in vivo and in vitro for mechanical properties of the skin. In this study, we presented a method to measure the mechanical properties of facial skin in vivo. Digital image correlation (DIC) and indentation tests were used to obtain the experiment data, including the deformation of facial surface and indentation force-displacement curve. Then, the experiment was simulated using a finite element (FE) model. Application of Computed Tomography (CT) and reconstruction techniques obtained the real tissue geometry. A three-dimensional FE model of facial skin, including a bi-layer system, was obtained. As the epidermis is relatively thin, the epidermis and dermis were regarded as one layer and below it was hypodermis in this study. The upper layer was modeled as a Gasser-Ogden-Holzapfel (GOH) model to describe hyperelastic and anisotropic behaviors of the dermis. The under layer was modeled as a linear elastic model. In conclusion, the material properties of two-layer were determined by minimizing the error between the FE data and experimental data.

Keywords: facial skin, indentation test, finite element, digital image correlation, computed tomography

Procedia PDF Downloads 107
604 The Effect of Low Power Laser on CK and Some of Markers Delayed Onset Muscle Soreness (DOMS)

Authors: Bahareh Yazdanparast Chaharmahali

Abstract:

The study showed effect of low power laser therapy on knee range of motion (flexion and extension), resting angle of knee joint, knee circumference and rating of delayed onset muscle soreness induced pain, 24 and 48 hours after eccentric training of knee flexor muscle (hamstring muscle). We investigate the effects of pulsed ultrasound on swelling, relaxed, flexion and extension knee angle and pain. 20 volunteers among girl students of college voluntary participated in this research. After eccentric training, subjects were randomly divided into two groups, control and laser therapy. In day 1 and in order to induce delayed onset muscle soreness, subjects eccentrically trained their knee flexor muscles. In day 2, subjects were randomly divided into two groups: control and low power laser therapy. 24 and 48 hours after eccentric training. Variables (knee flexion and extension, srang of motion, resting knee joint angle and knee circumferences) were measured and analyzed. Data are reported as means ± standard error (SE) and repeated measured was used to assess differences within groups. Methods of treatment (low power laser therapy) have significant effects on delayed onset muscle soreness markers. 24 and 48 hours after training a significant difference was observed between mean pains of 2 groups. This difference was significant between low power laser therapy and C groups. The Bonferroni post hock is significant. Low power laser therapy trophy as used in this study did significantly diminish the effects of delayed – onset muscle soreness on swelling, relaxed – knee extension and flexion angle.

Keywords: creatine kinase, DOMS, eccentric training, low power laser

Procedia PDF Downloads 241
603 Evaluation of Elements Impurities in Drugs According to Pharmacopoeia by use FESEM-EDS Technique

Authors: Rafid Doulab

Abstract:

Elemental Impurities in the Pharmaceuticals industryis are indispensable to ensure pharmaceuticalssafety for 24 elements. Although atomic absorption and inductively coupled plasma are used in the U.S Pharmacopeia and the European Pharmacopoeia, FESEM with energy dispersive spectrometers can be applied as an alternative analysis method for quantitative and qualitative results for a variety of elements without chemical pretreatment, unlike other techniques. This technique characterizes by shortest time, with more less contamination, no reagent consumption, and generation of minimal residue or waste, as well as sample preparations time limiting, with minimal analysis error. Simple dilution for powder or direct analysis for liquid, we analyzed the usefulness of EDS method in testing with field emission scanning electron microscopy (FESEM, SUPRA 55 Carl Zeiss Germany) with an X-ray energy dispersion (XFlash6l10 Bruker Germany). The samples analyzed directly without coating by applied 5µ of known concentrated diluted sample on carbon stub with accelerated voltage according to sample thickness, the result for this spot was in atomic percentage, and by Avogadro converted factor, the final result will be in microgram. Conclusion and recommendation: The conclusion of this study is application of FESEM-EDS in US pharmacopeia and ICH /Q3D guideline to reach a high-precision and accurate method in element impurities analysis of drugs or bulk materials to determine the permitted daily exposure PDE in liquid or solid specimens, and to obtain better results than other techniques, by the way it does not require complex methods or chemicals for digestion, which interfere with the final results with the possibility of to keep the sample at any time for re analysis. The recommendation is to use this technique in pharmacopeia as standard methods like inductively coupled plasma both ICP-AES, ICP-OES, and ICP-MS.

Keywords: pharmacopoeia, FESEM-EDS, element impurities, atomic concentration

Procedia PDF Downloads 113
602 Performance Comparison and Visualization of COMSOL Multiphysics, Matlab, and Fortran for Predicting the Reservoir Pressure on Oil Production in a Multiple Leases Reservoir with Boundary Element Method

Authors: N. Alias, W. Z. W. Muhammad, M. N. M. Ibrahim, M. Mohamed, H. F. S. Saipol, U. N. Z. Ariffin, N. A. Zakaria, M. S. Z. Suardi

Abstract:

This paper presents the performance comparison of some computation software for solving the boundary element method (BEM). BEM formulation is the numerical technique and high potential for solving the advance mathematical modeling to predict the production of oil well in arbitrarily shaped based on multiple leases reservoir. The limitation of data validation for ensuring that a program meets the accuracy of the mathematical modeling is considered as the research motivation of this paper. Thus, based on this limitation, there are three steps involved to validate the accuracy of the oil production simulation process. In the first step, identify the mathematical modeling based on partial differential equation (PDE) with Poisson-elliptic type to perform the BEM discretization. In the second step, implement the simulation of the 2D BEM discretization using COMSOL Multiphysic and MATLAB programming languages. In the last step, analyze the numerical performance indicators for both programming languages by using the validation of Fortran programming. The performance comparisons of numerical analysis are investigated in terms of percentage error, comparison graph and 2D visualization of pressure on oil production of multiple leases reservoir. According to the performance comparison, the structured programming in Fortran programming is the alternative software for implementing the accurate numerical simulation of BEM. As a conclusion, high-level language for numerical computation and numerical performance evaluation are satisfied to prove that Fortran is well suited for capturing the visualization of the production of oil well in arbitrarily shaped.

Keywords: performance comparison, 2D visualization, COMSOL multiphysic, MATLAB, Fortran, modelling and simulation, boundary element method, reservoir pressure

Procedia PDF Downloads 488
601 Emoji, the Language of the Future: An Analysis of the Usage and Understanding of Emoji across User-Groups

Authors: Sakshi Bhalla

Abstract:

On the one hand, given their seemingly simplistic, near universal usage and understanding, emoji are discarded as a potential step back in the evolution of communication. On the other, their effectiveness, pervasiveness, and adaptability across and within contexts are undeniable. In this study, the responses of 40 people (categorized by age) were recorded based on a uniform two-part questionnaire where they were required to a) identify the meaning of 15 emoji when placed in isolation, and b) interpret the meaning of the same 15 emoji when placed in a context-defining posting on Twitter. Their responses were studied on the basis of deviation from their responses that identified the emoji in isolation, as well as the originally intended meaning ascribed to the emoji. Based on an analysis of these results, it was discovered that each of the five age categories uses, understands and perceives emoji differently, which could be attributed to the degree of exposure they have undergone. For example, in the case of the youngest category (aged < 20), it was observed that they were the least accurate at correctly identifying emoji in isolation (~55%). Further, their proclivity to change their response with respect to the context was also the least (~31%). However, an analysis of each of their individual responses showed that these first-borns of social media seem to have reached a point where emojis no longer inspire their most literal meanings to them. The meaning and implication of these emoji have evolved to imply their context-derived meanings, even when placed in isolation. These trends carry forward meaningfully for the other four groups as well. In the case of the oldest category (aged > 35), however, the trends indicated inaccuracy and therefore, a higher incidence of a proclivity to change their responses. When studied in a continuum, the responses indicate that slowly and steadily, emoji are evolving from pictograms to ideograms. That is to suggest that they do not just indicate a one-to-one relation between a singular form and singular meaning. In fact, they communicate increasingly complicated ideas. This is much like the evolution of ancient hieroglyphics on papyrus reed or cuneiform on Sumerian clay tablets, which evolved from simple pictograms to progressively more complex ideograms. This evolution within communication is parallel to and contingent on the simultaneous evolution of communication. What’s astounding is the capacity of humans to leverage different platforms to facilitate such changes. Twiterese, as it is now called, is one of the instances where language is adapting to the demands of the digital world. That it does not have a spoken component, an ostensible grammar, and lacks standardization of use and meaning, as some might suggest, may seem like impediments in qualifying it as the 'language' of the digital world. However, that kind of a declarative remains a function of time, and time alone.

Keywords: communication, emoji, language, Twitter

Procedia PDF Downloads 93
600 Transparency Obligations under the AI Act Proposal: A Critical Legal Analysis

Authors: Michael Lognoul

Abstract:

In April 2021, the European Commission released its AI Act Proposal, which is the first policy proposal at the European Union level to target AI systems comprehensively, in a horizontal manner. This Proposal notably aims to achieve an ecosystem of trust in the European Union, based on the respect of fundamental rights, regarding AI. Among many other requirements, the AI Act Proposal aims to impose several generic transparency obligationson all AI systems to the benefit of natural persons facing those systems (e.g. information on the AI nature of systems, in case of an interaction with a human). The Proposal also provides for more stringent transparency obligations, specific to AI systems that qualify as high-risk, to the benefit of their users, notably on the characteristics, capabilities, and limitations of the AI systems they use. Against that background, this research firstly presents all such transparency requirements in turn, as well as related obligations, such asthe proposed obligations on record keeping. Secondly, it focuses on a legal analysis of their scope of application, of the content of the obligations, and on their practical implications. On the scope of transparency obligations tailored for high-risk AI systems, the research notably notes that it seems relatively narrow, given the proposed legal definition of the notion of users of AI systems. Hence, where end-users do not qualify as users, they may only receive very limited information. This element might potentially raise concern regarding the objective of the Proposal. On the content of the transparency obligations, the research highlights that the information that should benefit users of high-risk AI systems is both very broad and specific, from a technical perspective. Therefore, the information required under those obligations seems to create, prima facie, an adequate framework to ensure trust for users of high-risk AI systems. However, on the practical implications of these transparency obligations, the research notes that concern arises due to potential illiteracy of high-risk AI systems users. They might not benefit from sufficient technical expertise to fully understand the information provided to them, despite the wording of the Proposal, which requires that information should be comprehensible to its recipients (i.e. users).On this matter, the research points that there could be, more broadly, an important divergence between the level of detail of the information required by the Proposal and the level of expertise of users of high-risk AI systems. As a conclusion, the research provides policy recommendations to tackle (part of) the issues highlighted. It notably recommends to broaden the scope of transparency requirements for high-risk AI systems to encompass end-users. It also suggests that principles of explanation, as they were put forward in the Guidelines for Trustworthy AI of the High Level Expert Group, should be included in the Proposal in addition to transparency obligations.

Keywords: aI act proposal, explainability of aI, high-risk aI systems, transparency requirements

Procedia PDF Downloads 304
599 Diverse High-Performing Teams: An Interview Study on the Balance of Demands and Resources

Authors: Alana E. Jansen

Abstract:

With such a large proportion of organisations relying on the use of team-based structures, it is surprising that so few teams would be classified as high-performance teams. While the impact of team composition on performance has been researched frequently, there have been conflicting findings as to the effects, particularly when examined alongside other team factors. To broaden the theoretical perspectives on this topic and potentially explain some of the inconsistencies in research findings left open by other various models of team effectiveness and high-performing teams, the present study aims to use the Job-Demands-Resources model, typically applied to burnout and engagement, as a framework to examine how team composition factors (particularly diversity in team member characteristics) can facilitate or hamper team effectiveness. This study used a virtual interview design where participants were asked to both rate and describe their experiences, in one high-performing and one low-performing team, over several factors relating to demands, resources, team composition, and team effectiveness. A semi-structured interview protocol was developed, which combined the use of the Likert style and exploratory questions. A semi-targeted sampling approach was used to invite participants ranging in age, gender, and ethnic appearance (common surface-level diversity characteristics) and those from different specialties, roles, educational and industry backgrounds (deep-level diversity characteristics). While the final stages of data analyses are still underway, thematic analysis using a grounded theory approach was conducted concurrently with data collection to identify the point of thematic saturation, resulting in 35 interviews being completed. Analyses examine differences in perceptions of demands and resources as they relate to perceived team diversity. Preliminary results suggest that high-performing and low-performing teams differ in perceptions of the type and range of both demands and resources. The current research is likely to offer contributions to both theory and practice. The preliminary findings suggest there is a range of demands and resources which vary between high and low-performing teams, factors which may play an important role in team effectiveness research going forward. Findings may assist in explaining some of the more complex interactions between factors experienced in the team environment, making further progress towards understanding the intricacies of why only some teams achieve high-performance status.

Keywords: diversity, high-performing teams, job demands and resources, team effectiveness

Procedia PDF Downloads 184
598 Prescription of Maintenance Fluids in the Emergency Department

Authors: Adrian Craig, Jonathan Easaw, Rose Jordan, Ben Hall

Abstract:

The prescription of intravenous fluids is a fundamental component of inpatient management, but it is one which usually lacks thought. Fluids are a drug, which like any other can cause harm when prescribed inappropriately or wrongly. However, it is well recognised that it is poorly done, especially in the acute portals. The National Institute for Health and Care Excellence (NICE) recommends 1mmol/kg of potassium, sodium, and chloride per day. With various options of fluids, clinicians tend to face difficulty in choosing the most appropriate maintenance fluid, and there is a reluctance to prescribe potassium as part of an intravenous maintenance fluid regime. The aim was to prospectively audit the prescription of the first bag of intravenous maintenance fluids, the use of urea and electrolytes results to guide the choice of fluid and the use of fluid prescription charts, in a busy emergency department of a major trauma centre in Stoke-on-Trent, United Kingdom. This was undertaken over a week in early November 2016. Of those prescribed maintenance fluid only 8.9% were prescribed a fluid which was most appropriate for their daily electrolyte requirements. This audit has helped to highlight further the issues that are faced in busy Emergency Departments within hospitals that are stretched and lack capacity for prompt transfer to a ward. It has supported the findings of NICE, that emergency admission portals such as Emergency Departments poorly prescribed intravenous fluid therapy. The findings have enabled simple steps to be taken to educate clinicians about their fluid of choice. This has included: posters to remind clinicians to consider the urea and electrolyte values before prescription, suggesting the inclusion of a suggested intravenous fluid of choice in the prescription chart of the trust and the inclusion of a session within the introduction programme revising intravenous fluid therapy and daily electrolyte requirements. Moving forward, once the interventions have been implemented then, the data will be reaudited in six months to note any improvement in maintenance fluid choice. Alongside this, an audit of the rate of intravenous maintenance fluid therapy would be proposed to further increase patient safety by avoiding unintentional fluid overload which may cause unnecessary harm to patients within the hospital. In conclusion, prescription of maintenance fluid therapy was poor within the Emergency Department, and there is a great deal of opportunity for improvement. Therefore, the measures listed above will be implemented and the data reaudited.

Keywords: chloride, electrolyte, emergency department, emergency medicine, fluid, fluid therapy, intravenous, maintenance, major trauma, potassium, sodium, trauma

Procedia PDF Downloads 319
597 Rohingya Refugees and Bangladesh: Balance of Human Rights and Rationalization

Authors: Kudrat-E-Khuda Babu

Abstract:

Rohingya refugees are the most marginalized and persecuted section of people in the world. The heinous brutality of Myanmar has forced the Muslim minority community to flee themselves to their neighboring country, Bangladesh for quite a few times now. The recent atrocity of the Buddhist country has added insult to injury on the existing crisis. In lieu of protection, the rights of the Rohingya community in Myanmar are being violated through exclusion from citizenship and steamroller of persecution. The mass influx of Rohingya refugees to Bangladesh basically took place in 1978, 1992, 2012, and 2017. At present, there are around one million Rohingyas staying at Teknaf, Ukhiya of Cox’s Bazar, the southern part of Bangladesh. The country, despite being a poverty-stricken one, has shown unprecedented generosity in sheltering the Rohingya people. For sheltering half of the total refugees in 2017, the Prime Minister of Bangladesh, Sheikh Hasina is now being regarded as the lighthouse of humanity or the mother of humanity. Though Bangladesh is not a ratifying state of the UN Refugee Convention, 1951 and its Additional Protocol, 1967, the country cannot escape its obligation under international human rights jurisprudence. Bangladesh is a party to eight human rights instruments out of nine core instruments, and thus, the country has an indirect obligation to protect and promote the rights of the refugees. Pressure from international bodies has also made Bangladesh bound to provide refuge to Rohingya people. Even though the demographic vulnerability and socio-economic condition of the country do not suggest taking over extra responsibility, the principle of non-refoulment as a part of customary international law reminds us to stay beside those persecuted or believed to have well-founded fear of persecution. In the case of HM Ershad v. Bangladesh and Others, 7 BLC (AD) 67, it was held that any international treaty or document after signing or ratification is not directly enforceable unless and until the parliament enacts a similar statute howsoever sweet the document is. As per Article 33(2) of the 1951 Refugee Convention, there are even exceptions for a state party in case of serious consequences like threat to national security, apprehension of serious crime and danger to safeguard state population. Bangladesh is now at a cross-road of human rights and national interest. The world community should come forward to resolve the crisis of the persecuted Rohingya people through repatriation, resettlement, and reintegration.

Keywords: Rohingya refugees, human rights, Bangladesh, Myanmar

Procedia PDF Downloads 185
596 Tommy: Communication in Education about Disability

Authors: Karen V. Lee

Abstract:

The background and significance of this study involve communication in education by a faculty advisor exploring story and music that informs others about a disabled teacher. Social issues draw deep reflection about the emotional turmoil. As a musician becoming a teacher is a passionate yet complex endeavor, the faculty advisor shares a poetic but painful story about a disabled teacher being inducted into the teaching profession. The qualitative research method as theoretical framework draws on autoethnography of music and story where the faculty advisor approaches a professor for advice. His musicianship shifts her forward, backward, and sideways through feelings that evoke and provoke curriculum to remove communication barriers in education. They discover they do not transfer knowledge from educational method classes. Instead, the autoethnography embeds musical language as a metaphorical conduit for removing communication barriers in teacher education. Sub-themes involve communication barriers and educational technologies to ensure teachers receive social, emotional, physical, spiritual, and intervention disability resources that evoke visceral, emotional responses from the audience. Major findings of the study discover how autoethnography of music and story bring the authors to understand wider political issues of the practicum internship for teachers with disabilities. An epiphany reveals the irony of living in a culture of both uniformity and diversity. They explore the constructs of secrecy, ideology, abnormality, and marginalization by evoking visceral and emotional responses from the audience. As the voices harmonize plot, climax, characterization, and denouement, they dramatize meaning that is episodic yet incomplete to highlight the circumstances surrounding the disabled protagonist’s life. In conclusion, the qualitative research method argues for embracing storied experiences that depict communication in education. Scholarly significance embraces personal thoughts and feelings as a way of understanding social phenomena while highlighting the importance of removing communication barriers in education. The circumstance about a teacher with a disability is not uncommon in society. Thus, the authors resolve to removing barriers in education by using stories to transform the personal and cultural influences that provoke new ways of thinking about the curriculum for a disabled teacher.

Keywords: communication in education, communication barriers, autoethnography, teaching

Procedia PDF Downloads 236
595 A Segmentation Method for Grayscale Images Based on the Firefly Algorithm and the Gaussian Mixture Model

Authors: Donatella Giuliani

Abstract:

In this research, we propose an unsupervised grayscale image segmentation method based on a combination of the Firefly Algorithm and the Gaussian Mixture Model. Firstly, the Firefly Algorithm has been applied in a histogram-based research of cluster means. The Firefly Algorithm is a stochastic global optimization technique, centered on the flashing characteristics of fireflies. In this context it has been performed to determine the number of clusters and the related cluster means in a histogram-based segmentation approach. Successively these means are used in the initialization step for the parameter estimation of a Gaussian Mixture Model. The parametric probability density function of a Gaussian Mixture Model is represented as a weighted sum of Gaussian component densities, whose parameters are evaluated applying the iterative Expectation-Maximization technique. The coefficients of the linear super-position of Gaussians can be thought as prior probabilities of each component. Applying the Bayes rule, the posterior probabilities of the grayscale intensities have been evaluated, therefore their maxima are used to assign each pixel to the clusters, according to their gray-level values. The proposed approach appears fairly solid and reliable when applied even to complex grayscale images. The validation has been performed by using different standard measures, more precisely: the Root Mean Square Error (RMSE), the Structural Content (SC), the Normalized Correlation Coefficient (NK) and the Davies-Bouldin (DB) index. The achieved results have strongly confirmed the robustness of this gray scale segmentation method based on a metaheuristic algorithm. Another noteworthy advantage of this methodology is due to the use of maxima of responsibilities for the pixel assignment that implies a consistent reduction of the computational costs.

Keywords: clustering images, firefly algorithm, Gaussian mixture model, meta heuristic algorithm, image segmentation

Procedia PDF Downloads 212
594 Reconceptualizing “Best Practices” in Public Sector

Authors: Eftychia Kessopoulou, Styliani Xanthopoulou, Ypatia Theodorakioglou, George Tsiotras, Katerina Gotzamani

Abstract:

Public sector managers frequently herald that implementing best practices as a set of standards, may lead to superior organizational performance. However, recent research questions the objectification of best practices, highlighting: a) the inability of public sector organizations to develop innovative administrative practices, as well as b) the adoption of stereotypical renowned practices inculcated in the public sector by international governance bodies. The process through which organizations construe what a best practice is, still remains a black box that is yet to be investigated, given the trend of continuous changes in public sector performance, as well as the burgeoning interest of sharing popular administrative practices put forward by international bodies. This study aims to describe and understand how organizational best practices are constructed by public sector performance management teams, like benchmarkers, during the benchmarking-mediated performance improvement process and what mechanisms enable this construction. A critical realist action research methodology is employed, starting from a description of various approaches on best practice nature when a benchmarking-mediated performance improvement initiative, such as the Common Assessment Framework, is applied. Firstly, we observed the benchmarker’s management process of best practices in a public organization, so as to map their theories-in-use. As a second step we contextualized best administrative practices by reflecting the different perspectives emerged from the previous stage on the design and implementation of an interview protocol. We used this protocol to conduct 30 semi-structured interviews with “best practice” process owners, in order to examine their experiences and performance needs. Previous research on best practices has shown that needs and intentions of benchmarkers cannot be detached from the causal mechanisms of the various contexts in which they work. Such causal mechanisms can be found in: a) process owner capabilities, b) the structural context of the organization, and c) state regulations. Therefore, we developed an interview protocol theoretically informed in the first part to spot causal mechanisms suggested by previous research studies and supplemented it with questions regarding the provision of best practice support from the government. Findings of this work include: a) a causal account of the nature of best administrative practices in the Greek public sector that shed light on explaining their management, b) a description of the various contexts affecting best practice conceptualization, and c) a description of how their interplay changed the organization’s best practice management.

Keywords: benchmarking, action research, critical realism, best practices, public sector

Procedia PDF Downloads 122
593 LTE Performance Analysis in the City of Bogota Northern Zone for Two Different Mobile Broadband Operators over Qualipoc

Authors: Víctor D. Rodríguez, Edith P. Estupiñán, Juan C. Martínez

Abstract:

The evolution in mobile broadband technologies has allowed to increase the download rates in users considering the current services. The evaluation of technical parameters at the link level is of vital importance to validate the quality and veracity of the connection, thus avoiding large losses of data, time and productivity. Some of these failures may occur between the eNodeB (Evolved Node B) and the user equipment (UE), so the link between the end device and the base station can be observed. LTE (Long Term Evolution) is considered one of the IP-oriented mobile broadband technologies that work stably for data and VoIP (Voice Over IP) for those devices that have that feature. This research presents a technical analysis of the connection and channeling processes between UE and eNodeB with the TAC (Tracking Area Code) variables, and analysis of performance variables (Throughput, Signal to Interference and Noise Ratio (SINR)). Three measurement scenarios were proposed in the city of Bogotá using QualiPoc, where two operators were evaluated (Operator 1 and Operator 2). Once the data were obtained, an analysis of the variables was performed determining that the data obtained in transmission modes vary depending on the parameters BLER (Block Error Rate), performance and SNR (Signal-to-Noise Ratio). In the case of both operators, differences in transmission modes are detected and this is reflected in the quality of the signal. In addition, due to the fact that both operators work in different frequencies, it can be seen that Operator 1, despite having spectrum in Band 7 (2600 MHz), together with Operator 2, is reassigning to another frequency, a lower band, which is AWS (1700 MHz), but the difference in signal quality with respect to the establishment with data by the provider Operator 2 and the difference found in the transmission modes determined by the eNodeB in Operator 1 is remarkable.

Keywords: BLER, LTE, network, qualipoc, SNR.

Procedia PDF Downloads 113
592 A Hybrid Block Multistep Method for Direct Numerical Integration of Fourth Order Initial Value Problems

Authors: Adamu S. Salawu, Ibrahim O. Isah

Abstract:

Direct solution to several forms of fourth-order ordinary differential equations is not easily obtained without first reducing them to a system of first-order equations. Thus, numerical methods are being developed with the underlying techniques in the literature, which seeks to approximate some classes of fourth-order initial value problems with admissible error bounds. Multistep methods present a great advantage of the ease of implementation but with a setback of several functions evaluation for every stage of implementation. However, hybrid methods conventionally show a slightly higher order of truncation for any k-step linear multistep method, with the possibility of obtaining solutions at off mesh points within the interval of solution. In the light of the foregoing, we propose the continuous form of a hybrid multistep method with Chebyshev polynomial as a basis function for the numerical integration of fourth-order initial value problems of ordinary differential equations. The basis function is interpolated and collocated at some points on the interval [0, 2] to yield a system of equations, which is solved to obtain the unknowns of the approximating polynomial. The continuous form obtained, its first and second derivatives are evaluated at carefully chosen points to obtain the proposed block method needed to directly approximate fourth-order initial value problems. The method is analyzed for convergence. Implementation of the method is done by conducting numerical experiments on some test problems. The outcome of the implementation of the method suggests that the method performs well on problems with oscillatory or trigonometric terms since the approximations at several points on the solution domain did not deviate too far from the theoretical solutions. The method also shows better performance compared with an existing hybrid method when implemented on a larger interval of solution.

Keywords: Chebyshev polynomial, collocation, hybrid multistep method, initial value problems, interpolation

Procedia PDF Downloads 120
591 Fem Models of Glued Laminated Timber Beams Enhanced by Bayesian Updating of Elastic Moduli

Authors: L. Melzerová, T. Janda, M. Šejnoha, J. Šejnoha

Abstract:

Two finite element (FEM) models are presented in this paper to address the random nature of the response of glued timber structures made of wood segments with variable elastic moduli evaluated from 3600 indentation measurements. This total database served to create the same number of ensembles as was the number of segments in the tested beam. Statistics of these ensembles were then assigned to given segments of beams and the Latin Hypercube Sampling (LHS) method was called to perform 100 simulations resulting into the ensemble of 100 deflections subjected to statistical evaluation. Here, a detailed geometrical arrangement of individual segments in the laminated beam was considered in the construction of two-dimensional FEM model subjected to in four-point bending to comply with the laboratory tests. Since laboratory measurements of local elastic moduli may in general suffer from a significant experimental error, it appears advantageous to exploit the full scale measurements of timber beams, i.e. deflections, to improve their prior distributions with the help of the Bayesian statistical method. This, however, requires an efficient computational model when simulating the laboratory tests numerically. To this end, a simplified model based on Mindlin’s beam theory was established. The improved posterior distributions show that the most significant change of the Young’s modulus distribution takes place in laminae in the most strained zones, i.e. in the top and bottom layers within the beam center region. Posterior distributions of moduli of elasticity were subsequently utilized in the 2D FEM model and compared with the original simulations.

Keywords: Bayesian inference, FEM, four point bending test, laminated timber, parameter estimation, prior and posterior distribution, Young’s modulus

Procedia PDF Downloads 281
590 Hybrid Knowledge and Data-Driven Neural Networks for Diffuse Optical Tomography Reconstruction in Medical Imaging

Authors: Paola Causin, Andrea Aspri, Alessandro Benfenati

Abstract:

Diffuse Optical Tomography (DOT) is an emergent medical imaging technique which employs NIR light to estimate the spatial distribution of optical coefficients in biological tissues for diagnostic purposes, in a noninvasive and non-ionizing manner. DOT reconstruction is a severely ill-conditioned problem due to prevalent scattering of light in the tissue. In this contribution, we present our research in adopting hybrid knowledgedriven/data-driven approaches which exploit the existence of well assessed physical models and build upon them neural networks integrating the availability of data. Namely, since in this context regularization procedures are mandatory to obtain a reasonable reconstruction [1], we explore the use of neural networks as tools to include prior information on the solution. 2. Materials and Methods The idea underlying our approach is to leverage neural networks to solve PDE-constrained inverse problems of the form 𝒒 ∗ = 𝒂𝒓𝒈 𝒎𝒊𝒏𝒒 𝐃(𝒚, 𝒚̃), (1) where D is a loss function which typically contains a discrepancy measure (or data fidelity) term plus other possible ad-hoc designed terms enforcing specific constraints. In the context of inverse problems like (1), one seeks the optimal set of physical parameters q, given the set of observations y. Moreover, 𝑦̃ is the computable approximation of y, which may be as well obtained from a neural network but also in a classic way via the resolution of a PDE with given input coefficients (forward problem, Fig.1 box ). Due to the severe ill conditioning of the reconstruction problem, we adopt a two-fold approach: i) we restrict the solutions (optical coefficients) to lie in a lower-dimensional subspace generated by auto-decoder type networks. This procedure forms priors of the solution (Fig.1 box ); ii) we use regularization procedures of type 𝒒̂ ∗ = 𝒂𝒓𝒈𝒎𝒊𝒏𝒒 𝐃(𝒚, 𝒚̃)+ 𝑹(𝒒), where 𝑹(𝒒) is a regularization functional depending on regularization parameters which can be fixed a-priori or learned via a neural network in a data-driven modality. To further improve the generalizability of the proposed framework, we also infuse physics knowledge via soft penalty constraints (Fig.1 box ) in the overall optimization procedure (Fig.1 box ). 3. Discussion and Conclusion DOT reconstruction is severely hindered by ill-conditioning. The combined use of data-driven and knowledgedriven elements is beneficial and allows to obtain improved results, especially with a restricted dataset and in presence of variable sources of noise.

Keywords: inverse problem in tomography, deep learning, diffuse optical tomography, regularization

Procedia PDF Downloads 72
589 Impact of Diabetes Mellitus Type 2 on Clinical In-Stent Restenosis in First Elective Percutaneous Coronary Intervention Patients

Authors: Leonard Simoni, Ilir Alimehmeti, Ervina Shirka, Endri Hasimi, Ndricim Kallashi, Verona Beka, Suerta Kabili, Artan Goda

Abstract:

Background: Diabetes Mellitus type 2, small vessel calibre, stented length of vessel, complex lesion morphology, and prior bypass surgery have resulted risk factors for In-Stent Restenosis (ISR). However, there are some contradictory results about body mass index (BMI) as a risk factor for ISR. Purpose: We want to identify clinical, lesional and procedural factors that can predict clinical ISR in our patients. Methods: Were enrolled 759 patients who underwent first-time elective PCI with Bare Metal Stents (BMS) from September 2011 to December 2013 in our Department of Cardiology and followed them for at least 1.5 years with a median of 862 days (2 years and 4 months). Only the patients re-admitted with ischemic heart disease underwent control coronary angiography but no routine angiographic control was performed. Patients were categorized in ISR and non-ISR groups and compared between them. Multivariate analysis - Binary Logistic Regression: Forward Conditional Method was used to identify independent predictive risk factors. P was considered statistically significant when <0.05. Results: ISR compared to non-ISR individuals had a significantly lower BMI (25.7±3.3 vs. 26.9±3.7, p=0.004), higher risk anatomy (LM + 3-vessel CAD) (23% vs. 14%, p=0.03), higher number of stents/person used (2.1±1.1 vs. 1.75±0.96, p=0.004), greater length of stents/person used (39.3±21.6 vs. 33.3±18.5, p=0.01), and a lower use of clopidogrel and ASA (together) (95% vs. 99%, p=0.012). They also had a higher, although not statistically significant, prevalence of Diabetes Mellitus (42% vs. 32%, p=0.072) and a greater number of treated vessels (1.36±0.5 vs. 1.26±0.5, p=0.08). In the multivariate analysis, Diabetes Mellitus type 2 and multiple stents used were independent predictors risk factors for In-Stent Restenosis, OR 1.66 [1.03-2.68], p=0.039, and OR 1.44 [1.16-1.78,] p=0.001, respectively. On the other side higher BMI and use of clopidogrel and ASA together resulted protective factors OR 0.88 [0.81-0.95], p=0.001 and OR 0.2 [0.06-0.72] p=0.013, respectively. Conclusion: Diabetes Mellitus and multiple stents are strong predictive risk factors, whereas the use of clopidogrel and ASA together are protective factors for clinical In-Stent Restenosis. Paradoxically High BMI is a protective factor for In-stent Restenosis, probably related to a larger diameter of vessels and consequently a larger diameter of stents implanted in these patients. Further studies are needed to clarify this finding.

Keywords: body mass index, diabetes mellitus, in-stent restenosis, percutaneous coronary intervention

Procedia PDF Downloads 207
588 An Intelligent Controller Augmented with Variable Zero Lag Compensation for Antilock Braking System

Authors: Benjamin Chijioke Agwah, Paulinus Chinaenye Eze

Abstract:

Antilock braking system (ABS) is one of the important contributions by the automobile industry, designed to ensure road safety in such way that vehicles are kept steerable and stable when during emergency braking. This paper presents a wheel slip-based intelligent controller with variable zero lag compensation for ABS. It is required to achieve a very fast perfect wheel slip tracking during hard braking condition and eliminate chattering with improved transient and steady state performance, while shortening the stopping distance using effective braking torque less than maximum allowable torque to bring a braking vehicle to a stop. The dynamic of a vehicle braking with a braking velocity of 30 ms⁻¹ on a straight line was determined and modelled in MATLAB/Simulink environment to represent a conventional ABS system without a controller. Simulation results indicated that system without a controller was not able to track desired wheel slip and the stopping distance was 135.2 m. Hence, an intelligent control based on fuzzy logic controller (FLC) was designed with a variable zero lag compensator (VZLC) added to enhance the performance of FLC control variable by eliminating steady state error, provide improve bandwidth to eliminate the effect of high frequency noise such as chattering during braking. The simulation results showed that FLC- VZLC provided fast tracking of desired wheel slip, eliminate chattering, and reduced stopping distance by 70.5% (39.92 m), 63.3% (49.59 m), 57.6% (57.35 m) and 50% (69.13 m) on dry, wet, cobblestone and snow road surface conditions respectively. Generally, the proposed system used effective braking torque that is less than the maximum allowable braking torque to achieve efficient wheel slip tracking and overall robust control performance on different road surfaces.

Keywords: ABS, fuzzy logic controller, variable zero lag compensator, wheel slip tracking

Procedia PDF Downloads 145
587 Harmonization of Accreditation Standards in Education of Central Asian Countries: Theoretical Aspect

Authors: Yskak Nabi, Onolkan Umankulova, Ilyas Seitov

Abstract:

Tempus project about “Central Asian network for quality assurance – CANQA” had been implemented in 2009-2012. As the result of the project, two accreditation agencies were established: the agency for quality assurance in the field of education, “EdNet” in Kyrgyzstan, center of progressive technologies in Tajikistan. The importance of the research studies of the project is supported by the idea that the creation of Central-Asian network for quality assurance in education is still relevant, and results of the International forum “Global in regional: Kazakhstan in Bologna process and EU projects,” that was held in Nur-Sultan in October 2020, proves this. At the same time, the previous experience of the partnership between accreditation agencies of Central Asia shows that recommendations elaborated within the CANQA project were not theoretically justified. But there are a number of facts and arguments that prove the practical appliance of these recommendations. In this respect, joint activities of accreditation agencies of Kyrgyzstan and Kazakhstan are representative. For example, independent Kazakh agency of accreditation and rating successfully conducts accreditation of Kyrgyz universities; based on the memorandum about joint activity between the agency for quality assurance in the field of education “EdNet” (Kyrgyzstan) and Astana accreditation agency (Kazakhstan), the last one provides its experts for accreditation procedures in EdNet. Exchange of experience among the agencies shows an effective approach towards adaptation of European standards to the reality of education systems of Central Asia with consideration of not only a legal framework but also from the point of European practices view. Therefore, the relevance of the research is identified as there is a practical partnership between accreditation agencies of Central Asian countries, but the absence of theoretical justification of integrational processes in the accreditation field. As a result, the following hypothesis was put forward: “if to develop theoretical aspects for harmonization of accreditation standards, then integrational processes would be improved since the implementation of Bologna process principles would be supported with wider possibilities, and particularly, students and academic mobility would be improved.” Indeed, for example, in Kazakhstan, the total share of foreign students was 5,04% in 2020, and most of them are coming from Kyrgyzstan, Tajikistan, and Uzbekistan, and if integrational processes will be improved, then this share can increase.

Keywords: accreditation standards in education, Central Asian countries, pedagogical theory, model

Procedia PDF Downloads 195
586 Hedgerow Detection and Characterization Using Very High Spatial Resolution SAR DATA

Authors: Saeid Gharechelou, Stuart Green, Fiona Cawkwell

Abstract:

Hedgerow has an important role for a wide range of ecological habitats, landscape, agriculture management, carbon sequestration, wood production. Hedgerow detection accurately using satellite imagery is a challenging problem in remote sensing techniques, because in the special approach it is very similar to line object like a road, from a spectral viewpoint, a hedge is very similar to a forest. Remote sensors with very high spatial resolution (VHR) recently enable the automatic detection of hedges by the acquisition of images with enough spectral and spatial resolution. Indeed, recently VHR remote sensing data provided the opportunity to detect the hedgerow as line feature but still remain difficulties in monitoring the characterization in landscape scale. In this research is used the TerraSAR-x Spotlight and Staring mode with 3-5 m resolution in wet and dry season in the test site of Fermoy County, Ireland to detect the hedgerow by acquisition time of 2014-2015. Both dual polarization of Spotlight data in HH/VV is using for detection of hedgerow. The varied method of SAR image technique with try and error way by integration of classification algorithm like texture analysis, support vector machine, k-means and random forest are using to detect hedgerow and its characterization. We are applying the Shannon entropy (ShE) and backscattering analysis in single and double bounce in polarimetric analysis for processing the object-oriented classification and finally extracting the hedgerow network. The result still is in progress and need to apply the other method as well to find the best method in study area. Finally, this research is under way to ahead to get the best result and here just present the preliminary work that polarimetric image of TSX potentially can detect the hedgerow.

Keywords: TerraSAR-X, hedgerow detection, high resolution SAR image, dual polarization, polarimetric analysis

Procedia PDF Downloads 228
585 Translation And Cultural Adaptation Of The Rivermead Behavioural Memory Test–3rd Edition Into the Arabic Language

Authors: Mai Alharthy, Agnes Shiel, Hynes Sinead

Abstract:

Objectives: The objectives of the study are to translate and culturally adapt the RBMT-3 to be appropriate for use within an Arabic-speaking population and to achieve maximum equivalency between the translated and original versions and to evaluate the psychometric properties of the Arabic version of the RBMT-3. Participants' numbers are 16 (10 females and 6 males). All participants are bilingual speakers of Arabic and English, above 18 years old and with no current nor past memory impairment. Methods: The study was conducted in two stages: Translation and cultural adaptation stage: Forward and backward translations were completed by professional translators. Five out of the 14 RBMT-3 subtests required cultural adaptations. Half of the faces in the face recognition subtests were replaced with Arabic faces by a professional photographer. Pictures that are irrelevant to the Arabic culture in the picture recognition subtests were replaced. Names, story and orientations subtests were also adapted to suit the Arabic culture. An expert committee was formed to compare the translated and original versions and to advise on further changes required for test materials. Validation of the Arabic RBMT-3- pilot: 16 Participants were tested on version 1 of the English version and the two versions of the Arabic RBMT-3 ( counterbalanced ). The assessment period was 6 weeks long, with two weeks gap between tests. All assessments took place in a quiet room in the National University of Ireland Galway. Two qualified occupational therapists completed the assessments. Results: Wilcox signed-rank test was used to compare between subtest scores. Significant differences were found in the story, orientation and names subtests between the English and Arabic versions. No significant differences were found in subtests from both Arabic versions except for the story subtest. Conclusion: The story and orientation subtests should be revised by the expert committee members to make further adaptations. The rest of the Arabic RBMT-3 subtests are equivalent to the subtests of the English version. The psychometric properties of the Arabic RBMT-3 will be investigated in a larger Arabic-speaking sample in Saudi Arabia. The outcome of this research is to provide clinicians and researchers with a reliable tool to assess memory problems in Arabic speaking population.

Keywords: memory impairment, neuropsychological assessment, cultural adaptation, cognitive assessment

Procedia PDF Downloads 253
584 Proposed Design of an Optimized Transient Cavity Picosecond Ultraviolet Laser

Authors: Marilou Cadatal-Raduban, Minh Hong Pham, Duong Van Pham, Tu Nguyen Xuan, Mui Viet Luong, Kohei Yamanoi, Toshihiko Shimizu, Nobuhiko Sarukura, Hung Dai Nguyen

Abstract:

There is a great deal of interest in developing all-solid-state tunable ultrashort pulsed lasers emitting in the ultraviolet (UV) region for applications such as micromachining, investigation of charge carrier relaxation in conductors, and probing of ultrafast chemical processes. However, direct short-pulse generation is not as straight forward in solid-state gain media as it is for near-IR tunable solid-state lasers such as Ti:sapphire due to the difficulty of obtaining continuous wave laser operation, which is required for Kerr lens mode-locking schemes utilizing spatial or temporal Kerr type nonlinearity. In this work, the transient cavity method, which was reported to generate ultrashort laser pulses in dye lasers, is extended to a solid-state gain medium. Ce:LiCAF was chosen among the rare-earth-doped fluoride laser crystals emitting in the UV region because of its broad tunability (from 280 to 325 nm) and enough bandwidth to generate 3-fs pulses, sufficiently large effective gain cross section (6.0 x10⁻¹⁸ cm²) favorable for oscillators, and a high saturation fluence (115 mJ/cm²). Numerical simulations are performed to investigate the spectro-temporal evolution of the broadband UV laser emission from Ce:LiCAF, represented as a system of two homogeneous broadened singlet states, by solving the rate equations extended to multiple wavelengths. The goal is to find the appropriate cavity length and Q-factor to achieve the optimal photon cavity decay time and pumping energy for resonator transients that will lead to ps UV laser emission from a Ce:LiCAF crystal pumped by the fourth harmonics (266nm) of a Nd:YAG laser. Results show that a single ps pulse can be generated from a 1-mm, 1 mol% Ce³⁺-doped LiCAF crystal using an output coupler with 10% reflectivity (low-Q) and an oscillator cavity that is 2-mm long (short cavity). This technique can be extended to other fluoride-based solid-state laser gain media.

Keywords: rare-earth-doped fluoride gain medium, transient cavity, ultrashort laser, ultraviolet laser

Procedia PDF Downloads 353
583 Improving Fingerprinting-Based Localization System Using Generative AI

Authors: Getaneh Berie Tarekegn

Abstract:

A precise localization system is crucial for many artificial intelligence Internet of Things (AI-IoT) applications in the era of smart cities. Their applications include traffic monitoring, emergency alarming, environmental monitoring, location-based advertising, intelligent transportation, and smart health care. The most common method for providing continuous positioning services in outdoor environments is by using a global navigation satellite system (GNSS). Due to nonline-of-sight, multipath, and weather conditions, GNSS systems do not perform well in dense urban, urban, and suburban areas.This paper proposes a generative AI-based positioning scheme for large-scale wireless settings using fingerprinting techniques. In this article, we presented a semi-supervised deep convolutional generative adversarial network (S-DCGAN)-based radio map construction method for real-time device localization. It also employed a reliable signal fingerprint feature extraction method with t-distributed stochastic neighbor embedding (t-SNE), which extracts dominant features while eliminating noise from hybrid WLAN and long-term evolution (LTE) fingerprints. The proposed scheme reduced the workload of site surveying required to build the fingerprint database by up to 78.5% and significantly improved positioning accuracy. The results show that the average positioning error of GAILoc is less than 0.39 m, and more than 90% of the errors are less than 0.82 m. According to numerical results, SRCLoc improves positioning performance and reduces radio map construction costs significantly compared to traditional methods.

Keywords: location-aware services, feature extraction technique, generative adversarial network, long short-term memory, support vector machine

Procedia PDF Downloads 54
582 The Dark Side of the Fight against Organised Crime

Authors: Ana M. Prieto del Pino

Abstract:

As is well known, UN Convention against Illicit Traffic in Narcotic Drugs and Psychotropic Substances (1988) was a landmark regarding the seizure of proceeds of crime. Depriving criminals of the profits from their activity became a priority at an international level in the fight against organised crime. Enabling confiscation of proceeds of illicit traffic in narcotic drugs and psychotropic substances, criminalising money laundering and confiscating the proceeds thereof are the three measures taken in order to achieve that purpose. The beginning of 21st century brought the declaration of war on corruption and on the illicit enjoyment of the profits thereof onto the international scene. According to the UN Convention against Transnational Organised Crime (2000), States Parties should adopt the necessary measures to enable the confiscation of proceeds of crime derived from offences (or property of equivalent value) and property, equipment and other instrumentalities used in offences covered by that Convention. The UN Convention against Corruption (2003) states asset recovery explicitly as a fundamental principle and sets forth measures aiming at the direct recovery of property through international cooperation in confiscation. Furthermore, European legislation has made many significant strides forward in less than twenty years concerning money laundering, confiscation, and asset recovery. Crime does not pay, let there be no doubt about it. Nevertheless, we must be very careful not to sing out of tune with individual rights and legal guarantees. On the one hand, innocent individuals and businesses must be protected, since they should not pay for the guilty ones’ faults. On the other hand, the rule of law must be preserved and not be tossed aside regarding those who have carried out criminal activities. An in-depth analysis of judicial decisions on money laundering and confiscation of proceeds of crime issued by European national courts and by the European Court of Human Rights in the last decade has been carried out from a human rights, legal guarantees and criminal law basic principles’ perspective. The undertaken study has revealed the violation of the right to property, of the proportionality principle legal and the infringement of basic principles of states’ domestic substantive and procedural criminal law systems. The most relevant ones have to do with the punishment of money laundering committed through negligence, non-conviction based confiscation and a too-far reaching interpretation of the notion of ‘proceeds of crime’. Almost everything in life has a bright and a dark side. Confiscation of criminal proceeds and asset recovery are not an exception to this rule.

Keywords: confiscation, human rights, money laundering, organized crime

Procedia PDF Downloads 138
581 Factors Impacting Training and Adult Education Providers’ Business Performance: The Singapore Context

Authors: Zan Chen, D. Kwok

Abstract:

The SkillsFuture Singapore’s mission to develop a responsive and forward-looking Training and Adult Education (TAE) and workforce development system is undergirded by how successful TAE providers are in their business performance and strategies that strengthen their operational efficiency and processes. Therefore, understanding the factors that drive the business performance of TAE providers is critical to the success of SkillsFuture Singapore’s initiatives. This study aims to investigate how business strategy, work autonomy, work intensity and professional development support impact the business performance of private TAE providers. Specifically, the three research questions are: (1) Are there significant relationships between the above-mentioned four factors and TAE providers’ business performance?; (2) Are there significant differences on the four factors between low and high TAE providers’ business performance groups?; and (3) To what extent and in what manner do the four factors predict TAE providers’ business performance? This was part of the first national study on organizations and professionals working in the Training and Adult Education (TAE) sector. Data from 265 private TAE providers where respondents were Chief Executive Officers representatives from the Senior Management were analyzed. The results showed that business strategy (the extent that the organization leads the way in terms of developing new products and services; uses up-to-date learning technologies; customizes its products and services to the client’s needs), work autonomy (the extent that the staff personally have an influence on how hard they work; deciding what tasks they are to do; deciding how they are to do the tasks, and deciding the quality standards to which they work) and professional development support (both monetary and non-monetary support and incentives) had positive and significant relationships with business performance. However, no significant relationship is found between work intensity and business performance. A business strategy, work autonomy and professional development support were significantly higher in the high business performance group compared to the low-performance group among the TAE providers. Results of hierarchical regression analyses controlling for the size of the TAE providers showed significant impacts of business strategy, work autonomy and professional development support on TAE providers’ business performance. Overall, the model accounted for 27% of the variance in TAE providers’ business performance. This study provides policymakers with insights into improving existing policies, designing new initiatives and implementing targeting interventions to support TAE providers. The findings also have implications on how the TAE providers could better formulate their organizational strategies and business models. Finally, limitations of study, along with directions for future research will be discussed in the paper.

Keywords: adult education, business performance, business strategy, training, work autonomy

Procedia PDF Downloads 203
580 The Co-Existence of Multidominance and Movement in the Syntax of Chinese Bi-Comparatives

Authors: Yaqing Hu

Abstract:

This paper puts forward a syntactic analysis involving multidominance and rightward movement in Chinese bi-comparatives, as in 'Yuehan bi Mali gao (John is taller than Mary).' It is argued here that the predicate of comparison is a shared constituent in two small clauses, namely one for the target and one for the standard; and then it moves rightward to form a degree phrase with the comparative morpheme. This proposal comes from four aspects. First, the example above can also be expressed in this way, 'A: Yuehan he Mali, shui gao? (John and Mary, who is taller?) B: Yuehan gao./Yuehan geng gao. (John is taller).' This shows that the gradable adjective is predicated of the target. In addition, according to a constraint on Chinese bi-comparatives, namely the target and the standard must be arguments of the predicate simultaneously, it is not unreasonable to assume that the gradable adjective may also be predicated of the standard. Second, subcomparatives are totally disallowed in Chinese, as in '*zhe-zhang zhuozi bi zhe-zhang yizi kuan chang. (This table is longer than this chair is wide.)' In order to save it from ungrammaticality, the target and the standard should be compared along the same dimension denoted by the gradable adjective. It may follow that in Chinese comparatives, having equal roles in the same eventuality, the target and the standard bear the same thematic relationship with the predicate of comparison. Third, verb-copy can appear in Chinese bi-comparatives, as in 'Yuehan qi ma bi Mali qi ma qi de kuai. (John rides horses faster than Mary does.)' The predicate qi seems to form a small clause with both the target and the standard. This might be supporting evidence that both the target and the standard share the predicate of comparison. Fourth, Chinese comparatives do have comparative morphemes, as in 'Yuehan bi Mali geng gao. (John is taller than Mary)', which is semantically equivalent to the first example above. Thus, it follows that one feature of Chinese comparative morphemes is that they can remain overt or covert in the syntax, which will not affect semantics. This further shows that comparative morphemes in bi-comparatives may not be able to saturate the degree argument denoted by the predicate of comparison due to its optionality in the structure. These four aspects present a challenge to the Direct Analysis used in Chinese comparatives since this approach would presume that the target and the standard somehow show independency with the predicate in the syntax. Meanwhile, this study also rejects the previous analysis of multidomiance in bi-comparatives in which the degree phrase comprised of the comparative morpheme and the gradable adjective may be shared by the standard when the comparative morpheme is covert. This syntactic analysis proposed in this study will therefore offer a different perspective of how to treat degree phrase in Chinese comparatives and may offer evidence to argue whether there is degree phrase movement in bi-comparatives as in its English counterparts.

Keywords: Chinese comparatives, degree phrase, movement, multidominance, syntactic analysis

Procedia PDF Downloads 324
579 Depth Camera Aided Dead-Reckoning Localization of Autonomous Mobile Robots in Unstructured GNSS-Denied Environments

Authors: David L. Olson, Stephen B. H. Bruder, Adam S. Watkins, Cleon E. Davis

Abstract:

In global navigation satellite systems (GNSS), denied settings such as indoor environments, autonomous mobile robots are often limited to dead-reckoning navigation techniques to determine their position, velocity, and attitude (PVA). Localization is typically accomplished by employing an inertial measurement unit (IMU), which, while precise in nature, accumulates errors rapidly and severely degrades the localization solution. Standard sensor fusion methods, such as Kalman filtering, aim to fuse precise IMU measurements with accurate aiding sensors to establish a precise and accurate solution. In indoor environments, where GNSS and no other a priori information is known about the environment, effective sensor fusion is difficult to achieve, as accurate aiding sensor choices are sparse. However, an opportunity arises by employing a depth camera in the indoor environment. A depth camera can capture point clouds of the surrounding floors and walls. Extracting attitude from these surfaces can serve as an accurate aiding source, which directly combats errors that arise due to gyroscope imperfections. This configuration for sensor fusion leads to a dramatic reduction of PVA error compared to traditional aiding sensor configurations. This paper provides the theoretical basis for the depth camera aiding sensor method, initial expectations of performance benefit via simulation, and hardware implementation, thus verifying its veracity. Hardware implementation is performed on the Quanser Qbot 2™ mobile robot, with a Vector-Nav VN-200™ IMU and Kinect™ camera from Microsoft.

Keywords: autonomous mobile robotics, dead reckoning, depth camera, inertial navigation, Kalman filtering, localization, sensor fusion

Procedia PDF Downloads 201
578 AI-Driven Solutions for Optimizing Master Data Management

Authors: Srinivas Vangari

Abstract:

In the era of big data, ensuring the accuracy, consistency, and reliability of critical data assets is crucial for data-driven enterprises. Master Data Management (MDM) plays a crucial role in this endeavor. This paper investigates the role of Artificial Intelligence (AI) in enhancing MDM, focusing on how AI-driven solutions can automate and optimize various stages of the master data lifecycle. By integrating AI (Quantitative and Qualitative Analysis) into processes such as data creation, maintenance, enrichment, and usage, organizations can achieve significant improvements in data quality and operational efficiency. Quantitative analysis is employed to measure the impact of AI on key metrics, including data accuracy, processing speed, and error reduction. For instance, our study demonstrates an 18% improvement in data accuracy and a 75% reduction in duplicate records across multiple systems post-AI implementation. Furthermore, AI’s predictive maintenance capabilities reduced data obsolescence by 22%, as indicated by statistical analyses of data usage patterns over a 12-month period. Complementing this, a qualitative analysis delves into the specific AI-driven strategies that enhance MDM practices, such as automating data entry and validation, which resulted in a 28% decrease in manual errors. Insights from case studies highlight how AI-driven data cleansing processes reduced inconsistencies by 25% and how AI-powered enrichment strategies improved data relevance by 24%, thus boosting decision-making accuracy. The findings demonstrate that AI significantly enhances data quality and integrity, leading to improved enterprise performance through cost reduction, increased compliance, and more accurate, real-time decision-making. These insights underscore the value of AI as a critical tool in modern data management strategies, offering a competitive edge to organizations that leverage its capabilities.

Keywords: artificial intelligence, master data management, data governance, data quality

Procedia PDF Downloads 9