Search results for: digital space
3887 Left Posterior Pericardiotomy in the Prevention of Post-Operative Atrial Fibrillation and Cardiac Tamponade: A Retrospective Study of 2118 Isolated Coronary Artery Bypass Graft Patients
Authors: Ayeshmanthe Rathnayake, Siew Goh, Carmel Fenton, Ashutosh Hardikar
Abstract:
Post-Operative Atrial Fibrillation (POAF) is the most frequent complication of cardiac surgery and is associated with reduced survival, increased rates of cognitive changes and cerebrovascular accident, heart failure, renal dysfunction, infection and length of stay, and hospital costs. Cardiac tamponade, although less common, carries high morbidity and mortality. Shed mediastinal blood in the pericardial space is a major source of intrapericardial oxidative stress and inflammation that triggers POAF. The utilisation of a left posterior pericardiotomy aims to shunt blood from the pericardium into the pleural space and have a role in the prevention of POAF as well as cardiac tamponade. 2118 patients had undergone isolated Coronary Artery Bypass Graft (CABG) at Royal Hobart Hospital from 2008-2021. They were divided into pericardiotomy vs control group. Patient baseline demographics, intraoperative data, and post-operative outcomes were reviewed retrospectively. Total incidence of new POAF and cardiac tamponade was 26.1% and 0.75%, respectively. Primary outcome of both the incidence of POAF(22.9% vs27.8%OR 0.77 p<0.05) and Cardiac Tamponade (0% vs 1.1% OR 0.85 p<0.05) were less in the pericardiotomy group.Increasing age, BMI, poor left ventricular function (EF <30%), and return to theatre were independent predictors of developing POAF. There were similar rates of return to theatre for bleeding however, no cases of tamponade in the pericardiotomy group. There were no complications attributable to left posterior pericardiotomy and the time added to the duration of surgery was minimal. Left posterior pericardiotomy is associated with a significant reduction in the incidence of POAFand cardiac tamponade and issafe and efficient.Keywords: cardiac surgery, pericardiotomy, post-operative atrial fibrillation, cardiac tamponade
Procedia PDF Downloads 913886 On a Negative Relation between Bacterial Taxis and Turing Pattern Formation
Authors: A. Elragig, S. Townley, H. Dreiwi
Abstract:
In this paper we introduce a bacteria-leukocyte model with bacteria chemotaxsis. We assume that bacteria develop a tactic defense mechanism as a response to Leukocyte phagocytosis. We explore the effect of this tactic motion on Turing space in two parameter spaces. A fine tuning of bacterial chemotaxis shows a significant effect on developing a non-uniform steady state.Keywords: chemotaxis-diffusion driven instability, bacterial chemotaxis, mathematical biology, ecology
Procedia PDF Downloads 3683885 Measurement of Radon Exhalation Rate, Natural Radioactivity, and Radiation Hazard Assessment in Soil Samples from the Surrounding Area of Kasimpur Thermal Power Plant Kasimpur (U. P.), India
Authors: Anil Sharma, Ajay Kumar Mahur, R. G. Sonkawade, A. C. Sharma, R. Prasad
Abstract:
In coal fired thermal power stations, large amount of fly ash is produced after burning of coal. Fly ash is spread and distributed in the surrounding area by air and may be deposited on the soil of the region surrounding the power plant. Coal contains increased levels of these radionuclides and fly ash may increase the radioactivity in the soil around the power plant. Radon atoms entering into the pore space from the mineral grain are transported by diffusion and advection through this space until they in turn decay or are released into the atmosphere. In the present study, Soil samples were collected from the region around a Kasimpur Thermal Power Plant, Kasimpur, Aligarh (U.P.). Radon activity, radon surface exhalation and mass exhalation rates were measured using “sealed can technique” using LR 115-type II nuclear track detectors. Radon activities vary from 92.9 to 556.8 Bq m-3 with mean value of 279.8 Bq m-3. Surface exhalation rates (EX) in these samples are found to vary from 33.4 to 200.2 mBq m-2 h-1 with an average value of 100.5 mBq m-2 h-1 whereas, Mass exhalation rates (EM) vary from 1.2 to 7.7 mBq kg-1 h-1 with an average value of 3.8 mBq kg-1 h-1. Activity concentrations of radionuclides were measured in these samples by using a low level NaI (Tl) based gamma ray spectrometer. Activity concentrations of 226Ra 232Th and 40K vary from 12 to 49 Bq kg-1, 24 to 49 Bq kg-1 and 135 to 546 Bq kg-1 with overall mean values of 30.3 Bq kg-1, 38.5 Bq kg-1 and 317.8 Bq kg-1, respectively. Radium equivalent activity has been found to vary from 80.0 to 143.7 Bq kg-1 with an average value of 109.7 Bq kg-1. Absorbed dose rate varies from 36.1 to 66.4 nGy h-1 with an average value of 50.4 nGy h-1 and corresponding outdoor annual effective dose varies from 0.044 to 0.081 mSv with an average value of 0.061 mSv. Values of external and internal hazard index Hex, Hin in this study vary from 0.21 to 0.38 and 0.27 to 0.50 with an average value of 0.29 and 0.37, Respectively. The results will be discussed in light of various factors.Keywords: natural radioactivity, radium equivalent activity, absorbed dose rate, gamma ray spectroscopy
Procedia PDF Downloads 3623884 Automatic Content Curation of Visual Heritage
Authors: Delphine Ribes Lemay, Valentine Bernasconi, André Andrade, Lara DéFayes, Mathieu Salzmann, FréDéRic Kaplan, Nicolas Henchoz
Abstract:
Digitization and preservation of large heritage induce high maintenance costs to keep up with the technical standards and ensure sustainable access. Creating impactful usage is instrumental to justify the resources for long-term preservation. The Museum für Gestaltung of Zurich holds one of the biggest poster collections of the world from which 52’000 were digitised. In the process of building a digital installation to valorize the collection, one objective was to develop an algorithm capable of predicting the next poster to show according to the ones already displayed. The work presented here describes the steps to build an algorithm able to automatically create sequences of posters reflecting associations performed by curator and professional designers. The exposed challenge finds similarities with the domain of song playlist algorithms. Recently, artificial intelligence techniques and more specifically, deep-learning algorithms have been used to facilitate their generations. Promising results were found thanks to Recurrent Neural Networks (RNN) trained on manually generated playlist and paired with clusters of extracted features from songs. We used the same principles to create the proposed algorithm but applied to a challenging medium, posters. First, a convolutional autoencoder was trained to extract features of the posters. The 52’000 digital posters were used as a training set. Poster features were then clustered. Next, an RNN learned to predict the next cluster according to the previous ones. RNN training set was composed of poster sequences extracted from a collection of books from the Gestaltung Museum of Zurich dedicated to displaying posters. Finally, within the predicted cluster, the poster with the best proximity compared to the previous poster is selected. The mean square distance between features of posters was used to compute the proximity. To validate the predictive model, we compared sequences of 15 posters produced by our model to randomly and manually generated sequences. Manual sequences were created by a professional graphic designer. We asked 21 participants working as professional graphic designers to sort the sequences from the one with the strongest graphic line to the one with the weakest and to motivate their answer with a short description. The sequences produced by the designer were ranked first 60%, second 25% and third 15% of the time. The sequences produced by our predictive model were ranked first 25%, second 45% and third 30% of the time. The sequences produced randomly were ranked first 15%, second 29%, and third 55% of the time. Compared to designer sequences, and as reported by participants, model and random sequences lacked thematic continuity. According to the results, the proposed model is able to generate better poster sequencing compared to random sampling. Eventually, our algorithm is sometimes able to outperform a professional designer. As a next step, the proposed algorithm should include a possibility to create sequences according to a selected theme. To conclude, this work shows the potentiality of artificial intelligence techniques to learn from existing content and provide a tool to curate large sets of data, with a permanent renewal of the presented content.Keywords: Artificial Intelligence, Digital Humanities, serendipity, design research
Procedia PDF Downloads 1843883 Subjective Evaluation of Mathematical Morphology Edge Detection on Computed Tomography (CT) Images
Authors: Emhimed Saffor
Abstract:
In this paper, the problem of edge detection in digital images is considered. Three methods of edge detection based on mathematical morphology algorithm were applied on two sets (Brain and Chest) CT images. 3x3 filter for first method, 5x5 filter for second method and 7x7 filter for third method under MATLAB programming environment. The results of the above-mentioned methods are subjectively evaluated. The results show these methods are more efficient and satiable for medical images, and they can be used for different other applications.Keywords: CT images, Matlab, medical images, edge detection
Procedia PDF Downloads 3383882 Design and Development of an Optimal Fault Tolerant 3 Degree of Freedom Robotic Manipulator
Authors: Ramish, Farhan Khalique Awan
Abstract:
Kinematic redundancy within the manipulators presents extended dexterity and manipulability to the manipulators. Redundant serial robotic manipulators are very popular in industries due to its competencies to keep away from singularities during normal operation and fault tolerance because of failure of one or more joints. Such fault tolerant manipulators are extraordinarily beneficial in applications where human interference for repair and overhaul is both impossible or tough; like in case of robotic arms for space programs, nuclear applications and so on. The design of this sort of fault tolerant serial 3 DoF manipulator is presented in this paper. This work was the extension of the author’s previous work of designing the simple 3R serial manipulator. This work is the realization of the previous design with optimizing the link lengths for incorporating the feature of fault tolerance. Various measures have been followed by the researchers to quantify the fault tolerance of such redundant manipulators. The fault tolerance in this work has been described in terms of the worst-case measure of relative manipulability that is, in fact, a local measure of optimization that works properly for certain configuration of the manipulators. An optimum fault tolerant Jacobian matrix has been determined first based on prescribed null space properties after which the link parameters have been described to meet the given Jacobian matrix. A solid model of the manipulator was then developed to realize the mathematically rigorous design. Further work was executed on determining the dynamic properties of the fault tolerant design and simulations of the movement for various trajectories have been carried out to evaluate the joint torques. The mathematical model of the system was derived via the Euler-Lagrange approach after which the same has been tested using the RoboAnalyzer© software. The results have been quite in agreement. From the CAD model and dynamic simulation data, the manipulator was fabricated in the workshop and Advanced Machining lab of NED University of Engineering and Technology.Keywords: fault tolerant, Graham matrix, Jacobian, kinematics, Lagrange-Euler
Procedia PDF Downloads 2223881 Improving Preconception Health and Lifestyle Behaviours through Digital Health Intervention: The OptimalMe Program
Authors: Bonnie R. Brammall, Rhonda M. Garad, Helena J. Teede, Cheryce L. Harrison
Abstract:
Introduction: Reproductive aged women are at high-risk for accelerated weight gain and obesity development, with pregnancy recognised as a critical contributory life phase. Healthy lifestyle interventions during the preconception and antenatal period improve maternal and infant health outcomes. Yet, interventions from preconception through to postpartum and translation and implementation into real-world healthcare settings remain limited. OptimalMe is a randomised, hybrid implementation effectiveness study of evidence-based healthy lifestyle intervention. Here, we report engagement, acceptability of the intervention during preconception, and self-reported behaviour change outcomes as a result of the preconception phase of the intervention. Methods: Reproductive aged women who upgraded their private health insurance to include pregnancy and birth cover, signalling a pregnancy intention, were invited to participate. Women received access to an online portal with preconception health and lifestyle modules, goal-setting and behaviour change tools, monthly SMS messages, and two coaching sessions (randomised to video or phone) prior to pregnancy. Results: Overall n=527 expressed interest in participating. Of these, n=33 did not meet inclusion criteria, n=8 were not contactable for eligibility screening, and n=177 failed to engage after the screening, leaving n=309 who were enrolled in OptimalMe and randomised to intervention delivery method. Engagement with coaching sessions dropped by 25% for session two, with no difference between intervention groups. Women had a mean (SD) age of 31.7 (4.3) years and, at baseline, a self-reported mean BMI of 25.7 (6.1) kg/m², with 55.8% (n=172) of a healthy BMI. Behaviour was sub-optimal with infrequent self-weighing (38.1%), alcohol consumption prevalent (57.1%), sub-optimal pre-pregnancy supplementation (61.5%), and incomplete medical screening. Post-intervention 73.2% of women reported engagement with a GP for preconception care and improved lifestyle behaviour (85.5%), since starting OptimalMe. Direct pre-and-post comparison of individual participant data showed that of 322 points of potential change (up-to-date cervical screening, elimination of high-risk behaviours [alcohol, drugs, smoking], uptake of preconception supplements and improved weighing habits) 158 (49.1%) points of change were achieved. Health coaching sessions were found to improve accountability and confidence, yet further personalisation and support were desired. Engagement with video and phone sessions was comparable, having similar impacts on behaviour change, and both methods were well accepted and increased women's accountability. Conclusion: A low-intensity digital health and lifestyle program with embedded health coaching can improve the uptake of preconception care and lead to self-reported behaviour change. This is the first program of its kind to reach an otherwise healthy population of women planning a pregnancy. Women who were otherwise healthy showed divergence from preconception health and lifestyle objectives and benefited from the intervention. OptimalMe shows promising results for population-based behaviour change interventions that can improve preconception lifestyle habits and increase engagement with clinical health care for pregnancy preparation.Keywords: preconception, pregnancy, preventative health, weight gain prevention, self-management, behaviour change, digital health, telehealth, intervention, women's health
Procedia PDF Downloads 913880 Variational Explanation Generator: Generating Explanation for Natural Language Inference Using Variational Auto-Encoder
Authors: Zhen Cheng, Xinyu Dai, Shujian Huang, Jiajun Chen
Abstract:
Recently, explanatory natural language inference has attracted much attention for the interpretability of logic relationship prediction, which is also known as explanation generation for Natural Language Inference (NLI). Existing explanation generators based on discriminative Encoder-Decoder architecture have achieved noticeable results. However, we find that these discriminative generators usually generate explanations with correct evidence but incorrect logic semantic. It is due to that logic information is implicitly encoded in the premise-hypothesis pairs and difficult to model. Actually, logic information identically exists between premise-hypothesis pair and explanation. And it is easy to extract logic information that is explicitly contained in the target explanation. Hence we assume that there exists a latent space of logic information while generating explanations. Specifically, we propose a generative model called Variational Explanation Generator (VariationalEG) with a latent variable to model this space. Training with the guide of explicit logic information in target explanations, latent variable in VariationalEG could capture the implicit logic information in premise-hypothesis pairs effectively. Additionally, to tackle the problem of posterior collapse while training VariaztionalEG, we propose a simple yet effective approach called Logic Supervision on the latent variable to force it to encode logic information. Experiments on explanation generation benchmark—explanation-Stanford Natural Language Inference (e-SNLI) demonstrate that the proposed VariationalEG achieves significant improvement compared to previous studies and yields a state-of-the-art result. Furthermore, we perform the analysis of generated explanations to demonstrate the effect of the latent variable.Keywords: natural language inference, explanation generation, variational auto-encoder, generative model
Procedia PDF Downloads 1513879 Serious Digital Video Game for Solving Algebraic Equations
Authors: Liliana O. Martínez, Juan E González, Manuel Ramírez-Aranda, Ana Cervantes-Herrera
Abstract:
A serious game category mobile application called Math Dominoes is presented. The main objective of this applications is to strengthen the teaching-learning process of solving algebraic equations and is based on the board game "Double 6" dominoes. Math Dominoes allows the practice of solving first, second-, and third-degree algebraic equations. This application is aimed to students who seek to strengthen their skills in solving algebraic equations in a dynamic, interactive, and fun way, to reduce the risk of failure in subsequent courses that require mastery of this algebraic tool.Keywords: algebra, equations, dominoes, serious games
Procedia PDF Downloads 1303878 Women’s Colours in Digital Innovation
Authors: Daniel J. Patricio Jiménez
Abstract:
Digital reality demands new ways of thinking, flexibility in learning, acquisition of new competencies, visualizing reality under new approaches, generating open spaces, understanding dimensions in continuous change, etc. We need inclusive growth, where colors are not lacking, where lights do not give a distorted reality, where science is not half-truth. In carrying out this study, the documentary or bibliographic collection has been taken into account, providing a reflective and analytical analysis of current reality. In this context, deductive and inductive methods have been used on different multidisciplinary information sources. Women today and tomorrow are a strategic element in science and arts, which, under the umbrella of sustainability, implies ‘meeting current needs without detriment to future generations’. We must build new scenarios, which qualify ‘the feminine and the masculine’ as an inseparable whole, encouraging cooperative behavior; nothing is exclusive or excluding, and that is where true respect for diversity must be based. We are all part of an ecosystem, which we will make better as long as there is a real balance in terms of gender. It is the time of ‘the lifting of the veil’, in other words, it is the time to discover the pseudonyms, the women who painted, wrote, investigated, recorded advances, etc. However, the current reality demands much more; we must remove doors where they are not needed. Mass processing of data, big data, needs to incorporate algorithms under the perspective of ‘the feminine’. However, most STEM students (science, technology, engineering, and math) are men. Our way of doing science is biased, focused on honors and short-term results to the detriment of sustainability. Historically, the canons of beauty, the way of looking, of perceiving, of feeling, depended on the circumstances and interests of each moment, and women had no voice in this. Parallel to science, there is an under-representation of women in the arts, but not so much in the universities, but when we look at galleries, museums, art dealers, etc., colours impoverish the gaze and once again highlight the gender gap and the silence of the feminine. Art registers sensations by divining the future, science will turn them into reality. The uniqueness of the so-called new normality requires women to be protagonists both in new forms of emotion and thought, and in the experimentation and development of new models. This will result in women playing a decisive role in the so-called "5.0 society" or, in other words, in a more sustainable, more humane world.Keywords: art, digitalization, gender, science
Procedia PDF Downloads 1653877 Deorbiting Performance of Electrodynamic Tethers to Mitigate Space Debris
Authors: Giulia Sarego, Lorenzo Olivieri, Andrea Valmorbida, Carlo Bettanini, Giacomo Colombatti, Marco Pertile, Enrico C. Lorenzini
Abstract:
International guidelines recommend removing any artificial body in Low Earth Orbit (LEO) within 25 years from mission completion. Among disposal strategies, electrodynamic tethers appear to be a promising option for LEO, thanks to the limited storage mass and the minimum interface requirements to the host spacecraft. In particular, recent technological advances make it feasible to deorbit large objects with tether lengths of a few kilometers or less. To further investigate such an innovative passive system, the European Union is currently funding the project E.T.PACK – Electrodynamic Tether Technology for Passive Consumable-less Deorbit Kit in the framework of the H2020 Future Emerging Technologies (FET) Open program. The project focuses on the design of an end of life disposal kit for LEO satellites. This kit aims to deploy a taped tether that can be activated at the spacecraft end of life to perform autonomous deorbit within the international guidelines. In this paper, the orbital performance of the E.T.PACK deorbiting kit is compared to other disposal methods. Besides, the orbital decay prediction is parametrized as a function of spacecraft mass and tether system performance. Different values of length, width, and thickness of the tether will be evaluated for various scenarios (i.e., different initial orbital parameters). The results will be compared to other end-of-life disposal methods with similar allocated resources. The analysis of the more innovative system’s performance with the tape coated with a thermionic material, which has a low work-function (LWT), for which no active component for the cathode is required, will also be briefly discussed. The results show that the electrodynamic tether option can be a competitive and performant solution for satellite disposal compared to other deorbit technologies.Keywords: deorbiting performance, H2020, spacecraft disposal, space electrodynamic tethers
Procedia PDF Downloads 1773876 A Regulatory Analysis on Legal Problems of BitCoin
Authors: Fady Tawakol
Abstract:
BitCoin is a decentralized cryptocurrency that can be used without the need of traditional central banks to accomplish any e-commerce trade. The use of such currency could facilitate new economic interactions and linkages. However, without effective and efficient regulations, cryptocurrency transactions are mostly used by criminals to commit crimes such as money laundering, theft, and blackmailing. And because law is one step behind technological developments, this paper discusses the importance of regulations and supervision for the BitCoin-system, to provide unified regulatory solutions for our digital future in the Middle East. It will provide a detailed analysis of the legal nature of BitCoin along with, its regulation with respect to criminal and civil law.Keywords: BitCoin, financial protection, crypto currency, money laundering
Procedia PDF Downloads 2093875 The Impact of Information and Communication Technology in Education: Opportunities and Challenges
Authors: M. Nadeem, S. Nasir, K. A. Moazzam, R. Kashif
Abstract:
The remarkable growth and evolution in information and communication technology (ICT) in the past few decades has transformed modern society in almost every aspect of life. The impact and application of ICT have been observed in almost all walks of life including science, arts, business, health, management, engineering, sports, and education. ICT in education is being used extensively for student learning, creativity, interaction, and knowledge sharing and as a valuable source of teaching instrument. Apart from the student’s perspective, it plays a vital role for teacher education, instructional methods and curriculum development. There is a significant difference in growth of ICT enabled education in developing countries compared to developed nations and according to research, this gap is widening. ICT gradually infiltrate in almost every aspect of life. It has a deep and profound impact on our social, economic, health, environment, development, work, learning, and education environments. ICT provides very effective and dominant tools for information and knowledge processing. It is firmly believed that the coming generation should be proficient and confident in the use of ICT to cope with the existing international standards. This is only possible if schools can provide basic ICT infrastructure to students and to develop an ICT-integrated curriculum which covers all aspects of learning and creativity in students. However, there is a digital divide and steps must be taken to reduce this digital divide considerably to have the profound impact of ICT in education all around the globe. This study is based on theoretical approach and an extensive literature review is being conducted to see the successful implementations of ICT integration in education and to identify technologies and models which have been used in education in developed countries. This paper deals with the modern applications of ICT in schools for both teachers and students to uplift the learning and creativity amongst the students. A brief history of technology in education is presented and discussed are some important ICT tools for both student and teacher’s perspective. Basic ICT-based infrastructure for academic institutions is presented. The overall conclusion leads to the positive impact of ICT in education by providing an interactive, collaborative and challenging environment to students and teachers for knowledge sharing, learning and critical thinking.Keywords: information and communication technology, ICT, education, ICT infrastructure, learning
Procedia PDF Downloads 1243874 Digitial Communication – The Future of Chronic Disease Management Is Healthcare Apps
Authors: Kirstin Griffin
Abstract:
During a period of increased anxiety and stress, communication became the essential tool to help the public stay informed and feel prepared during the Covid-19 pandemic. However, certain groups of patients were not feeling as reassured. The news and media blasted the message that patients with diabetes were “high-risk" in regards to contracting the Covid-19 infection. Routine clinics were being cancelled, GP practices were closing their doors, and patients with type 1 diabetes were understandably scared. The influx of calls to diabetes specialists nurses from concerned patients highlighted the need for better and more specialised information. An Application specifically for patients with type 1 diabetes was created to deliver this information, and it proved to be the essential communication tool that was desperately needed. The Application for patients with type 1 diabetes aimed to deliver specialist information to patients in regards to their diagnosis, management, and ongoing follow-up commitments. The Application gives practical advice on multiple areas of diabetes management, including sick-day rules and diabetic emergencies, as well as up-to-date information on technology, including setting up Libre devices and downloading glucose meters to facilitate attending virtual clinics. Delivery of this information in an easy-to-understand and comprehensive way is intended to improve patient engagement with diabetes services and ultimately empower patients in the control of their own disease. The application also offers a messaging service to allow the diabetes team to send out alerts to patient groups on specific issues, such as changes to clinics, or respond to recent news updates regarding Covid-19. The App was launched in NHS Fife in June 2020 and has amassed 800 active users so far. There is growing engagement with the App since its launch, with over 1000 user interactions in the last month alone. Feedback shows that 100% of users like the App and have found it useful in the management of their diabetes. The App has proven to be an essential tool in communication with one of the most vulnerable groups during the Covid-19 pandemic, and its ongoing development will continue to increase patient engagement and improve glycaemic control for patients with type 1 diabetes. The future of chronic disease management should involve digital solutions such as apps to further empower patients in their healthcare.Keywords: diabetes, endocrinology, digital healthcare, medical apps
Procedia PDF Downloads 873873 Taking What Each Needs - The Basic Logic of Everyday Practice in State-backed Cultural Infrastructure in China
Authors: Yiling Shao, Megan Dai
Abstract:
This paper attempts to explore whether the cultural infrastructure supported by the Chinese government is still subject to a logic of “strict regulation”.Previous studies have pointed out that the "paternalism" tendency of China's cultural policy always leads to excessive government intervention in cultural development, while Chinese cultural practitioners can only seek cultural autonomy in the cracks of supervision. This can also explain why Chinese cultural policies sometimes have different effects than the official expectations.But this only reflects one aspect of China's cultural policy. In fact, the welfare cultural infrastructure funded by the government seems to highlight the principles of "safeguarding citizens' cultural rights" and "citizens' voluntary" rather than "indoctrination" and "enlightenment", What new features of China's cultural policy are reflected behind this policy orientation that is completely different from the logic of "regulation", which has also become an important issue in this paper. Based on the field survey of a cultural infrastructure (Gao ming District Cultural Center) in Gao ming District, Fo shan City, Guangdong Province, China, for nearly one year, the authors have obtained many text and picture materials.The paper discusses the dual role of cultural centers in China's cultural policy -both as a formal commitment by the state to protect citizens' basic cultural rights and as a social space for citizens to use preferential policies to obtain cultural capital. All in all, the author have conclued three operational logics of the cultural infrastructure currently supported by the Chinese government (at least in developed areas): first, the cultural center has become a versatile cultural space; second, grass-roots cultural cadres can be described as "policy entrepreneurs"; third, ordinary citizens will use the officially supported cultural infrastructure to increase cultural capital. This paper argues that, in comparison to the common “regulatory hand” in the field of cultural industries, in cultural infrastructure supported by state, the authorities and citizens are not in conflict. On the contrary, authorities must adopt a de-regulatory "pleasing" strategy to gain the support of citizens.Keywords: cultural infrastructure, cultural capital, deregulation, policy entrepreneur
Procedia PDF Downloads 973872 The Effect of Visual Access to Greenspace and Urban Space on a False Memory Learning Task
Authors: Bryony Pound
Abstract:
This study investigated how views of green or urban space affect learning performance. It provides evidence of the value of visual access to greenspace in work and learning environments, and builds on the extensive research into the cognitive and learning-related benefits of access to green and natural spaces, particularly in learning environments. It demonstrates that benefits of visual access to natural spaces whilst learning can produce statistically significant faster responses than those facing urban views after only 5 minutes. The primary hypothesis of this research was that a greenspace view would improve short-term learning. Participants were randomly assigned to either a view of parkland or of urban buildings from the same room. They completed a psychological test of two stages. The first stage consisted of a presentation of words from eight different categories (four manmade and four natural). Following this a 2.5 minute break was given; participants were not prompted to look out of the window, but all were observed doing so. The second stage of the test involved a word recognition/false memory test of three types. Type 1 was presented words from each category; Type 2 was non-presented words from those same categories; and Type 3 was non-presented words from different categories. Participants were asked to respond with whether they thought they had seen the words before or not. Accuracy of responses and reaction times were recorded. The key finding was that reaction times for Type 2 words (highest difficulty) were significantly different between urban and green view conditions. Those with an urban view had slower reaction times for these words, so a view of greenspace resulted in better information retrieval for word and false memory recognition. Importantly, this difference was found after only 5 minutes of exposure to either view, during winter, and with a sample size of only 26. Greenspace views improve performance in a learning task. This provides a case for better visual access to greenspace in work and learning environments.Keywords: benefits, greenspace, learning, restoration
Procedia PDF Downloads 1273871 Next-Generation Lunar and Martian Laser Retro-Reflectors
Authors: Simone Dell'Agnello
Abstract:
There are laser retroreflectors on the Moon and no laser retroreflectors on Mars. Here we describe the design, construction, qualification and imminent deployment of next-generation, optimized laser retroreflectors on the Moon and on Mars (where they will be the first ones). These instruments are positioned by time-of-flight measurements of short laser pulses, the so-called 'laser ranging' technique. Data analysis is carried out with PEP, the Planetary Ephemeris Program of CfA (Center for Astrophysics). Since 1969 Lunar Laser Ranging (LLR) to Apollo/Lunokhod laser retro-reflector (CCR) arrays supplied accurate tests of General Relativity (GR) and new gravitational physics: possible changes of the gravitational constant Gdot/G, weak and strong equivalence principle, gravitational self-energy (Parametrized Post Newtonian parameter beta), geodetic precession, inverse-square force-law; it can also constraint gravitomagnetism. Some of these measurements also allowed for testing extensions of GR, including spacetime torsion, non-minimally coupled gravity. LLR has also provides significant information on the composition of the deep interior of the Moon. In fact, LLR first provided evidence of the existence of a fluid component of the deep lunar interior. In 1969 CCR arrays contributed a negligible fraction of the LLR error budget. Since laser station range accuracy improved by more than a factor 100, now, because of lunar librations, current array dominate the error due to their multi-CCR geometry. We developed a next-generation, single, large CCR, MoonLIGHT (Moon Laser Instrumentation for General relativity high-accuracy test) unaffected by librations that supports an improvement of the space segment of the LLR accuracy up to a factor 100. INFN also developed INRRI (INstrument for landing-Roving laser Retro-reflector Investigations), a microreflector to be laser-ranged by orbiters. Their performance is characterized at the SCF_Lab (Satellite/lunar laser ranging Characterization Facilities Lab, INFN-LNF, Frascati, Italy) for their deployment on the lunar surface or the cislunar space. They will be used to accurately position landers, rovers, hoppers, orbiters of Google Lunar X Prize and space agency missions, thanks to LLR observations from station of the International Laser Ranging Service in the USA, in France and in Italy. INRRI was launched in 2016 with the ESA mission ExoMars (Exobiology on Mars) EDM (Entry, descent and landing Demonstration Module), deployed on the Schiaparelli lander and is proposed for the ExoMars 2020 Rover. Based on an agreement between NASA and ASI (Agenzia Spaziale Italiana), another microreflector, LaRRI (Laser Retro-Reflector for InSight), was delivered to JPL (Jet Propulsion Laboratory) and integrated on NASA’s InSight Mars Lander in August 2017 (launch scheduled in May 2018). Another microreflector, LaRA (Laser Retro-reflector Array) will be delivered to JPL for deployment on the NASA Mars 2020 Rover. The first lunar landing opportunities will be from early 2018 (with TeamIndus) to late 2018 with commercial missions, followed by opportunities with space agency missions, including the proposed deployment of MoonLIGHT and INRRI on NASA’s Resource Prospectors and its evolutions. In conclusion, we will extend significantly the CCR Lunar Geophysical Network and populate the Mars Geophysical Network. These networks will enable very significantly improved tests of GR.Keywords: general relativity, laser retroreflectors, lunar laser ranging, Mars geodesy
Procedia PDF Downloads 2703870 Delivery of Patient-Directed Wound Care Via Mobile Application-Based Qualitative Analysis
Authors: Amulya Srivatsa, Gayatri Prakash, Deeksha Sarda, Varshni Nandakumar, Duncan Salmon
Abstract:
Delivery of Patient-Directed Wound Care Via Mobile Application-Based Qualitative Analysis Chronic wounds are difficult for patients to manage at-home due to their unpredictable healing process. These wounds are associated with increased morbidity and negatively affect physical and mental health. The solution is a mobile application that will have an algorithm-based checklist to determine the state of the wound based on different factors that vary from person to person. Once this information is gathered, the application will recommend a plan of care to the user and subsequent steps to be taken. The mobile application will allow users to perform a digital scan of the wound to extract quantitative information regarding wound width, length, and depth, which will then be uploaded to the EHR to notify the patient’s provider. This scan utilizes a photo taken by the user, who is prompted appropriately. Furthermore, users will enter demographic information and answer multiple choice and drop-down menus describing the wound state. The proposed solution can save patients from unnecessary trips to the hospital for chronic wound care. The next iteration of the application can incorporate AI to allow users to perform a digital scan of the wound to extract quantitative information regarding wound width, length, and depth, which can be shared with the patient’s provider to allow for more efficient treatment. Ultimately, this product can provide immediate and economical medical advice for patients that suffer from chronic wounds. Research Objectives: The application should be capable of qualitative analysis of a wound and recommend a plan of care to the user. Additionally, the results of the wound analysis should automatically upload to the patient’s EMR. Research Methodologies: The app has two components: the first is a checklist with tabs for varying factors that assists users in the assessment of their skin. Subsequently, the algorithm will create an at-home regimen for patients to follow to manage their wounds. Research Contributions: The app aims to return autonomy back to the patient and reduce the number of visits to a physician for chronic wound care. The app also serves to educate the patient on how best to care for their wounds.Keywords: wound, app, qualitative, analysis, home, chronic
Procedia PDF Downloads 673869 Processing Big Data: An Approach Using Feature Selection
Authors: Nikat Parveen, M. Ananthi
Abstract:
Big data is one of the emerging technology, which collects the data from various sensors and those data will be used in many fields. Data retrieval is one of the major issue where there is a need to extract the exact data as per the need. In this paper, large amount of data set is processed by using the feature selection. Feature selection helps to choose the data which are actually needed to process and execute the task. The key value is the one which helps to point out exact data available in the storage space. Here the available data is streamed and R-Center is proposed to achieve this task.Keywords: big data, key value, feature selection, retrieval, performance
Procedia PDF Downloads 3413868 Application Research of Stilbene Crystal for the Measurement of Accelerator Neutron Sources
Authors: Zhao Kuo, Chen Liang, Zhang Zhongbing, Ruan Jinlu. He Shiyi, Xu Mengxuan
Abstract:
Stilbene, C₁₄H₁₂, is well known as one of the most useful organic scintillators for pulse shape discrimination (PSD) technique for its good scintillation properties. An on-line acquisition system and an off-line acquisition system were developed with several CAMAC standard plug-ins, NIM plug-ins, neutron/γ discriminating plug-in named 2160A and a digital oscilloscope with high sampling rate respectively for which stilbene crystals and photomultiplier tube detectors (PMT) as detector for accelerator neutron sources measurement carried out in China Institute of Atomic Energy. Pulse amplitude spectrums and charge amplitude spectrums were real-time recorded after good neutron/γ discrimination whose best PSD figure-of-merits (FoMs) are 1.756 for D-D accelerator neutron source and 1.393 for D-T accelerator neutron source. The probability of neutron events in total events was 80%, and neutron detection efficiency was 5.21% for D-D accelerator neutron sources, which were 50% and 1.44% for D-T accelerator neutron sources after subtracting the background of scattering observed by the on-line acquisition system. Pulse waveform signals were acquired by the off-line acquisition system randomly while the on-line acquisition system working. The PSD FoMs obtained by the off-line acquisition system were 2.158 for D-D accelerator neutron sources and 1.802 for D-T accelerator neutron sources after waveform digitization off-line processing named charge integration method for just 1000 pulses. In addition, the probabilities of neutron events in total events obtained by the off-line acquisition system matched very well with the probabilities of the on-line acquisition system. The pulse information recorded by the off-line acquisition system could be repetitively used to adjust the parameters or methods of PSD research and obtain neutron charge amplitude spectrums or pulse amplitude spectrums after digital analysis with a limited number of pulses. The off-line acquisition system showed equivalent or better measurement effects compared with the online system with a limited number of pulses which indicated a feasible method based on stilbene crystals detectors for the measurement of prompt neutrons neutron sources like prompt accelerator neutron sources emit a number of neutrons in a short time.Keywords: stilbene crystal, accelerator neutron source, neutron / γ discrimination, figure-of-merits, CAMAC, waveform digitization
Procedia PDF Downloads 1873867 Unleashing Potential in Pedagogical Innovation for STEM Education: Applying Knowledge Transfer Technology to Guide a Co-Creation Learning Mechanism for the Lingering Effects Amid COVID-19
Authors: Lan Cheng, Harry Qin, Yang Wang
Abstract:
Background: COVID-19 has induced the largest digital learning experiment in history. There is also emerging research evidence that students have paid a high cost of learning loss from virtual learning. University-wide survey results demonstrate that digital learning remains difficult for students who struggle with learning challenges, isolation, or a lack of resources. Large-scale efforts are therefore increasingly utilized for digital education. To better prepare students in higher education for this grand scientific and technological transformation, STEM education has been prioritized and promoted as a strategic imperative in the ongoing curriculum reform essential for unfinished learning needs and whole-person development. Building upon five key elements identified in the STEM education literature: Problem-based Learning, Community and Belonging, Technology Skills, Personalization of Learning, Connection to the External Community, this case study explores the potential of pedagogical innovation that integrates computational and experimental methodologies to support, enrich, and navigate STEM education. Objectives: The goal of this case study is to create a high-fidelity prototype design for STEM education with knowledge transfer technology that contains a Cooperative Multi-Agent System (CMAS), which has the objectives of (1) conduct assessment to reveal a virtual learning mechanism and establish strategies to facilitate scientific learning engagement, accessibility, and connection within and beyond university setting, (2) explore and validate an interactional co-creation approach embedded in project-based learning activities under the STEM learning context, which is being transformed by both digital technology and student behavior change,(3) formulate and implement the STEM-oriented campaign to guide learning network mapping, mitigate the loss of learning, enhance the learning experience, scale-up inclusive participation. Methods: This study applied a case study strategy and a methodology informed by Social Network Analysis Theory within a cross-disciplinary communication paradigm (students, peers, educators). Knowledge transfer technology is introduced to address learning challenges and to increase the efficiency of Reinforcement Learning (RL) algorithms. A co-creation learning framework was identified and investigated in a context-specific way with a learning analytic tool designed in this study. Findings: The result shows that (1) CMAS-empowered learning support reduced students’ confusion, difficulties, and gaps during problem-solving scenarios while increasing learner capacity empowerment, (2) The co-creation learning phenomenon have examined through the lens of the campaign and reveals that an interactive virtual learning environment fosters students to navigate scientific challenge independently and collaboratively, (3) The deliverables brought from the STEM educational campaign provide a methodological framework both within the context of the curriculum design and external community engagement application. Conclusion: This study brings a holistic and coherent pedagogy to cultivates students’ interest in STEM and develop them a knowledge base to integrate and apply knowledge across different STEM disciplines. Through the co-designing and cross-disciplinary educational content and campaign promotion, findings suggest factors to empower evidence-based learning practice while also piloting and tracking the impact of the scholastic value of co-creation under the dynamic learning environment. The data nested under the knowledge transfer technology situates learners’ scientific journey and could pave the way for theoretical advancement and broader scientific enervators within larger datasets, projects, and communities.Keywords: co-creation, cross-disciplinary, knowledge transfer, STEM education, social network analysis
Procedia PDF Downloads 1143866 Identification of Ideal Plain Sufu (Fermented Soybean Curds) Based on Ideal Profile Method and Assessment of the Consistency of Ideal Profiles Obtained from Consumers
Authors: Yan Ping Chen, Hau Yin Chung
Abstract:
The Ideal Profile Method (IPM) is a newly developed descriptive sensory analysis conducted by consumers without previous training. To perform this test, both the perceived and the ideal intensities from the judgements of consumers on products’ attributes, as well as their hedonic ratings were collected for formulating an ideal product (the most liked one). In addition, Ideal Profile Analysis (IPA) was conducted to check the consistency of the ideal data at both the panel and consumer levels. In this test, 12 commercial plain sufus bought from Hong Kong local market were tested by 113 consumers according to the IPM, and rated on 22 attributes. Principal component analysis was used to profile the perceived and the ideal spaces of tested products. The consistency of ideal data was then checked by IPA. The result showed that most consumers shared a common ideal. It was observed that the sensory product space and the ideal product space were structurally similar. Their first dimensions all opposed products with intense fermented related aroma to products with less fermented related aroma. And the predicted ideal profile (the estimated liking score around 7.0 in a 9.0-point scale) got higher hedonic score than the tested products (the average liking score around 6.0 in a 9.0-point scale). For the majority of consumers (95.2%), the stated ideal product considered as a potential ideal through checking the R2 coefficient value. Among all the tested products, sample-6 was the most popular one with consumer liking percentage around 30%. This product with less fermented and moldy flavour but easier to melt in mouth texture possessed close sensory profile according to the ideal product. This experiment validated that data from untrained consumers could be guided as useful information. Appreciated sensory characteristics could be served as reference in the optimization of the commercial plain sufu.Keywords: ideal profile method, product development, sensory evaluation, sufu (fermented soybean curd)
Procedia PDF Downloads 1883865 A Survey on Compression Methods for Table Constraints
Authors: N. Gharbi
Abstract:
Constraint Satisfaction problems are mathematical problems that are often used to model many real-world problems for which we look if there exists a solution satisfying all its constraints. Table constraints are important for modeling parts of many problems since they list all combinations of allowed or forbidden values. However, they admit practical limitations because they are sometimes too large to be represented in a direct way. In this paper, we present a survey of the different categories of the proposed approaches to compress table constraints in order to reduce both space and time complexities.Keywords: constraint programming, compression, data mining, table constraints
Procedia PDF Downloads 3253864 A Leader-Follower Kinematic-Based Control System for a Cable-Driven Hyper-Redundant Manipulator
Authors: Abolfazl Zaraki, Yoshikatsu Hayashi, Harry Thorpe, Vincent Strong, Gisle-Andre Larsen, William Holderbaum
Abstract:
Thanks to the high maneuverability of the cable-driven hyper-redundant manipulators (HRMs), this class of robots has shown a superior capability in highly confined and unstructured space applications. Although the large number of degrees of freedom (DOF) of HRMs enhances the motion flexibility and the robot’s reachability range, it highly increases the complexity of the kinematic configuration which makes the kinematic control problem very challenging or even impossible to solve. This paper presents our current progress achieved on the development of a kinematic-based leader-follower control system which is designed to control not only the robot’s body posture but also to control the trajectory of the robot’s movement in a semi-autonomous manner (the human operator is retained in the robot’s control loop). To obtain the forward kinematic model, the coordinate frames are established by the classical Denavit–Hartenburg (D-H) convention for a hyper-redundant serial manipulator which has a controlled cables-driven mechanism. To solve the inverse kinematics of the robot, unlike the conventional methods, a leader-follower mechanism, based on the sequential inverse kinematic, is followed. Using this mechanism, the inverse kinematic problem is solved for all sequential joints starting from the head joint to the base joint of the robot. To verify the kinematic design and simulate the robot motion, the MATLAB robotic toolbox is used. The simulation result demonstrated the promising capability of the proposed leader-follower control system in controlling the robot motion and trajectory in our confined space application.Keywords: hyper-redundant robots, kinematic analysis, semi-autonomous control, serial manipulators
Procedia PDF Downloads 1573863 An International Curriculum Development for Languages and Technology
Authors: Miguel Nino
Abstract:
When considering the challenges of a changing and demanding globalizing world, it is important to reflect on how university students will be prepared for the realities of internationalization, marketization and intercultural conversation. The present study is an interdisciplinary program designed to respond to the needs of the global community. The proposal bridges the humanities and science through three different fields: Languages, graphic design and computer science, specifically, fundamentals of programming such as python, java script and software animation. Therefore, the goal of the four year program is twofold: First, enable students for intercultural communication between English and other languages such as Spanish, Mandarin, French or German. Second, students will acquire knowledge in practical software and relevant employable skills to collaborate in assisted computer projects that most probable will require essential programing background in interpreted or compiled languages. In order to become inclusive and constructivist, the cognitive linguistics approach is suggested for the three different fields, particularly for languages that rely on the traditional method of repetition. This methodology will help students develop their creativity and encourage them to become independent problem solving individuals, as languages enhance their common ground of interaction for culture and technology. Participants in this course of study will be evaluated in their second language acquisition at the Intermediate-High level. For graphic design and computer science students will apply their creative digital skills, as well as their critical thinking skills learned from the cognitive linguistics approach, to collaborate on a group project design to find solutions for media web design problems or marketing experimentation for a company or the community. It is understood that it will be necessary to apply programming knowledge and skills to deliver the final product. In conclusion, the program equips students with linguistics knowledge and skills to be competent in intercultural communication, where English, the lingua franca, remains the medium for marketing and product delivery. In addition to their employability, students can expand their knowledge and skills in digital humanities, computational linguistics, or increase their portfolio in advertising and marketing. These students will be the global human capital for the competitive globalizing community.Keywords: curriculum, international, languages, technology
Procedia PDF Downloads 4433862 Comparison between RILM, JSTOR, and WorldCat Used to Search for Secondary Literature
Authors: Stacy Jarvis
Abstract:
Databases such as JSTOR, RILM and WorldCat have been the main source and storage of literature in the music orb. The Reference Index to Music Literature is a bibliographic database of over 2.6 million citations to writings about music from over 70 countries. The Research Institute produces RILM for the Study of Music at the University of Buffalo. JSTOR is an e-library of academic journals, books, and primary sources. Database JSTOR helps scholars find, utilise, and build upon a vast range of literature through a powerful teaching and research platform. Another database, WorldCat, is the world's biggest library catalogue, assisting scholars in finding library materials online. An evaluation of these databases in the music sphere is conducted by looking into the description and intended use and finding similarities and differences among them. Through comparison, it is found that these aim to serve different purposes, though they have the same goal of providing and storing literature. Also, since each database has different parts of literature that it majors on, the intended use of the three databases is evaluated. This can be found in the description, scope, and intended uses section. These areas are crucial to the research as it addresses the functional or literature differences among the three databases. It is also found that these databases have different quantitative potentials. This is determined by addressing the year each database began collecting literature and the number of articles, periodicals, albums, conference proceedings, music, dissertations, digital media, essays collections, journal articles, monographs, online resources, reviews, and reference materials that can be found in each one of them. This can be found in the sections- description, scope and intended uses and the importance of the database in identifying literature on different topics. To compare the delivery of services to the users, the importance of databases in identifying literature on different topics is also addressed in the section -the importance of databases in identifying literature on different topics. Even though these databases are used in research, they all have disadvantages and advantages. This is addressed in the sections on advantages and disadvantages. This will be significant in determining which of the three is the best. Also, it will help address how the shortcomings of one database can be addressed by utilising two databases together while conducting research. It is addressed in the section- a combination of RILM and JSTOR. All this information revolves around the idea that a huge amount of quantitative and qualitative data can be found in the presented databases on music and digital content; however, each of the given databases has a different construction and material features contributing to the musical scholarship in its way.Keywords: RILM, JSTOR, WorldCat, database, literature, research
Procedia PDF Downloads 833861 Establishment of Precision System for Underground Facilities Based on 3D Absolute Positioning Technology
Authors: Yonggu Jang, Jisong Ryu, Woosik Lee
Abstract:
The study aims to address the limitations of existing underground facility exploration equipment in terms of exploration depth range, relative depth measurement, data processing time, and human-centered ground penetrating radar image interpretation. The study proposed the use of 3D absolute positioning technology to develop a precision underground facility exploration system. The aim of this study is to establish a precise exploration system for underground facilities based on 3D absolute positioning technology, which can accurately survey up to a depth of 5m and measure the 3D absolute location of precise underground facilities. The study developed software and hardware technologies to build the precision exploration system. The software technologies developed include absolute positioning technology, ground surface location synchronization technology of GPR exploration equipment, GPR exploration image AI interpretation technology, and integrated underground space map-based composite data processing technology. The hardware systems developed include a vehicle-type exploration system and a cart-type exploration system. The data was collected using the developed exploration system, which employs 3D absolute positioning technology. The GPR exploration images were analyzed using AI technology, and the three-dimensional location information of the explored precise underground facilities was compared to the integrated underground space map. The study successfully developed a precision underground facility exploration system based on 3D absolute positioning technology. The developed exploration system can accurately survey up to a depth of 5m and measure the 3D absolute location of precise underground facilities. The system comprises software technologies that build a 3D precise DEM, synchronize the GPR sensor's ground surface 3D location coordinates, automatically analyze and detect underground facility information in GPR exploration images and improve accuracy through comparative analysis of the three-dimensional location information, and hardware systems, including a vehicle-type exploration system and a cart-type exploration system. The study's findings and technological advancements are essential for underground safety management in Korea. The proposed precision exploration system significantly contributes to establishing precise location information of underground facility information, which is crucial for underground safety management and improves the accuracy and efficiency of exploration. The study addressed the limitations of existing equipment in exploring underground facilities, proposed 3D absolute positioning technology-based precision exploration system, developed software and hardware systems for the exploration system, and contributed to underground safety management by providing precise location information. The developed precision underground facility exploration system based on 3D absolute positioning technology has the potential to provide accurate and efficient exploration of underground facilities up to a depth of 5m. The system's technological advancements contribute to the establishment of precise location information of underground facility information, which is essential for underground safety management in Korea.Keywords: 3D absolute positioning, AI interpretation of GPR exploration images, complex data processing, integrated underground space maps, precision exploration system for underground facilities
Procedia PDF Downloads 623860 A Generalized Framework for Adaptive Machine Learning Deployments in Algorithmic Trading
Authors: Robert Caulk
Abstract:
A generalized framework for adaptive machine learning deployments in algorithmic trading is introduced, tested, and released as open-source code. The presented software aims to test the hypothesis that recent data contains enough information to form a probabilistically favorable short-term price prediction. Further, the framework contains various adaptive machine learning techniques that are geared toward generating profit during strong trends and minimizing losses during trend changes. Results demonstrate that this adaptive machine learning approach is capable of capturing trends and generating profit. The presentation also discusses the importance of defining the parameter space associated with the dynamic training data-set and using the parameter space to identify and remove outliers from prediction data points. Meanwhile, the generalized architecture enables common users to exploit the powerful machinery while focusing on high-level feature engineering and model testing. The presentation also highlights common strengths and weaknesses associated with the presented technique and presents a broad range of well-tested starting points for feature set construction, target setting, and statistical methods for enforcing risk management and maintaining probabilistically favorable entry and exit points. The presentation also describes the end-to-end data processing tools associated with FreqAI, including automatic data fetching, data aggregation, feature engineering, safe and robust data pre-processing, outlier detection, custom machine learning and statistical tools, data post-processing, and adaptive training backtest emulation, and deployment of adaptive training in live environments. Finally, the generalized user interface is also discussed in the presentation. Feature engineering is simplified so that users can seed their feature sets with common indicator libraries (e.g. TA-lib, pandas-ta). The user also feeds data expansion parameters to fill out a large feature set for the model, which can contain as many as 10,000+ features. The presentation describes the various object-oriented programming techniques employed to make FreqAI agnostic to third-party libraries and external data sources. In other words, the back-end is constructed in such a way that users can leverage a broad range of common regression libraries (Catboost, LightGBM, Sklearn, etc) as well as common Neural Network libraries (TensorFlow, PyTorch) without worrying about the logistical complexities associated with data handling and API interactions. The presentation finishes by drawing conclusions about the most important parameters associated with a live deployment of the adaptive learning framework and provides the road map for future development in FreqAI.Keywords: machine learning, market trend detection, open-source, adaptive learning, parameter space exploration
Procedia PDF Downloads 893859 Navigating through Organizational Change: TAM-Based Manual for Digital Skills and Safety Transitions
Authors: Margarida Porfírio Tomás, Paula Pereira, José Palma Oliveira
Abstract:
Robotic grasping is advancing rapidly, but transferring techniques from rigid to deformable objects remains a challenge. Deformable and flexible items, such as food containers, demand nuanced handling due to their changing shapes. Bridging this gap is crucial for applications in food processing, surgical robotics, and household assistance. AGILEHAND, a Horizon project, focuses on developing advanced technologies for sorting, handling, and packaging soft and deformable products autonomously. These technologies serve as strategic tools to enhance flexibility, agility, and reconfigurability within the production and logistics systems of European manufacturing companies. Key components include intelligent detection, self-adaptive handling, efficient sorting, and agile, rapid reconfiguration. The overarching goal is to optimize work environments and equipment, ensuring both efficiency and safety. As new technologies emerge in the food industry, there will be some implications, such as labour force, safety problems and acceptance of the new technologies. To overcome these implications, AGILEHAND emphasizes the integration of social sciences and humanities, for example, the application of the Technology Acceptance Model (TAM). The project aims to create a change management manual, that will outline strategies for developing digital skills and managing health and safety transitions. It will also provide best practices and models for organizational change. Additionally, AGILEHAND will design effective training programs to enhance employee skills and knowledge. This information will be obtained through a combination of case studies, structured interviews, questionnaires, and a comprehensive literature review. The project will explore how organizations adapt during periods of change and identify factors influencing employee motivation and job satisfaction. This project received funding from European Union’s Horizon 2020/Horizon Europe research and innovation program under grant agreement No101092043 (AGILEHAND).Keywords: change management, technology acceptance model, organizational change, health and safety
Procedia PDF Downloads 453858 Informal Self-Governance: The Formation of an Alternative Urban Framework in a Cairo Region
Authors: Noor Abdelhamid
Abstract:
Almost half of Cairo’s growing population is housed in self-built, self-governed informal settlements serving as an alternative in the absence of government-provided public housing. These settlements emerged as the spatial expression of informal practices or activities operating outside regulated, formal frameworks. A comprehensive narrative of political events, administrative decisions, and urban policies set the stage for the growth of informal expression in Egypt. The purpose of this qualitative inquiry is to portray informal self-governance practiced by residents in the Cairo region. This research argues that informal spatial practices offer an alternative urban framework for bottom-up development in the absence of government provisions. In the context of this study, informal self-governance is defined as the residents’ autonomous control and use of public urban space in informal settlements. The case study for this research is Ard al-Liwa, a semi-formal settlement representing the majority of informal settlement typologies in Egypt, which consist of the formal occupation of land through an uncontrolled land subdivision, zoning, and construction. An inductive methodological approach is adopted to first study informal practices as singular activities and then as components of a larger environment. The collected set of empirical data consists of audiovisual material and observations obtained during regular site visits and interviews with residents native to the settlement. Methods of analysis are synthesized to identify themes in the data: the static and dynamic use of sidewalks, the urban traces of informal building allocation and construction, the de facto right to urban space, and the resultant spatial patterns. The paper concludes by positioning the research in the context of the current architectural practice, questioning the role, and responsibility, of designers in these self-governed urban regions.Keywords: Egypt, informal settlements, self-governance, urban framework
Procedia PDF Downloads 160