Search results for: edge application framework
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 13082

Search results for: edge application framework

872 Holographic Art as an Approach to Enhance Visual Communication in Egyptian Community: Experimental Study

Authors: Diaa Ahmed Mohamed Ahmedien

Abstract:

Nowadays, it cannot be denied that the most important interactive arts trends have appeared as a result of significant scientific mutations in the modern sciences, and holographic art is not an exception, where it is considered as a one of the most important major contemporary interactive arts trends in visual arts. Holographic technique had been evoked through the modern physics application in late 1940s, for the improvement of the quality of electron microscope images by Denis Gabor, until it had arrived to Margaret Benyon’s art exhibitions, and then it passed through a lot of procedures to enhance its quality and artistic applications technically and visually more over 70 years in visual arts. As a modest extension to these great efforts, this research aimed to invoke extraordinary attempt to enroll sample of normal people in Egyptian community in holographic recording program to record their appreciated objects or antiques, therefore examine their abilities to interact with modern techniques in visual communication arts. So this research tried to answer to main three questions: 'can we use the analog holographic techniques to unleash new theoretical and practical knowledge in interactive arts for public in Egyptian community?', 'to what extent holographic art can be familiar with public and make them able to produce interactive artistic samples?', 'are there possibilities to build holographic interactive program for normal people which lead them to enhance their understanding to visual communication in public and, be aware of interactive arts trends?' This research was depending in its first part on experimental methods, where it conducted in Laser lab at Cairo University, using Nd: Yag Laser 532 nm, and holographic optical layout, with selected samples of Egyptian people that they have been asked to record their appreciated object, after they had already learned recording methods, and in its second part on a lot of discussion panel had conducted to discuss the result and how participants felt towards their holographic artistic products through survey, questionnaires, take notes and critiquing holographic artworks. Our practical experiments and final discussions have already lead us to say that this experimental research was able to make most of participants pass through paradigm shift in their visual and conceptual experiences towards more interaction with contemporary visual arts trends, as an attempt to emphasize to the role of mature relationship between the art, science and technology, to spread interactive arts out in our community through the latest scientific and artistic mutations around the world and the role of this relationship in our societies particularly with those who have never been enrolled in practical arts programs before.

Keywords: Egyptian community, holographic art, laser art, visual art

Procedia PDF Downloads 468
871 Polyvinyl Alcohol Incorporated with Hibiscus Extract Microcapsules as Combined Active and Intelligent Composite Film for Meat Preservation: Antimicrobial, Antioxidant, and Physicochemical Investigations

Authors: Ahmed F. Ghanem, Marwa I. Wahba, Asmaa N. El-Dein, Mohamed A. EL-Raey, Ghada E. A. Awad

Abstract:

Numerous attempts are being performed in order to formulate suitable packaging materials for the meat products. However, to the best of our knowledge, the incorporation of the free hibiscus extract or its microcapsules in the pure polyvinyl alcohol (PVA) matrix as packaging materials for the meats is seldom reported. Therefore, this study aims at the protection of the aqueous crude extract of the hibiscus flowers utilizing the spry drying encapsulation technique. Results of the Fourier transform infrared (FTIR), the scanning electron microscope (SEM), and the particle size analyzer confirmed the successful formation of the assembled capsules via strong interactions, the spherical rough microparticles, and the particle size of ~ 235 nm, respectively. Also, the obtained microcapsules enjoy higher thermal stability than the free extract. Then, the obtained spray-dried particles were incorporated into the casting solution of the pure PVA film with a concentration of 10 wt. %. The segregated free-standing composite films were investigated, compared to the neat matrix, with several characterization techniques such as FTIR, SEM, thermal gravimetric analysis (TGA), mechanical tester, contact angle, water vapor permeability, and oxygen transmission. The results demonstrated variations in the physicochemical properties of the PVA film after the inclusion of the free and the extract microcapsules. Moreover, biological studies emphasized the biocidal potential of the hybrid films against the microorganisms contaminating the meat. Specifically, the microcapsules imparted not only antimicrobial but also antioxidant activities to the PVA matrix. Application of the prepared films on the real meat samples displayed a low bacterial growth with a slight increase in the pH over the storage time which continued up to 10 days at 4 oC, as further evidence to the meat safety. Moreover, the colors of the films did not significantly changed except after 21 days indicating the spoilage of the meat samples. No doubt, the dual-functional of the prepared composite films pave the way towards combined active and smart food packaging applications. This would play a vital role in the food hygiene, including also the quality control and the assurance.

Keywords: PVA, hibiscus, extraction, encapsulation, active packaging, smart and intelligent packaging, meat spoilage

Procedia PDF Downloads 77
870 Vapour Liquid Equilibrium Measurement of CO₂ Absorption in Aqueous 2-Aminoethylpiperazine (AEP)

Authors: Anirban Dey, Sukanta Kumar Dash, Bishnupada Mandal

Abstract:

Carbondioxide (CO2) is a major greenhouse gas responsible for global warming and fossil fuel power plants are the main emitting sources. Therefore the capture of CO2 is essential to maintain the emission levels according to the standards. Carbon capture and storage (CCS) is considered as an important option for stabilization of atmospheric greenhouse gases and minimizing global warming effects. There are three approaches towards CCS: Pre combustion capture where carbon is removed from the fuel prior to combustion, Oxy-fuel combustion, where coal is combusted with oxygen instead of air and Post combustion capture where the fossil fuel is combusted to produce energy and CO2 is removed from the flue gases left after the combustion process. Post combustion technology offers some advantage as existing combustion technologies can still be used without adopting major changes on them. A number of separation processes could be utilized part of post –combustion capture technology. These include (a) Physical absorption (b) Chemical absorption (c) Membrane separation (d) Adsorption. Chemical absorption is one of the most extensively used technologies for large scale CO2 capture systems. The industrially important solvents used are primary amines like Monoethanolamine (MEA) and Diglycolamine (DGA), secondary amines like diethanolamine (DEA) and Diisopropanolamine (DIPA) and tertiary amines like methyldiethanolamine (MDEA) and Triethanolamine (TEA). Primary and secondary amines react fast and directly with CO2 to form stable carbamates while Tertiary amines do not react directly with CO2 as in aqueous solution they catalyzes the hydrolysis of CO2 to form a bicarbonate ion and a protonated amine. Concentrated Piperazine (PZ) has been proposed as a better solvent as well as activator for CO2 capture from flue gas with a 10 % energy benefit compared to conventional amines such as MEA. However, the application of concentrated PZ is limited due to its low solubility in water at low temperature and lean CO2 loading. So following the performance of PZ its derivative 2-Aminoethyl piperazine (AEP) which is a cyclic amine can be explored as an activator towards the absorption of CO2. Vapour liquid equilibrium (VLE) in CO2 capture systems is an important factor for the design of separation equipment and gas treating processes. For proper thermodynamic modeling accurate equilibrium data for the solvent system over a wide range of temperatures, pressure and composition is essential. The present work focuses on the determination of VLE data for (AEP + H2O) system at 40 °C for various composition range.

Keywords: absorption, aminoethyl piperazine, carbondioxide, vapour liquid equilibrium

Procedia PDF Downloads 251
869 Bioanalytical Method Development and Validation of Aminophylline in Rat Plasma Using Reverse Phase High Performance Liquid Chromatography: An Application to Preclinical Pharmacokinetics

Authors: S. G. Vasantharaju, Viswanath Guptha, Raghavendra Shetty

Abstract:

Introduction: Aminophylline is a methylxanthine derivative belonging to the class bronchodilator. From the literature survey, reported methods reveals the solid phase extraction and liquid liquid extraction which is highly variable, time consuming, costly and laborious analysis. Present work aims to develop a simple, highly sensitive, precise and accurate high-performance liquid chromatography method for the quantification of Aminophylline in rat plasma samples which can be utilized for preclinical studies. Method: Reverse Phase high-performance liquid chromatography method. Results: Selectivity: Aminophylline and the internal standard were well separated from the co-eluted components and there was no interference from the endogenous material at the retention time of analyte and the internal standard. The LLOQ measurable with acceptable accuracy and precision for the analyte was 0.5 µg/mL. Linearity: The developed and validated method is linear over the range of 0.5-40.0 µg/mL. The coefficient of determination was found to be greater than 0.9967, indicating the linearity of this method. Accuracy and precision: The accuracy and precision values for intra and inter day studies at low, medium and high quality control samples concentrations of aminophylline in the plasma were within the acceptable limits Extraction recovery: The method produced consistent extraction recovery at all 3 QC levels. The mean extraction recovery of aminophylline was 93.57 ± 1.28% while that of internal standard was 90.70 ± 1.30%. Stability: The results show that aminophylline is stable in rat plasma under the studied stability conditions and that it is also stable for about 30 days when stored at -80˚C. Pharmacokinetic studies: The method was successfully applied to the quantitative estimation of aminophylline rat plasma following its oral administration to rats. Discussion: Preclinical studies require a rapid and sensitive method for estimating the drug concentration in the rat plasma. The method described in our article includes a simple protein precipitation extraction technique with ultraviolet detection for quantification. The present method is simple and robust for fast high-throughput sample analysis with less analysis cost for analyzing aminophylline in biological samples. In this proposed method, no interfering peaks were observed at the elution times of aminophylline and the internal standard. The method also had sufficient selectivity, specificity, precision and accuracy over the concentration range of 0.5 - 40.0 µg/mL. An isocratic separation technique was used underlining the simplicity of the presented method.

Keywords: Aminophyllin, preclinical pharmacokinetics, rat plasma, RPHPLC

Procedia PDF Downloads 208
868 Consolidated Predictive Model of the Natural History of Breast Cancer Considering Primary Tumor and Secondary Distant Metastases Growth

Authors: Ella Tyuryumina, Alexey Neznanov

Abstract:

This study is an attempt to obtain reliable data on the natural history of breast cancer growth. We analyze the opportunities for using classical mathematical models (exponential and logistic tumor growth models, Gompertz and von Bertalanffy tumor growth models) to try to describe growth of the primary tumor and the secondary distant metastases of human breast cancer. The research aim is to improve predicting accuracy of breast cancer progression using an original mathematical model referred to CoMPaS and corresponding software. We are interested in: 1) modelling the whole natural history of the primary tumor and the secondary distant metastases; 2) developing adequate and precise CoMPaS which reflects relations between the primary tumor and the secondary distant metastases; 3) analyzing the CoMPaS scope of application; 4) implementing the model as a software tool. The foundation of the CoMPaS is the exponential tumor growth model, which is described by determinate nonlinear and linear equations. The CoMPaS corresponds to TNM classification. It allows to calculate different growth periods of the primary tumor and the secondary distant metastases: 1) ‘non-visible period’ for the primary tumor; 2) ‘non-visible period’ for the secondary distant metastases; 3) ‘visible period’ for the secondary distant metastases. The CoMPaS is validated on clinical data of 10-years and 15-years survival depending on the tumor stage and diameter of the primary tumor. The new predictive tool: 1) is a solid foundation to develop future studies of breast cancer growth models; 2) does not require any expensive diagnostic tests; 3) is the first predictor which makes forecast using only current patient data, the others are based on the additional statistical data. The CoMPaS model and predictive software: a) fit to clinical trials data; b) detect different growth periods of the primary tumor and the secondary distant metastases; c) make forecast of the period of the secondary distant metastases appearance; d) have higher average prediction accuracy than the other tools; e) can improve forecasts on survival of breast cancer and facilitate optimization of diagnostic tests. The following are calculated by CoMPaS: the number of doublings for ‘non-visible’ and ‘visible’ growth period of the secondary distant metastases; tumor volume doubling time (days) for ‘non-visible’ and ‘visible’ growth period of the secondary distant metastases. The CoMPaS enables, for the first time, to predict ‘whole natural history’ of the primary tumor and the secondary distant metastases growth on each stage (pT1, pT2, pT3, pT4) relying only on the primary tumor sizes. Summarizing: a) CoMPaS describes correctly the primary tumor growth of IA, IIA, IIB, IIIB (T1-4N0M0) stages without metastases in lymph nodes (N0); b) facilitates the understanding of the appearance period and inception of the secondary distant metastases.

Keywords: breast cancer, exponential growth model, mathematical model, metastases in lymph nodes, primary tumor, survival

Procedia PDF Downloads 331
867 Educational Debriefing in Prehospital Medicine: A Qualitative Study Exploring Educational Debrief Facilitation and the Effects of Debriefing

Authors: Maria Ahmad, Michael Page, Danë Goodsman

Abstract:

‘Educational’ debriefing – a construct distinct from clinical debriefing – is used following simulated scenarios and is central to learning and development in fields ranging from aviation to emergency medicine. However, little research into educational debriefing in prehospital medicine exists. This qualitative study explored the facilitation and effects of prehospital educational debriefing and identified obstacles to debriefing, using the London’s Air Ambulance Pre-Hospital Care Course (PHCC) as a model. Method: Ethnographic observations of moulages and debriefs were conducted over two consecutive days of the PHCC in October 2019. Detailed contemporaneous field notes were made and analysed thematically. Subsequently, seven one-to-one, semi-structured interviews were conducted with four PHCC debrief facilitators and three course participants to explore their experiences of prehospital educational debriefing. Interview data were manually transcribed and analysed thematically. Results: Four overarching themes were identified: the approach to the facilitation of debriefs, effects of debriefing, facilitator development, and obstacles to debriefing. The unpredictable debriefing environment was seen as both hindering and paradoxically benefitting educational debriefing. Despite using varied debriefing structures, facilitators emphasised similar key debriefing components, including exploring participants’ reasoning and sharing experiences to improve learning and prevent future errors. Debriefing was associated with three principal effects: releasing emotion; learning and improving, particularly participant compound learning as they progressed through scenarios; and the application of learning to clinical practice. Facilitator training and feedback were central to facilitator learning and development. Several obstacles to debriefing were identified, including mismatch of participant and facilitator agendas, performance pressure, and time. Interestingly, when used appropriately in the educational environment, these obstacles may paradoxically enhance learning. Conclusions: Educational debriefing in prehospital medicine is complex. It requires the establishment of a safe learning environment, an understanding of participant agendas, and facilitator experience to maximise participant learning. Aspects unique to prehospital educational debriefing were identified, notably the unpredictable debriefing environment, interdisciplinary working, and the paradoxical benefit of educational obstacles for learning. This research also highlights aspects of educational debriefing not extensively detailed in the literature, such as compound participant learning, display of ‘professional honesty’ by facilitators, and facilitator learning, which require further exploration. Future research should also explore educational debriefing in other prehospital services.

Keywords: debriefing, prehospital medicine, prehospital medical education, pre-hospital care course

Procedia PDF Downloads 201
866 Mechanical Behavior of Sandwiches with Various Glass Fiber/Epoxy Skins under Bending Load

Authors: Emre Kara, Metehan Demir, Şura Karakuzu, Kadir Koç, Ahmet F. Geylan, Halil Aykul

Abstract:

While the polymeric foam cored sandwiches have been realized for many years, recently there is a growing and outstanding interest on the use of sandwiches consisting of aluminum foam core because of their some of the distinct mechanical properties such as high bending stiffness, high load carrying and energy absorption capacities. These properties make them very useful in the transportation industry (automotive, aerospace, shipbuilding industry), where the "lightweight design" philosophy and the safety of vehicles are very important aspects. Therefore, in this study, the sandwich panels with aluminum alloy foam core and various types and thicknesses of glass fiber reinforced polymer (GFRP) skins produced via Vacuum Assisted Resin Transfer Molding (VARTM) technique were obtained by using a commercial toughened epoxy based adhesive with two components. The aim of this contribution was the analysis of the bending response of sandwiches with various glass fiber reinforced polymer skins. The three point bending tests were performed on sandwich panels at different values of support span distance using a universal static testing machine in order to clarify the effects of the type and thickness of the GFRP skins in terms of peak load, energy efficiency and absorbed energy values. The GFRP skins were easily bonded to the aluminum alloy foam core under press machine with a very low pressure. The main results of the bending tests are: force-displacement curves, peak force values, absorbed energy, collapse mechanisms and the influence of the support span length and GFRP skins. The obtained results of the experimental investigation presented that the sandwich with the skin made of thicker S-Glass fabric failed at the highest load and absorbed the highest amount of energy compared to the other sandwich specimens. The increment of the support span distance made the decrease of the peak force and absorbed energy values for each type of panels. The common collapse mechanism of the panels was obtained as core shear failure which was not affected by the skin materials and the support span distance.

Keywords: aluminum foam, collapse mechanisms, light-weight structures, transport application

Procedia PDF Downloads 386
865 A Bottleneck-Aware Power Management Scheme in Heterogeneous Processors for Web Apps

Authors: Inyoung Park, Youngjoo Woo, Euiseong Seo

Abstract:

With the advent of WebGL, Web apps are now able to provide high quality graphics by utilizing the underlying graphic processing units (GPUs). Despite that the Web apps are becoming common and popular, the current power management schemes, which were devised for the conventional native applications, are suboptimal for Web apps because of the additional layer, the Web browser, between OS and application. The Web browser running on a CPU issues GL commands, which are for rendering images to be displayed by the Web app currently running, to the GPU and the GPU processes them. The size and number of issued GL commands determine the processing load of the GPU. While the GPU is processing the GL commands, CPU simultaneously executes the other compute intensive threads. The actual user experience will be determined by either CPU processing or GPU processing depending on which of the two is the more demanded resource. For example, when the GPU work queue is saturated by the outstanding commands, lowering the performance level of the CPU does not affect the user experience because it is already deteriorated by the retarded execution of GPU commands. Consequently, it would be desirable to lower CPU or GPU performance level to save energy when the other resource is saturated and becomes a bottleneck in the execution flow. Based on this observation, we propose a power management scheme that is specialized for the Web app runtime environment. This approach incurs two technical challenges; identification of the bottleneck resource and determination of the appropriate performance level for unsaturated resource. The proposed power management scheme uses the CPU utilization level of the Window Manager to tell which one is the bottleneck if exists. The Window Manager draws the final screen using the processed results delivered from the GPU. Thus, the Window Manager is on the critical path that determines the quality of user experience and purely executed by the CPU. The proposed scheme uses the weighted average of the Window Manager utilization to prevent excessive sensitivity and fluctuation. We classified Web apps into three categories using the analysis results that measure frame-per-second (FPS) changes under diverse CPU/GPU clock combinations. The results showed that the capability of the CPU decides user experience when the Window Manager utilization is above 90% and consequently, the proposed scheme decreases the performance level of CPU by one step. On the contrary, when its utilization is less than 60%, the bottleneck usually lies in the GPU and it is desirable to decrease the performance of GPU. Even the processing unit that is not on critical path, excessive performance drop can occur and that may adversely affect the user experience. Therefore, our scheme lowers the frequency gradually, until it finds an appropriate level by periodically checking the CPU utilization. The proposed scheme reduced the energy consumption by 10.34% on average in comparison to the conventional Linux kernel, and it worsened their FPS by 1.07% only on average.

Keywords: interactive applications, power management, QoS, Web apps, WebGL

Procedia PDF Downloads 182
864 Identification of Tangible and Intangible Heritage and Preparation of Conservation Proposal for the Historic City of Karanja Laad

Authors: Prachi Buche Marathe

Abstract:

Karanja Laad is a city located in the Vidarbha region in the state of Maharashtra, India. It has a huge amount of tangible and intangible heritage in the form of monuments, precincts, a group of structures, festivals and procession route, which is neglected and lost with time. Three different religions Hinduism, Islam and Jainism along with associations of being a birthplace of Swami Nrusinha Saraswati, an exponent of Datta Sampradaya sect and the British colonial layer have shaped the culture and society of the place over the period. The architecture of the town Karanja Laad has enhanced its unique historic and cultural value with a combination of all these historic layers. Karanja Laad is also a traditional trading historic town with unique hybrid architectural style and has a good potential for developing as a tourist place along with the present image of a pilgrim destination of Datta Sampradaya. The aim of the research is to prepare a conservation proposal for the historic town along with the management framework. Objectives of the research are to study the evolution of Karanja town, to identify the cultural resources along with issues of the historic core of the city, to understand Datta sampradaya, and contribution of Saint Nrusinha Saraswati in the religious sect and his association as an important personality with Karanja. The methodology of the research is site visits to the Karanja city, making field surveys for documentation and discussions and questionnaires with the residents to establish heritage and identify potential and issues within the historic core thereby establishing a case for conservation. Field surveys are conducted for town level study of land use, open spaces, occupancy, ownership, traditional commodity and community, infrastructure, streetscapes, and precinct activities during the festival and non-festival period. Building level study includes establishing various typologies like residential, institutional commercial, religious, and traditional infrastructure from the mythological references like waterbodies (kund), lake and wells. One of the main issues is that the loss of the traditional footprint as well as the traditional open spaces which are getting lost due to the new illegal encroachments and lack of guidelines for the new additions to conserve the original fabric of the structures. Traditional commodities are getting lost since there is no promotion of these skills like pottery and painting. Lavish bungalows like Kannava mansion, main temple Wada (birthplace of the saint) have a huge potential to be developed as a museum by adaptive re-use which will, in turn, attract many visitors during festivals which will boost the economy. Festival procession routes can be identified and a heritage walk can be developed so as to highlight the traditional features of the town. Overall study has resulted in establishing a heritage map with 137 heritage structures identified as potential. Conservation proposal is worked out on the town level, precinct level and building level with interventions such as developing construction guidelines for further development and establishing a heritage cell consisting architects and engineers for the upliftment of the existing rich heritage of the Karanja city.

Keywords: built heritage, conservation, Datta Sampradaya, Karanja Laad, Swami Nrusinha Saraswati, procession route

Procedia PDF Downloads 150
863 Barriers and Enablers to Climate and Health Adaptation Planning in Small Urban Areas in the Great Lakes Region

Authors: Elena Cangelosi, Wayne Beyea

Abstract:

This research expands the resilience planning literature by exploring the barriers and enablers to climate and health adaptation planning for small urban, coastal Great Lakes communities. With funding from the United States Centers for Disease Control and Prevention (CDC) Climate Ready City and States Initiative, this research took place during a 3-year pilot intervention project which integrates urban planning and public health. The project used the CDC’s Building Resilience Against Climate Effects (BRACE) framework to prevent or reduce the human health impacts from climate change in Marquette County, Michigan. Using a deliberation with the analysis planning process, interviews, focus groups, and community meetings with over 25 stakeholder groups and over 100 participants identified the area’s climate-related health concerns and adaptation interventions to address those concerns. Marquette County, on the shores of Lake Superior, the largest of the Great Lakes, was selected for the project based on their existing adaptive capacity and proactive approach to climate adaptation planning. With Marquette County as the context, this study fills a gap in the adaptation literature, which currently heavily emphasizes large-urban or agriculturally-based rural areas, and largely neglects small urban areas. This research builds on the qualitative case-study, survey, and interview approach established by previous researchers on contextual barriers and enablers for adaptation planning. This research uses a case study approach, including surveys and interviews of public officials, to identify the barriers and enablers for climate and health adaptation planning for small-urban areas within a large, non-agricultural, Great Lakes county. The researchers hypothesize that the barriers and enablers will, in some cases, overlap those found in other contexts, but in many cases, will be unique to a rural setting. The study reveals that funding, staff capacity, and communication across a large, rural geography act as the main barriers, while strong networks and collaboration, interested leaders, and community interest through a strong human-land connection act as the primary enablers. Challenges unique to rural areas are revealed, including weak opportunities for grant funding, large geographical distances, communication challenges with an aging and remote population, and the out-migration of education residents. Enablers that may be unique to rural contexts include strong collaborative relationships across jurisdictions for regional work and strong connections between residents and the land. As the factors that enable and prevent climate change planning are highly contextual, understanding, and appropriately addressing the unique factors at play for small-urban communities is key for effective planning in those areas. By identifying and addressing the barriers and enablers to climate and health adaptation planning for small-urban, coastal areas, this study can help Great Lakes communities appropriately build resilience to the adverse impacts of climate change. In addition, this research expands the breadth of research and understanding of the challenges and opportunities planners confront in the face of climate change.

Keywords: climate adaptation and resilience, climate change adaptation, climate change and urban resilience, governance and urban resilience

Procedia PDF Downloads 107
862 The Healing 'Touch' of Music: A Neuro-Acoustics Approach to Understand Its Therapeutic Effect

Authors: Jagmeet S. Kanwal, Julia F. Langley

Abstract:

Music can heal the body, but a mechanistic understanding of this phenomenon is lacking. This study explores the effects of music presentation on neurologic and physiologic responses leading to metabolic changes in the human body. The mind and body co-exist in a corporeal entity and within this framework, sickness ensues when the mind-body balance goes awry. It is further hypothesized that music has the capacity to directly reset this balance. Two lines of inquiry taken together can provide a mechanistic understanding of this phenomenon 1) Empirical evidence for a sound-sensitive pressure sensor system in the body, and 2) The notion of a “healing center” within the brain that is activated by specific patterns of sounds. From an acoustics perspective, music is spatially distributed as pressure waves ranging from a few cm to several meters in wavelength. These waves interact and propagate in three-dimensions in unique ways, depending on the wavelength. Furthermore, music creates dynamically changing wave-fronts. Frequencies between 200 Hz and 1 kHz generate wavelengths that range from 5'6" to 1 foot. These dimensions are in the range of the body size of most people making it plausible that these pressure waves can geometrically interact with the body surface and create distinct patterns of pressure stimulation across the skin surface. For humans, short wavelength, high frequency (> 200 Hz) sounds are best received via cochlear receptors. For low frequency (< 200 Hz), long wavelength sound vibrations, however, the whole body may act as an ideal receiver. A vast array of highly sensitive pressure receptors (Pacinian corpuscles) is present just beneath the skin surface, as well as in the tendons, bones, several organs in the abdomen, and the sexual organs. Per the available empirical evidence, these receptors contribute to music perception by allowing the whole body to function as a sound receiver, and knowledge of how they function is essential to fully understanding the therapeutic effect of music. Neuroscientific studies have established that music stimulates the limbic system that can trigger states of anxiety, arousal, fear, and other emotions. These emotional states of brain activity play a crucial role in filtering top-down feedback from thoughts and bottom-up sensory inputs to the autonomic system, which automatically regulates bodily functions. Music likely exerts its pleasurable and healing effects by enhancing functional and effective connectivity and feedback mechanisms between brain regions that mediate reward, autonomic, and cognitive processing. Stimulation of pressure receptors under the skin by low-frequency music-induced sensations can activate multiple centers in the brain, including the amygdala, the cingulate cortex, and nucleus accumbens. Melodies in music in the low (< 600 Hz) frequency range may augment auditory inputs after convergence of the pressure-sensitive inputs from the vagus nerve onto emotive processing regions within the limbic system. The integration of music-generated auditory and somato-visceral inputs may lead to a synergistic input to the brain that promotes healing. Thus, music can literally heal humans through “touch” as it energizes the brain’s autonomic system for restoring homeostasis.

Keywords: acoustics, brain, music healing, pressure receptors

Procedia PDF Downloads 150
861 Urban Dynamics Modelling of Mixed Land Use for Sustainable Urban Development in Indian Context

Authors: Rewati Raman, Uttam K. Roy

Abstract:

One of the main adversaries of city planning in present times is the ever-expanding problem of urbanization and the antagonistic issues accompanying it. The prevalent challenges in urbanization such as population growth, urban sprawl, poverty, inequality, pollution, congestion, etc. call for reforms in the urban fabric as well as in planning theory and practice. One of the various paradigms of city planning, land use planning, has been the major instruments for spatial planning of cities and regions in India. Zoning regulation based land use planning in the form of land use and development control plans (LUDCP) and development control regulations (DCR) have been considered mainstream guiding principles in land use planning for decades. In spite of many advantages of such zoning based regulations, over a period of time, it has been critiqued by scholars for its own limitations of isolation and lack of vitality, inconvenience in business in terms of proximity to residence and low operating cost, unsuitable environment for small investments, higher travel distance for facilities, amenities and thereby higher expenditure, safety issues etc. Mixed land use has been advocated as a tool to avoid such limitations in city planning by researchers. In addition, mixed land use can offer many advantages like housing variety and density, the creation of an economic blend of compatible land use, compact development, stronger neighborhood character, walkability, and generation of jobs, etc. Alternatively, the mixed land use beyond a suitable balance of use can also bring disadvantages like traffic congestion, encroachments, very high-density housing leading to a slum like condition, parking spill out, non-residential uses operating on residential premises paying less tax, chaos hampering residential privacy, pressure on existing infrastructure facilities, etc. This research aims at studying and outlining the various challenges and potentials of mixed land use zoning, through modeling tools, as a competent instrument for city planning in lieu of the present urban scenario. The methodology of research adopted in this paper involves the study of a mixed land use neighborhood in India, identification of indicators and parameters related to its extent and spatial pattern and the subsequent use of system dynamics as a modeling tool for simulation. The findings from this analysis helped in identifying the various advantages and challenges associated with the dynamic nature of a mixed use urban settlement. The results also confirmed the hypothesis that mixed use neighborhoods are catalysts for employment generation, socioeconomic gains while improving vibrancy, health, safety, and security. It is also seen that certain challenges related to chaos, lack of privacy and pollution prevail in mixed use neighborhoods, which can be mitigated by varying the percentage of mixing as per need, ensuring compatibility of adjoining use, institutional interventions in the form of policies, neighborhood micro-climatic interventions, etc. Therefore this paper gives a consolidated and holistic framework and quantified outcome pertaining to the extent and spatial pattern of mixed land use that should be adopted to ensure sustainable urban planning.

Keywords: mixed land use, sustainable development, system dynamics analysis, urban dynamics modelling

Procedia PDF Downloads 168
860 Sustainability in the Purchase of Airline Tickets: Analysis of Digital Communication from the Perspective of Neuroscience

Authors: Rodríguez Sánchez Carla, Sancho-Esper Franco, Guillen-Davo Marina

Abstract:

Tourism is one of the most important sectors worldwide since it is an important economic engine for today's society. It is also one of the sectors that most negatively affect the environment in terms of CO₂ emissions due to this expansion. In light of this, airlines are developing Voluntary Carbon Offset (VCO). There is important evidence focused on analyzing the features of these VCO programs and their efficacy in reducing CO₂ emissions, and findings are mixed without a clear consensus. Different research approaches have centered on analyzing factors and consequences of VCO programs, such as economic modelling based on panel data, survey research based on traveler responses or experimental research analyzing customer decisions in a simulated context. This study belongs to the latter group because it tries to understand how different characteristics of an online ticket purchase website affect the willingness of a traveler to choose a sustainable one. The proposed behavioral model is based on several theories, such as the nudge theory, the dual processing ELM and the cognitive dissonance theory. This randomized experiment aims at overcoming previous studies based on self-reported measures that mainly study sustainable behavioral intention rather than actual decision-making. It also complements traditional self-reported independent variables by gathering objective information from an eye-tracking device. This experiment analyzes the influence of two characteristics of the online purchase website: i) the type of information regarding flight CO₂ emissions (quantitative vs. qualitative) and the comparison framework related to the sustainable purchase decision (negative: alternative with more emissions than the average flight of the route vs. positive: alternative with less emissions than the average flight of the route), therefore it is a 2x2 experiment with four alternative scenarios. A pretest was run before the actual experiment to refine the experiment features and to check the manipulations. Afterward, a different sample of students answered the pre-test questionnaire aimed at recruiting the cases and measuring several pre-stimulus measures. One week later, students came to the neurolab at the University setting to be part of the experiment, made their decision regarding online purchases and answered the post-test survey. A final sample of 21 students was gathered. The committee of ethics of the institution approved the experiment. The results show that qualitative information generates more sustainable decisions (less contaminant alternative) than quantitative information. Moreover, evidence shows that subjects are more willing to choose the sustainable decision to be more ecological (comparison of the average with the less contaminant alternative) rather than to be less contaminant (comparison of the average with the more contaminant alternative). There are also interesting differences in the information processing variables from the eye tracker. Both the total time to make the choice and the specific times by area of interest (AOI) differ depending on the assigned scenario. These results allow for a better understanding of the factors that condition the decision of a traveler to be part of a VCO program and provide useful information for airline managers to promote these programs to reduce environmental impact.

Keywords: voluntary carbon offset, airline, online purchase, carbon emission, sustainability, randomized experiment

Procedia PDF Downloads 56
859 Evaluation of an Integrated Supersonic System for Inertial Extraction of CO₂ in Post-Combustion Streams of Fossil Fuel Operating Power Plants

Authors: Zarina Chokparova, Ighor Uzhinsky

Abstract:

Carbon dioxide emissions resulting from burning of the fossil fuels on large scales, such as oil industry or power plants, leads to a plenty of severe implications including global temperature raise, air pollution and other adverse impacts on the environment. Besides some precarious and costly ways for the alleviation of CO₂ emissions detriment in industrial scales (such as liquefaction of CO₂ and its deep-water treatment, application of adsorbents and membranes, which require careful consideration of drawback effects and their mitigation), one physically and commercially available technology for its capture and disposal is supersonic system for inertial extraction of CO₂ in after-combustion streams. Due to the flue gas with a carbon dioxide concentration of 10-15 volume percent being emitted from the combustion system, the waste stream represents a rather diluted condition at low pressure. The supersonic system induces a flue gas mixture stream to expand using a converge-and-diverge operating nozzle; the flow velocity increases to the supersonic ranges resulting in rapid drop of temperature and pressure. Thus, conversion of potential energy into the kinetic power causes a desublimation of CO₂. Solidified carbon dioxide can be sent to the separate vessel for further disposal. The major advantages of the current solution are its economic efficiency, physical stability, and compactness of the system, as well as needlessness of addition any chemical media. However, there are several challenges yet to be regarded to optimize the system: the way for increasing the size of separated CO₂ particles (as they are represented on a micrometers scale of effective diameter), reduction of the concomitant gas separated together with carbon dioxide and provision of CO₂ downstream flow purity. Moreover, determination of thermodynamic conditions of the vapor-solid mixture including specification of the valid and accurate equation of state remains to be an essential goal. Due to high speeds and temperatures reached during the process, the influence of the emitted heat should be considered, and the applicable solution model for the compressible flow need to be determined. In this report, a brief overview of the current technology status will be presented and a program for further evaluation of this approach is going to be proposed.

Keywords: CO₂ sequestration, converging diverging nozzle, fossil fuel power plant emissions, inertial CO₂ extraction, supersonic post-combustion carbon dioxide capture

Procedia PDF Downloads 132
858 Green Synthesis of Nanosilver-Loaded Hydrogel Nanocomposites for Antibacterial Application

Authors: D. Berdous, H. Ferfera-Harrar

Abstract:

Superabsorbent polymers (SAPs) or hydrogels with three-dimensional hydrophilic network structure are high-performance water absorbent and retention materials. The in situ synthesis of metal nanoparticles within polymeric network as antibacterial agents for bio-applications is an approach that takes advantage of the existing free-space into networks, which not only acts as a template for nucleation of nanoparticles, but also provides long term stability and reduces their toxicity by delaying their oxidation and release. In this work, SAP/nanosilver nanocomposites were successfully developed by a unique green process at room temperature, which involves in situ formation of silver nanoparticles (AgNPs) within hydrogels as a template. The aim of this study is to investigate whether these AgNPs-loaded hydrogels are potential candidates for antimicrobial applications. Firstly, the superabsorbents were prepared through radical copolymerization via grafting and crosslinking of acrylamide (AAm) onto chitosan backbone (Cs) using potassium persulfate as initiator and N,N’-methylenebisacrylamide as the crosslinker. Then, they were hydrolyzed to achieve superabsorbents with ampholytic properties and uppermost swelling capacity. Lastly, the AgNPs were biosynthesized and entrapped into hydrogels through a simple, eco-friendly and cost-effective method using aqueous silver nitrate as a silver precursor and curcuma longa tuber-powder extracts as both reducing and stabilizing agent. The formed superabsorbents nanocomposites (Cs-g-PAAm)/AgNPs were characterized by X-ray Diffraction (XRD), UV-visible Spectroscopy, Attenuated Total reflectance Fourier Transform Infrared Spectroscopy (ATR-FTIR), Inductively Coupled Plasma (ICP), and Thermogravimetric Analysis (TGA). Microscopic surface structure analyzed by Transmission Electron Microscopy (TEM) has showed spherical shapes of AgNPs with size in the range of 3-15 nm. The extent of nanosilver loading was decreased by increasing Cs content into network. The silver-loaded hydrogel was thermally more stable than the unloaded dry hydrogel counterpart. The swelling equilibrium degree (Q) and centrifuge retention capacity (CRC) in deionized water were affected by both contents of Cs and the entrapped AgNPs. The nanosilver-embedded hydrogels exhibited antibacterial activity against Escherichia coli and Staphylococcus aureus bacteria. These comprehensive results suggest that the elaborated AgNPs-loaded nanomaterials could be used to produce valuable wound dressing.

Keywords: antibacterial activity, nanocomposites, silver nanoparticles, superabsorbent Hydrogel

Procedia PDF Downloads 226
857 Mapping Context, Roles, and Relations for Adjudicating Robot Ethics

Authors: Adam J. Bowen

Abstract:

Abstract— Should robots have rights or legal protections. Often debates concerning whether robots and AI should be afforded rights focus on conditions of personhood and the possibility of future advanced forms of AI satisfying particular intrinsic cognitive and moral attributes of rights-holding persons. Such discussions raise compelling questions about machine consciousness, autonomy, and value alignment with human interests. Although these are important theoretical concerns, especially from a future design perspective, they provide limited guidance for addressing the moral and legal standing of current and near-term AI that operate well below the cognitive and moral agency of human persons. Robots and AI are already being pressed into service in a wide range of roles, especially in healthcare and biomedical contexts. The design and large-scale implementation of robots in the context of core societal institutions like healthcare systems continues to rapidly develop. For example, we bring them into our homes, hospitals, and other care facilities to assist in care for the sick, disabled, elderly, children, or otherwise vulnerable persons. We enlist surgical robotic systems in precision tasks, albeit still human-in-the-loop technology controlled by surgeons. We also entrust them with social roles involving companionship and even assisting in intimate caregiving tasks (e.g., bathing, feeding, turning, medicine administration, monitoring, transporting). There have been advances to enable severely disabled persons to use robots to feed themselves or pilot robot avatars to work in service industries. As the applications for near-term AI increase and the roles of robots in restructuring our biomedical practices expand, we face pressing questions about the normative implications of human-robot interactions and collaborations in our collective worldmaking, as well as the moral and legal status of robots. This paper argues that robots operating in public and private spaces be afforded some protections as either moral patients or legal agents to establish prohibitions on robot abuse, misuse, and mistreatment. We already implement robots and embed them in our practices and institutions, which generates a host of human-to-machine and machine-to-machine relationships. As we interact with machines, whether in service contexts, medical assistance, or home health companions, these robots are first encountered in relationship to us and our respective roles in the encounter (e.g., surgeon, physical or occupational therapist, recipient of care, patient’s family, healthcare professional, stakeholder). This proposal aims to outline a framework for establishing limiting factors and determining the extent of moral or legal protections for robots. In doing so, it advocates for a relational approach that emphasizes the priority of mapping the complex contextually sensitive roles played and the relations in which humans and robots stand to guide policy determinations by relevant institutions and authorities. The relational approach must also be technically informed by the intended uses of the biomedical technologies in question, Design History Files, extensive risk assessments and hazard analyses, as well as use case social impact assessments.

Keywords: biomedical robots, robot ethics, robot laws, human-robot interaction

Procedia PDF Downloads 97
856 Psychophysiological Adaptive Automation Based on Fuzzy Controller

Authors: Liliana Villavicencio, Yohn Garcia, Pallavi Singh, Luis Fernando Cruz, Wilfrido Moreno

Abstract:

Psychophysiological adaptive automation is a concept that combines human physiological data and computer algorithms to create personalized interfaces and experiences for users. This approach aims to enhance human learning by adapting to individual needs and preferences and optimizing the interaction between humans and machines. According to neurosciences, the working memory demand during the student learning process is modified when the student is learning a new subject or topic, managing and/or fulfilling a specific task goal. A sudden increase in working memory demand modifies the level of students’ attention, engagement, and cognitive load. The proposed psychophysiological adaptive automation system will adapt the task requirements to optimize cognitive load, the process output variable, by monitoring the student's brain activity. Cognitive load changes according to the student’s previous knowledge, the type of task, the difficulty level of the task, and the overall psychophysiological state of the student. Scaling the measured cognitive load as low, medium, or high; the system will assign a task difficulty level to the next task according to the ratio between the previous-task difficulty level and student stress. For instance, if a student becomes stressed or overwhelmed during a particular task, the system detects this through signal measurements such as brain waves, heart rate variability, or any other psychophysiological variables analyzed to adjust the task difficulty level. The control of engagement and stress are considered internal variables for the hypermedia system which selects between three different types of instructional material. This work assesses the feasibility of a fuzzy controller to track a student's physiological responses and adjust the learning content and pace accordingly. Using an industrial automation approach, the proposed fuzzy logic controller is based on linguistic rules that complement the instrumentation of the system to monitor and control the delivery of instructional material to the students. From the test results, it can be proved that the implemented fuzzy controller can satisfactorily regulate the delivery of academic content based on the working memory demand without compromising students’ health. This work has a potential application in the instructional design of virtual reality environments for training and education.

Keywords: fuzzy logic controller, hypermedia control system, personalized education, psychophysiological adaptive automation

Procedia PDF Downloads 62
855 Polypeptide Modified Carbon Nanotubes – Mediated GFP Gene Transfection for H1299 Cells and Toxicity Assessment

Authors: Pei-Ying Lo, Jing-Hao Ciou, Kai-Cheng Yang, Jia-Huei Zheng, Shih-Hsiang Huang, Kuen-Chan Lee, Er-Chieh Cho

Abstract:

As-produced CNTs are insoluble in all organic solvents and aqueous solutions have imposed limitations to the use of CNTs. Therefore, how to debundle carbon nanotubes and to modify them for further uses is an important issue. There are several methods for the dispersion of CNTs in water using covalent attachment of hydrophilic groups to the surface of tubes. These methods, however, alter the electronic structure of the nanotubes by disrupting the network of sp2 hybridized carbons. In order to keep the nanotubes’ intrinsic mechanical and electrical properties intact, non-covalent interactions are increasingly being explored as an alternative route for dispersion. Apart from conventional surfactants such as sodium dodecylsulfate (SDS) or sodium dodecylbenzenesulfonate (SDBS) which are highly effective in dispersing CNTs, biopolymers have received much attention as dispersing agents due to the anticipated biocompatibility of the dispersed CNTs. Also, The pyrenyl group is known to interact strongly with the basal plane of graphene via π-stacking. In this study, a highly re-dispersible biopolymer is reported for the synthesis of pyrene-modified poly-L-lysine (PBPL) and poly(D-Glu, D-Lys) (PGLP). To provide the evidence of the safety of the PBPL/CNT & PGLP/CNT materials we use in this study, H1299 and HCT116 cells were incubated with PBPL/CNT & PGLP/CNT materials for toxicity analysis, MTS assays. The results from MTS assays indicated that no significant cellular toxicity was shown in H1299 and HCT116 cells. Furthermore, the fluorescence marker fluorescein isothiocyanate (FITC) was added to PBPL & PGLP dispersions. From the fluorescent measurements showed that the chemical functionalisation of the PBPL/CNT & PGLP/CNT conjugates with the fluorescence marker were successful. The fluorescent PBPL/CNT & PGLP/CNT conjugates could find application in medical imaging. In the next step, the GFP gene is immobilized onto PBPL/CNT conjugates by introducing electrostatic interaction. GFP-transfected cells that emitted fluorescence were imaged and counted under a fluorescence microscope. Due to the unique biocompatibility of PBPL modified CNTs, the GFP gene could be transported into H1299 cells without using antibodies. The applicability of such soluble and chemically functionalised polypeptide/CNT conjugates in biomedicine is currently investigated. We expect that this polypeptide/CNT system will be a safe and multi-functional nanomedical delivery platform and contribute to future medical therapy.

Keywords: carbon nanotube, nanotoxicology, GFP transfection, polypeptide/CNT hybrids

Procedia PDF Downloads 328
854 Multicenter Baseline Survey to Outline Antimicrobial Prescribing Practices at Six Public Sectortertiary Care Hospitals in a Low Middle Income Country

Authors: N. Khursheed, M. Fatima, S. Jamal, A. Raza, S. Rattani, Q. Ahsan, A. Rasheed, M. Jawed

Abstract:

Introduction: Antibiotics are among the commonly prescribed medicines to treat bacterial infections. Their misuse intensifies resistance, and overuse incurs heavy losses to the healthcare system in terms of increased treatment costs and enhanced disease burden. Studies show that 40% of empirically used antibiotics are irrationally utilized. The objective of this study was to evaluate prescribing pattern of antibiotics at six public sector tertiary care hospitals across Pakistan. Methods: A multicenter cross-sectional point prevalence survey (PPS) was conducted in selected wards of six public sector tertiary care hospitals in Pakistan as part of the Clinical Engagement program by Fleming Fund Country Grant Pakistan in collaboration with Indus Hospital & Health Network (IHHN) from February to March 2021, these included Jinnah Postgraduate Medical Center and Dr. Ruth K. M. Pfau Civil Hospital from Karachi, Sheikh Zayed Hospital Lahore, Nishtar Medical University Hospital Multan, Medical Teaching Institute Hayatabad Medical Complex Peshawar, and Provincial Headquarters Hospital Gilgit. WHO PPS methodology was used for data collection (Hospital, ward, and patient level data was collected). Data was entered into the open-source Kobo Collect application and was analyzed using SPSS (version 22.0). Findings: Medical records of 837 in-patients were surveyed, of which the prevalence of antibiotics use was 78.5%. The most commonly prescribed antimicrobial was Ceftriaxone (21.7%) which is categorized in the Watch group of WHO AWaRe Classification, followed by Metronidazole (17.3%), Cefoperazone/Sulbactam (8.4%), Co-Amoxiclav (6.3%) and Piperacillin/Tazobactam (5.9%). The antibiotics were prescribed largely for surgical prophylaxis (36.7%), followed by community-acquired infections (24.7%). One antibiotic was prescribed to 46.7%, two to 39.9%, and three or more to 12.5 %. Two of six (30%) hospitals had functional drug and therapeutic committees, three (50%) had infection prevention and control committees, and one facility had an antibiotic formulary. Conclusion: Findings demonstrate high consumption of broad-spectrum antimicrobials and emphasizes the importance of expanding the antimicrobial stewardship program. Mentoring clinical teams will help to rationalize antimicrobial use.

Keywords: antimicrobial resistance, antimicrobial stewardship, point prevalence survey, antibiotics

Procedia PDF Downloads 91
853 Energy Audit and Renovation Scenarios for a Historical Building in Rome: A Pilot Case Towards the Zero Emission Building Goal

Authors: Domenico Palladino, Nicolandrea Calabrese, Francesca Caffari, Giulia Centi, Francesca Margiotta, Giovanni Murano, Laura Ronchetti, Paolo Signoretti, Lisa Volpe, Silvia Di Turi

Abstract:

The aim to achieve a fully decarbonized building stock by 2050 stands as one of the most challenging issues within the spectrum of energy and climate objectives. Numerous strategies are imperative, particularly emphasizing the reduction and optimization of energy demand. Ensuring the high energy performance of buildings emerges as a top priority, with measures aimed at cutting energy consumptions. Concurrently, it is imperative to decrease greenhouse gas emissions by using renewable energy sources for the on-site energy production, thereby striving for an energy balance leading towards zero-emission buildings. Italy's predominant building stock comprises ancient buildings, many of which hold historical significance and are subject to stringent preservation and conservation regulations. Attaining high levels of energy efficiency and reducing CO2 emissions in such buildings poses a considerable challenge, given their unique characteristics and the imperative to adhere to principles of conservation and restoration. Additionally, conducting a meticulous analysis of these buildings' current state is crucial for accurately quantifying their energy performance and predicting the potential impacts of proposed renovation strategies on energy consumption reduction. Within this framework, the paper presents a pilot case in Rome, outlining a methodological approach for the renovation of historic buildings towards achieving Zero Emission Building (ZEB) objective. The building has a mixed function with offices, a conference hall, and an exposition area. The building envelope is made of historical and precious materials used as cladding which must be preserved. A thorough understanding of the building's current condition serves as a prerequisite for analyzing its energy performance. This involves conducting comprehensive archival research, undertaking on-site diagnostic examinations to characterize the building envelope and its systems, and evaluating actual energy usage data derived from energy bills. Energy simulations and audit are the first step in the analysis with the assessment of the energy performance of the actual current state. Subsequently, different renovation scenarios are proposed, encompassing advanced building techniques, to pinpoint the key actions necessary for improving mechanical systems, automation and control systems, and the integration of renewable energy production. These scenarios entail different levels of renovation, ranging from meeting minimum energy performance goals to achieving the highest possible energy efficiency level. The proposed interventions are meticulously analyzed and compared to ascertain the feasibility of attaining the Zero Emission Building objective. In conclusion, the paper provides valuable insights that can be extrapolated to inform a broader approach towards energy-efficient refurbishment of historical buildings that may have limited potential for renovation in their building envelopes. By adopting a methodical and nuanced approach, it is possible to reconcile the imperative of preserving cultural heritage with the pressing need to transition towards a sustainable, low-carbon future.

Keywords: energy conservation and transition, energy efficiency in historical buildings, buildings energy performance, energy retrofitting, zero emission buildings, energy simulation

Procedia PDF Downloads 47
852 Cord Blood Hematopoietic Stem Cell Expansion Ability of Mesenchymal Stem Cells Isolated From Different Sources

Authors: Ana M. Lara, Manuela Llano, Felipe Gaitán, Rosa H. Bustos, Ana Maria Perdomo-Arciniegas, Ximena Bonilla

Abstract:

Umbilical cord blood is used as a source of progenitor and stem cells for the regeneration of the hematopoietic and immune system to treat patients with different hematological or non-hematological diseases. This stem cell source represents an advantage over the use of bone marrow or mobilized peripheral blood because it has a lower incidence rate of graft-versus-host disease, probably due to fewer immunological compatibility restrictions. However, its low cellular dose limits its use in pediatric patients. This work proposes the standardization of a cell expansion technique to compensate for the dose of infused cells through the ex-vivo manipulation of hematopoietic progenitor cells from umbilical cord blood before transplantation. The expansion model is carried out through co-cultures with mesenchymal stem cells (MSC) from bone marrow (BM) and less explored fetal tissues such as Wharton's jelly (WJ) and umbilical cord blood (UCB). Initially, a master cell bank of primary mesenchymal stem cells isolated from different sources was established and characterized following International Society of Cell Therapies (ISCT) indications. Additionally, we assessed the effect of a short 25 Gy cycle of gamma irradiation on cell cycle arrest of mesenchymal cells over the support capacity for the expansion of hematopoietic stem cells from umbilical cord blood was evaluated. The results show that co-cultures with MSC from WJ and UCB allow the cellular dose of HSPC to be maximized between 5 and 16 times having a similar support capacity as BM. In addition, was evaluated the hematopoietic stem progenitor cell's HSPC functionality through the evaluation of migration capacity, their differentiation capacity during culture time by flow cytometry to evaluate the expression of membrane markers associated with lineage-committed progenitors, their clonogenic potential, and the evaluation of secretome profile in the expansion process was evaluated. So far, the treatment with gamma irradiation maintains the hematopoietic support capacity of mesenchymal stem cells from the three sources studied compared to treatments without irradiation, favoring the use of fetal tissues that are generally waste to obtain mesenchymal cell lines for ex-vivo expansion systems. With the results obtained, a standardized protocol that will contribute to the development of ex-vivo expansion with MSC on a larger scale will be achieved, enabling its clinical use and expanding its application in adults.

Keywords: ex-vivo expansion, hematopoietic stem cells, hematopoietic stem cell transplantation, mesenchymal stem cells, umbilical cord blood

Procedia PDF Downloads 102
851 Evaluating of Chemical Extractants for Assessment of Bioavailable Heavy Metals in Polluted Soils

Authors: Violina Angelova, Krasimir Ivanov, Stefan Krustev, Dimitar Dimitrov

Abstract:

Availability of a metal is characterised by its quantity transgressing from soil into different extractants or by its content in plants. In literature, the terms 'available forms of compounds' and 'mobile' are often considered as equivalents of the term 'accessible' to plants. Rapid and a sufficiently reliable method for defining the accessible for plants forms turns out to be their extraction through different extractants, imitating the functioning of the root system. As a criterion for the pertinence of the extractant to this purpose usually serves the significant statistic correlation between the extracted quantities of the element from soil and its content in plants. The aim of this work was to evaluate the effectiveness of various extractions (DTPA-TEA, AB-DTPA, Mehlich 3, 0.01 M CaCl₂, 1M NH₄NO₃) for the determination of bioavailability of heavy metals in industrially polluted soils from the metallurgical activity near Plovdiv and Kardjali, Bulgaria. Quantity measurements for contents of heavy metals were performed with ICP-OES. The results showed that extraction capacity was as follows: Mehlich 3>ABDTPA>DTPA-TEA>CaCl₂>NaNO₃. The content of the mobile form of heavy metals depends on the nature of metal ion, the nature of extractant and pH. The obtained results show that CaCl₂ extracts a greater quantity of mobile forms of heavy metals than NH₄NO₃. DTPA-TEA and AB-DTPA are capable of extracting from the soil not only the heavy metals participating in the exchange processes but also the heavy metals bound in carbonates and organic complexes, as well as bound and occluded in oxide and secondary clay minerals. AB-DTPA extracts a bit more heavy metals than DTPA-TEA. The darker color of the solutions obtained with AB-DTPA indicates that considerable quantities organic matter are being destructed. A comparison of the mobile forms of heavy metals extracted from clean and highly polluted soils has revealed that in the polluted soils the greater portion of heavy metals exists in a mobile form. High correlation coefficients are obtained between the metals extracted with different extractants and their total content in soil (r=0.9). A positive correlation between the pH, soil organic matter and the extracted quantities of heavy metals has been found. The results of correlation analysis revealed that the heavy metals extracted by DTPA-TEA, AB-DTPA, Mehlich 3, CaCl₂ and NaNO₃ correlated significantly with plant uptake. Significant correlation was found between DTPA-TEA, AB-DTPA, and CaCl₂ with heavy metals concentration in plants. Application of extracting methods contains chelating agents would be recommended in the future research onthe availabilityof heavy metals in polluted soils.

Keywords: availability, chemical extractants, heavy metals, mobile forms

Procedia PDF Downloads 338
850 Exploring Closed-Loop Business Systems Which Eliminates Solid Waste in the Textile and Fashion Industry: A Systematic Literature Review Covering the Developments Occurred in the Last Decade

Authors: Bukra Kalayci, Geraldine Brennan

Abstract:

Introduction: Over the last decade, a proliferation of literature related to textile and fashion business in the context of sustainable production and consumption has emerged. However, the economic and environmental benefits of solid waste recovery have not been comprehensively searched. Therefore at the end-of-life or end-of-use textile waste management remains a gap. Solid textile waste reuse and recycling principles of the circular economy need to be developed to close the disposal stage of the textile supply chain. The environmental problems associated with the over-production and –consumption of textile products arise. Together with growing population and fast fashion culture the share of solid textile waste in municipal waste is increasing. Focusing on post-consumer textile waste literature, this research explores the opportunities, obstacles and enablers or success factors associated with closed-loop textile business systems. Methodology: A systematic literature review was conducted in order to identify best practices and gaps from the existing body of knowledge related to closed-loop post-consumer textile waste initiatives over the last decade. Selected keywords namely: ‘cradle-to-cradle ‘, ‘circular* economy* ‘, ‘closed-loop* ‘, ‘end-of-life* ‘, ‘reverse* logistic* ‘, ‘take-back* ‘, ‘remanufacture* ‘, ‘upcycle* ‘ with the combination of (and) ‘fashion* ‘, ‘garment* ‘, ‘textile* ‘, ‘apparel* ‘, clothing* ‘ were used and the time frame of the review was set between 2005 to 2017. In order to obtain a broad coverage, Web of Knowledge and Science Direct databases were used, and peer-reviewed journal articles were chosen. The keyword search identified 299 number of papers which was further refined into 54 relevant papers that form the basis of the in-depth thematic analysis. Preliminary findings: A key finding was that the existing literature is predominantly conceptual rather than applied or empirical work. Moreover, the enablers or success factors, obstacles and opportunities to implement closed-loop systems in the textile industry were not clearly articulated and the following considerations were also largely overlooked in the literature. While the circular economy suggests multiple cycles of discarded products, components or materials, most research has to date tended to focus on a single cycle. Thus the calculations of environmental and economic benefits of closed-loop systems are limited to one cycle which does not adequately explore the feasibility or potential benefits of multiple cycles. Additionally, the time period textile products spend between point of sale, and end-of-use/end-of-life return is a crucial factor. Despite past efforts to study closed-loop textile systems a clear gap in the literature is the lack of a clear evaluation framework which enables manufacturers to clarify the reusability potential of textile products through consideration of indicators related too: quality, design, lifetime, length of time between manufacture and product return, volume of collected disposed products, material properties, and brand segment considerations (e.g. fast fashion versus luxury brands).

Keywords: circular fashion, closed loop business, product service systems, solid textile waste elimination

Procedia PDF Downloads 187
849 Effects of Stokes Shift and Purcell Enhancement in Fluorescence Assisted Radiative Cooling

Authors: Xue Ma, Yang Fu, Dangyuan Lei

Abstract:

Passive daytime radiative cooling is an emerging technology which has attracted worldwide attention in recent years due to its huge potential in cooling buildings without the use of electricity. Various coating materials with different optical properties have been developed to improve the daytime radiative cooling performance. However, commercial cooling coatings comprising functional fillers with optical bandgaps within the solar spectral range suffers from severe intrinsic absorption, limiting their cooling performance. Fortunately, it has recently been demonstrated that introducing fluorescent materials into polymeric coatings can covert the absorbed sunlight to fluorescent emissions and hence increase the effective solar reflectance and cooling performance. In this paper, we experimentally investigate the key factors for fluorescence-assisted radiative cooling with TiO2-based white coatings. The surrounding TiO2 nanoparticles, which enable spatial and temporal light confinement through multiple Mie scattering, lead to Purcell enhancement of phosphors in the coating. Photoluminescence lifetimes of two phosphors (BaMgAl10O17:Eu2+ and (Sr, Ba)SiO4:Eu2+) exhibit significant reduction of ~61% and ~23%, indicating Purcell factors of 2.6 and 1.3, respectively. Moreover, smaller Stokes shifts of the phosphors are preferred to further diminish solar absorption. Field test of fluorescent cooling coatings demonstrate an improvement of ~4% solar reflectance for the BaMgAl10O17:Eu2+-based fluorescent cooling coating. However, to maximize solar reflectance, a white appearance is introduced based on multiple Mie scattering by the broad size distribution of fillers, which is visually pressurized and aesthetically bored. Besides, most colored pigments absorb visible light significantly and convert it to non-radiative thermal energy, offsetting the cooling effect. Therefore, current colored cooling coatings are facing the compromise between color saturation and cooling effect. To solve this problem, we introduced colored fluorescent materials into white coating based on SiO2 microspheres as a top layer, covering a white cooling coating based on TiO2. Compared with the colored pigments, fluorescent materials could re-emit the absorbed light, reducing the solar absorption introduced by coloration. Our work investigated the scattering properties of SiO2 dielectric spheres with different diameters and detailly discussed their impact on the PL properties of phosphors, paving the way for colored fluorescent-assisted cooling coting to application and industrialization.

Keywords: solar reflection, infrared emissivity, mie scattering, photoluminescent emission, radiative cooling

Procedia PDF Downloads 72
848 Application of Metarhizium anisopliae against Meloidogyne javanica in Soil Amended with Oak Debris

Authors: Mohammad Abdollahi

Abstract:

Tomato (Lycopersicon esculentum Mill.) is one of the most popular, widely grown and the second most important vegetable crop, after potatoes. Nematodes have been identified as one of the major pests affecting tomato production throughout the world. The most destructive nematodes are the genus Meloidogyne. Most widespread and devastating species of this genus are M. incognita, M. javanica, and M. arenaria. These species can cause complete crop loss under adverse growing conditions. There are several potential methods for management of the root knot nematodes. Although the chemicals are widely used against the phytonematodes, because of hazardous effects of these compounds on non-target organisms and on the environment, there is a need to develop other control strategies. Nowadays, non-chemical measures are widely used to control the plant parasitic nematodes. Biocontrol of phytonematodes is an important method among environment-friendly measures of nematode management. There are some soil-inhabiting fungi that have biocontrol potential on phytonematodes, which can be used in nematode management program. The fungus Metarhizium anisopliae, originally is an entomopathogenic bioagent. Biocontrol potential of this fungus on some phytonematodes has been reported earlier. Recently, use of organic soil amendments as well as the use of bioagents is under special attention in sustainable agriculture. This research aimed to reduce the pesticide use in control of root-knot nematode, Meloidogyne javanica in tomato. The effects of M. anisopliae IMI 330189 and different levels of oak tree debris on M. javanica were determined. The combination effect of the fungus as well as the different rates of soil amendments was determined. Pots were filled with steam pasteurized soil mixture and the six leaf tomato seedlings were inoculated with 3000 second stage larvae of M. javanica/kg of soil. After eight weeks, plant growth parameters and nematode reproduction factors were compared. Based on the results of our experiment, combination of M. anisopliae IMI 330189 and oak debris caused more than 90% reduction in reproduction factor of nematode, at the rates of 100 and 150 g/kg soil (P ≤ 0.05). As compared to control, the reduction in number of galls was 76%. It was 86% for nematode reproduction factor, showing the significance of combined effect of both tested agents. Our results showed that plant debris can increase the biological activity of the tested bioagent. It was also proved that there was no adverse effect of oak debris, which potentially has antimicrobial activity, on antagonistic power of applied bioagent.

Keywords: biological control, nematode management, organic soil, Quercus branti, root knot nematode, soil amendment

Procedia PDF Downloads 156
847 Pivoting to Fortify our Digital Self: Revealing the Need for Personal Cyber Insurance

Authors: Richard McGregor, Carmen Reaiche, Stephen Boyle

Abstract:

Cyber threats are a relatively recent phenomenon and offer cyber insurers a dynamic and intelligent peril. As individuals en mass become increasingly digitally dependent, Personal Cyber Insurance (PCI) offers an attractive option to mitigate cyber risk at a personal level. This abstract proposes a literature review that conceptualises a framework for siting Personal Cyber Insurance (PCI) within the context of cyberspace. The lack of empirical research within this domain demonstrates an immediate need to define the scope of PCI to allow cyber insurers to understand personal cyber risk threats and vectors, customer awareness, capabilities, and their associated needs. Additionally, this will allow cyber insurers to conceptualise appropriate frameworks allowing effective management and distribution of PCI products and services within a landscape often in-congruent with risk attributes commonly associated with traditional personal line insurance products. Cyberspace has provided significant improvement to the quality of social connectivity and productivity during past decades and allowed enormous capability uplift of information sharing and communication between people and communities. Conversely, personal digital dependency furnish ample opportunities for adverse cyber events such as data breaches and cyber-attacksthus introducing a continuous and insidious threat of omnipresent cyber risk–particularly since the advent of the COVID-19 pandemic and wide-spread adoption of ‘work-from-home’ practices. Recognition of escalating inter-dependencies, vulnerabilities and inadequate personal cyber behaviours have prompted efforts by businesses and individuals alike to investigate strategies and tactics to mitigate cyber risk – of which cyber insurance is a viable, cost-effective option. It is argued that, ceteris parabus, the nature of cyberspace intrinsically provides characteristic peculiarities that pose significant and bespoke challenges to cyber insurers, often in-congruent with risk attributes commonly associated with traditional personal line insurance products. These challenges include (inter alia) a paucity of historical claim/loss data for underwriting and pricing purposes, interdependencies of cyber architecture promoting high correlation of cyber risk, difficulties in evaluating cyber risk, intangibility of risk assets (such as data, reputation), lack of standardisation across the industry, high and undetermined tail risks, and moral hazard among others. This study proposes a thematic overview of the literature deemed necessary to conceptualise the challenges to issuing personal cyber coverage. There is an evident absence of empirical research appertaining to PCI and the design of operational business models for this business domain, especially qualitative initiatives that (1) attempt to define the scope of the peril, (2) secure an understanding of the needs of both cyber insurer and customer, and (3) to identify elements pivotal to effective management and profitable distribution of PCI - leading to an argument proposed by the author that postulates that the traditional general insurance customer journey and business model are ill-suited for the lineaments of cyberspace. The findings of the review confirm significant gaps in contemporary research within the domain of personal cyber insurance.

Keywords: cyberspace, personal cyber risk, personal cyber insurance, customer journey, business model

Procedia PDF Downloads 89
846 Becoming a Good-Enough White Therapist: Experiences of International Students in Psychology Doctoral Programs

Authors: Mary T. McKinley

Abstract:

As socio-economic globalization impacts education and turns knowledge into a commodity, institutions of higher education are becoming more intentional about infusing a global and intercultural perspective into education via the recruitment of international students. Coming from dissimilar cultures, many of these students are evaluated and held accountable to Euro-American values of independence, self-reliance, and autonomy. Not surprisingly, these students often experience culture shock with deleterious effects on their mental health and academic functioning. Thus, it is critical to understand the experiences of international students with the hope that such knowledge will keep the field of psychology from promulgating Eurocentric ideals and values and prevent the training of these students as good-enough White therapists. Using a critical narrative inquiry framework, this study elicits stories about the challenges encountered by international students as they navigate their clinical training in the presence of acculturative stress and potentially different worldviews. With its emphasis on story-telling as meaning making, narrative research design is hinged on the assumption that people are interpretive beings who make meaning of themselves and their world through the language of stories. Also, dominant socially-constructed narratives play a central role in creating and maintaining hegemonic structures that privilege certain individuals and ideologies at the expense of others. On this premise, narrative inquiry begins with an exploration of the experiences of participants in their lived stories. Bounded narrative segments were read, interpreted, and analyzed using a critical events approach. Throughout the process, issues of reliability and researcher bias were addressed by keeping a reflective analytic memo, as well as triangulating the data using peer-reviewers and check-ins with participants. The findings situate culture at the epicenter of international students’ acculturation challenges as well as their resiliency in psychology doctoral programs. It was not uncommon for these international students to experience ethical dilemmas inherent in learning content that conflicted with their cultural beliefs and values. Issues of cultural incongruence appear to be further exacerbated by visible markers for differences like speech accent and clothing attire. These stories also link the acculturative stress reported by international students to the experiences of perceived racial discrimination and lack of support from the faculty, administration, peers, and the society at large. Beyond the impact on the international students themselves, there are implications for internationalization in psychology with the goal of equipping doctoral programs to be better prepared to meet the needs of their international students. More than ever before, programs need to liaise with international students’ services and work in tandem to meet the unique needs of this population of students. Also, there exists a need for multiculturally competent supervisors working with international students with varying degrees of acculturation. In addition to making social justice and advocacy salient in students’ multicultural training, it may be helpful for psychology doctoral programs to be more intentional about infusing cross-cultural theories, indigenous psychotherapies, and/or when practical, the possibility for geographically cross-cultural practicum experiences in the home countries of international students while taking into consideration the ethical issues for virtual supervision.

Keywords: decolonizing pedagogies, international students, multiculturalism, psychology doctoral programs

Procedia PDF Downloads 102
845 Higher-Level Return to Female Karate Competition Following Multiple Patella Dislocations

Authors: A. Maso, C. Bellissimo, G. Facchinetti, N. Milani, D. Panzin, D. Pogliana, L. Garlaschelli, L. Rivaroli, S. Rivaroli, M. Zurek, J. Konin

Abstract:

15 year-old female karate athlete experienced two unilateral patella dislocations: one contact and one non-contact. This challenged her from competing as planned at the regional and national competitions as a result of her inability to perform at a high level. Despite these injuries and other complicated factors, she was able to modify her training timeline and successfully perform, winning third at the National Cup. Initial pain numeric rating scale 8/10 during karate training isometric figures, taking the stairs, long walking, a positive rasp test, palpation pain on the lateral patella joint 9/10, pain performing open kinetic chain 0°-45° and close kinetic chain 30°-90°, tensor fascia lata, vastus lateralis, psoas muscles retraction/stiffness. Foot hyper pronation, internally rotated femur, and knee flexion 15° were the postural findings. Exercise prescription for three days/week for three weeks to include exercise-based rehabilitation and soft tissue mobilization with massage and foam rolling. After three weeks, the pain was improved during activity daily living 5/10, and soft tissue stiffness decreased. An additional four weeks of exercise-based rehabilitation was continued. At this time, axial x-rays and TA-GT TAC were taken, and an orthopaedic medical check was recommended to continue conservative treatment. At week seven, she performed 2/4 karate position technique without pain and 2/4 with pain. An isokinetic test was performed at week 12, demonstrating a 10% strength deficit and 6% resistance deficit both to the left hamstrings. Moreover, an 8% strength and resistance surplus to the left quadriceps was found. No pain was present during activity, daily living and sports activity, allowing a return to play training to begin. A plan for the return to play framework collaborated with her trainer, her father, a physiotherapist, a sports scientist, an osteopath, and a nutritionist. Within 4 and 5 months, both non-athlete and athlete movement quality analysis tests were performed. The plan agreed to establish a return to play goal of 7 months and the highest level return to competition goal of 9 months from the start of rehabilitation. This included three days/week of training and repeated testing of movement quality before return to competition with detectable improvements from 77% to 93%. Beginning goals of the rehabilitation plan included the importance of a team approach. The patient’s father and trainer were important to collaborate with to assure a safe and timely return to competition. The possibility of achieving the goals was strongly related to orthopaedic decision-making and progress during the first weeks of rehabilitation. Without complications or setbacks, the patient can successfully return to her highest level of competition. The patient returned to participation after five months of rehabilitation and training, and then she returned to competition at the national level in nine months. The successful return was the result of a team approach and a compliant patient with clear goals.

Keywords: karate, knee, performance, rehabilitation

Procedia PDF Downloads 91
844 Knowledge Graph Development to Connect Earth Metadata and Standard English Queries

Authors: Gabriel Montague, Max Vilgalys, Catherine H. Crawford, Jorge Ortiz, Dava Newman

Abstract:

There has never been so much publicly accessible atmospheric and environmental data. The possibilities of these data are exciting, but the sheer volume of available datasets represents a new challenge for researchers. The task of identifying and working with a new dataset has become more difficult with the amount and variety of available data. Datasets are often documented in ways that differ substantially from the common English used to describe the same topics. This presents a barrier not only for new scientists, but for researchers looking to find comparisons across multiple datasets or specialists from other disciplines hoping to collaborate. This paper proposes a method for addressing this obstacle: creating a knowledge graph to bridge the gap between everyday English language and the technical language surrounding these datasets. Knowledge graph generation is already a well-established field, although there are some unique challenges posed by working with Earth data. One is the sheer size of the databases – it would be infeasible to replicate or analyze all the data stored by an organization like The National Aeronautics and Space Administration (NASA) or the European Space Agency. Instead, this approach identifies topics from metadata available for datasets in NASA’s Earthdata database, which can then be used to directly request and access the raw data from NASA. By starting with a single metadata standard, this paper establishes an approach that can be generalized to different databases, but leaves the challenge of metadata harmonization for future work. Topics generated from the metadata are then linked to topics from a collection of English queries through a variety of standard and custom natural language processing (NLP) methods. The results from this method are then compared to a baseline of elastic search applied to the metadata. This comparison shows the benefits of the proposed knowledge graph system over existing methods, particularly in interpreting natural language queries and interpreting topics in metadata. For the research community, this work introduces an application of NLP to the ecological and environmental sciences, expanding the possibilities of how machine learning can be applied in this discipline. But perhaps more importantly, it establishes the foundation for a platform that can enable common English to access knowledge that previously required considerable effort and experience. By making this public data accessible to the full public, this work has the potential to transform environmental understanding, engagement, and action.

Keywords: earth metadata, knowledge graphs, natural language processing, question-answer systems

Procedia PDF Downloads 134
843 Governance in the Age of Artificial intelligence and E- Government

Authors: Mernoosh Abouzari, Shahrokh Sahraei

Abstract:

Electronic government is a way for governments to use new technology that provides people with the necessary facilities for proper access to government information and services, improving the quality of services and providing broad opportunities to participate in democratic processes and institutions. That leads to providing the possibility of easy use of information technology in order to distribute government services to the customer without holidays, which increases people's satisfaction and participation in political and economic activities. The expansion of e-government services and its movement towards intelligentization has the ability to re-establish the relationship between the government and citizens and the elements and components of the government. Electronic government is the result of the use of information and communication technology (ICT), which by implementing it at the government level, in terms of the efficiency and effectiveness of government systems and the way of providing services, tremendous commercial changes are created, which brings people's satisfaction at the wide level will follow. The main level of electronic government services has become objectified today with the presence of artificial intelligence systems, which recent advances in artificial intelligence represent a revolution in the use of machines to support predictive decision-making and Classification of data. With the use of deep learning tools, artificial intelligence can mean a significant improvement in the delivery of services to citizens and uplift the work of public service professionals while also inspiring a new generation of technocrats to enter government. This smart revolution may put aside some functions of the government, change its components, and concepts such as governance, policymaking or democracy will change in front of artificial intelligence technology, and the top-down position in governance may face serious changes, and If governments delay in using artificial intelligence, the balance of power will change and private companies will monopolize everything with their pioneering in this field, and the world order will also depend on rich multinational companies and in fact, Algorithmic systems will become the ruling systems of the world. It can be said that currently, the revolution in information technology and biotechnology has been started by engineers, large economic companies, and scientists who are rarely aware of the political complexities of their decisions and certainly do not represent anyone. Therefore, it seems that if liberalism, nationalism, or any other religion wants to organize the world of 2050, it should not only rationalize the concept of artificial intelligence and complex data algorithm but also mix them in a new and meaningful narrative. Therefore, the changes caused by artificial intelligence in the political and economic order will lead to a major change in the way all countries deal with the phenomenon of digital globalization. In this paper, while debating the role and performance of e-government, we will discuss the efficiency and application of artificial intelligence in e-government, and we will consider the developments resulting from it in the new world and the concepts of governance.

Keywords: electronic government, artificial intelligence, information and communication technology., system

Procedia PDF Downloads 82