Search results for: high dose rate brachytherapy
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 25526

Search results for: high dose rate brachytherapy

1346 An Analysis of the Strategic Pathway to Building a Successful Mobile Advertising Business in Nigeria: From Strategic Intent to Competitive Advantage

Authors: Pius A. Onobhayedo, Eugene A. Ohu

Abstract:

Nigeria has one of the fastest growing mobile telecommunications industry in the world. In the absence of fixed connection access to the Internet, access to the Internet is primarily via mobile devices. It, therefore, provides a test case for how to penetrate the mobile market in an emerging economy. We also hope to contribute to a sparse literature on strategies employed in building successful data-driven mobile businesses in emerging economies. We, therefore, sought to identify and analyse the strategic approach taken in a successful locally born mobile data-driven business in Nigeria. The analysis was carried out through the framework of strategic intent and competitive advantages developed from the conception of the company to date. This study is based on an exploratory investigation of an innovative digital company based in Nigeria specializing in the mobile advertising business. The projected growth and high adoption of mobile in this African country, coinciding with the smartphone revolution triggered by the launch of iPhone in 2007 opened a new entrepreneurial horizon for the founder of the company, who reached the conclusion that ‘the future is mobile’. This dream led to the establishment of three digital businesses, designed for convergence and complementarity of medium and content. The mobile Ad subsidiary soon grew to become a truly African network with operations and campaigns across West, East and South Africa, successfully delivering campaigns in several African countries including Nigeria, Kenya, South Africa, Ghana, Uganda, Zimbabwe, and Zambia amongst others. The company recently declared a 40% year-end profit which was nine times that of the previous financial year. This study drew from an in-depth interview with the company’s founder, analysis of primary and secondary data from and about the business, as well as case studies of digital marketing campaigns. We hinge our analysis on the strategic intent concept which has been proposed to be an engine that drives the quest for sustainable strategic advantage in the global marketplace. Our goal was specifically to identify the strategic intents of the founder and how these were transformed creatively into processes that may have led to some distinct competitive advantages. Along with the strategic intents, we sought to identify the respective absorptive capacities that constituted favourable antecedents to the creation of such competitive advantages. Our recommendations and findings will be pivotal information for anybody wishing to invest in the world’s fastest technology business space - Africa.

Keywords: Africa, competitive advantage, competitive strategy, digital, mobile business, marketing, strategic intent

Procedia PDF Downloads 422
1345 Advantages of Computer Navigation in Knee Arthroplasty

Authors: Mohammad Ali Al Qatawneh, Bespalchuk Pavel Ivanovich

Abstract:

Computer navigation has been introduced in total knee arthroplasty to improve the accuracy of the procedure. Computer navigation improves the accuracy of bone resection in the coronal and sagittal planes. It was also noted that it normalizes the rotational alignment of the femoral component and fully assesses and balances the deformation of soft tissues in the coronal plane. The work is devoted to the advantages of using computer navigation technology in total knee arthroplasty in 62 patients (11 men and 51 women) suffering from gonarthrosis, aged 51 to 83 years, operated using a computer navigation system, followed up to 3 years from the moment of surgery. During the examination, the deformity variant was determined, and radiometric parameters of the knee joints were measured using the Knee Society Score (KSS), Functional Knee Society Score (FKSS), and Western Ontario and McMaster University Osteoarthritis Index (WOMAC) scales. Also, functional stress tests were performed to assess the stability of the knee joint in the frontal plane and functional indicators of the range of motion. After surgery, improvement was observed in all scales; firstly, the WOMAC values decreased by 5.90 times, and the median value to 11 points (p < 0.001), secondly KSS increased by 3.91 times and reached 86 points (p < 0.001), and the third one is that FKSS data increased by 2.08 times and reached 94 points (p < 0.001). After TKA, the axis deviation of the lower limbs of more than 3 degrees was observed in 4 patients at 6.5% and frontal instability of the knee joint just in 2 cases at 3.2%., The lower incidence of sagittal instability of the knee joint after the operation was 9.6%. The range of motion increased by 1.25 times; the volume of movement averaged 125 degrees (p < 0.001). Computer navigation increases the accuracy of the spatial orientation of the endoprosthesis components in all planes, reduces the variability of the axis of the lower limbs within ± 3 °, allows you to achieve the best results of surgical interventions, and can be used to solve most basic tasks, allowing you to achieve excellent and good outcomes of operations in 100% of cases according to the WOMAC scale. With diaphyseal deformities of the femur and/or tibia, as well as with obstruction of their medullary canal, the use of computer navigation is the method of choice. The use of computer navigation prevents the occurrence of flexion contracture and hyperextension of the knee joint during the distal sawing of the femur. Using the navigation system achieves high-precision implantation for the endoprosthesis; in addition, it achieves an adequate balance of the ligaments, which contributes to the stability of the joint, reduces pain, and allows for the achievement of a good functional result of the treatment.

Keywords: knee joint, arthroplasty, computer navigation, advantages

Procedia PDF Downloads 65
1344 Railway Ballast Volumes Automated Estimation Based on LiDAR Data

Authors: Bahar Salavati Vie Le Sage, Ismaïl Ben Hariz, Flavien Viguier, Sirine Noura Kahil, Audrey Jacquin, Maxime Convert

Abstract:

The ballast layer plays a key role in railroad maintenance and the geometry of the track structure. Ballast also holds the track in place as the trains roll over it. Track ballast is packed between the sleepers and on the sides of railway tracks. An imbalance in ballast volume on the tracks can lead to safety issues as well as a quick degradation of the overall quality of the railway segment. If there is a lack of ballast in the track bed during the summer, there is a risk that the rails will expand and buckle slightly due to the high temperatures. Furthermore, the knowledge of the ballast quantities that will be excavated during renewal works is important for efficient ballast management. The volume of excavated ballast per meter of track can be calculated based on excavation depth, excavation width, volume of track skeleton (sleeper and rail) and sleeper spacing. Since 2012, SNCF has been collecting 3D points cloud data covering its entire railway network by using 3D laser scanning technology (LiDAR). This vast amount of data represents a modelization of the entire railway infrastructure, allowing to conduct various simulations for maintenance purposes. This paper aims to present an automated method for ballast volume estimation based on the processing of LiDAR data. The estimation of abnormal volumes in ballast on the tracks is performed by analyzing the cross-section of the track. Further, since the amount of ballast required varies depending on the track configuration, the knowledge of the ballast profile is required. Prior to track rehabilitation, excess ballast is often present in the ballast shoulders. Based on 3D laser scans, a Digital Terrain Model (DTM) was generated and automatic extraction of the ballast profiles from this data is carried out. The surplus in ballast is then estimated by performing a comparison between this ballast profile obtained empirically, and a geometric modelization of the theoretical ballast profile thresholds as dictated by maintenance standards. Ideally, this excess should be removed prior to renewal works and recycled to optimize the output of the ballast renewal machine. Based on these parameters, an application has been developed to allow the automatic measurement of ballast profiles. We evaluated the method on a 108 kilometers segment of railroad LiDAR scans, and the results show that the proposed algorithm detects ballast surplus that amounts to values close to the total quantities of spoil ballast excavated.

Keywords: ballast, railroad, LiDAR , cloud point, track ballast, 3D point

Procedia PDF Downloads 86
1343 Evaluation of Arsenic Removal in Synthetic Solutions and Natural Waters by Rhizofiltration

Authors: P. Barreto, A. Guevara, V. Ibujes

Abstract:

In this study, the removal of arsenic from synthetic solutions and natural water from Papallacta Lagoon was evaluated, by using the rhizofiltration method with terrestrial and aquatic plant species. Ecuador is a country of high volcanic activity, that is why most of water sources come from volcanic glaciers. Therefore, it is necessary to find new, affordable and effective methods for treating water. The water from Papallacta Lagoon shows levels from 327 µg/L to 803 µg/L of arsenic. The evaluation for the removal of arsenic began with the selection of 16 different species of terrestrial and aquatic plants. These plants were immersed to solutions of 4500 µg/L arsenic concentration, for 48 hours. Subsequently, 3 terrestrial species and 2 aquatic species were selected based on the highest amount of absorbed arsenic they showed, analyzed by plasma optical emission spectrometry (ICP-OES), and their best capacity for adaptation into the arsenic solution. The chosen terrestrial species were cultivated from their seed with hydroponics methods, using coconut fiber and polyurethane foam as substrates. Afterwards, the species that best adapted to hydroponic environment were selected. Additionally, a control of the development for the selected aquatic species was carried out using a basic nutrient solution to provide the nutrients that the plants required. Following this procedure, 30 plants from the 3 types of species selected were exposed to a synthetic solution with levels of arsenic concentration of 154, 375 and 874 µg/L, for 15 days. Finally, the plant that showed the highest level of arsenic absorption was placed in 3 L of natural water, with arsenic levels of 803 µg/L. The plant laid in the water until it reached the desired level of arsenic of 10 µg/L. This experiment was carried out in a total of 30 days, in which the capacity of arsenic absorption of the plant was measured. As a result, the five species initially selected to be used in the last part of the evaluation were: sunflower (Helianthus annuus), clover (Trifolium), blue grass (Poa pratensis), water hyacinth (Eichhornia crassipes) and miniature aquatic fern (Azolla). The best result of arsenic removal was showed by the water hyacinth with a 53,7% of absorption, followed by the blue grass with 31,3% of absorption. On the other hand, the blue grass was the plant that best responded to the hydroponic cultivation, by obtaining a germination percentage of 97% and achieving its full growth in two months. Thus, it was the only terrestrial species selected. In summary, the final selected species were blue grass, water hyacinth and miniature aquatic fern. These three species were evaluated by immersing them in synthetic solutions with three different arsenic concentrations (154, 375 and 874 µg/L). Out of the three plants, the water hyacinth was the one that showed the highest percentages of arsenic removal with 98, 58 and 64%, for each one of the arsenic solutions. Finally, 12 plants of water hyacinth were chosen to reach an arsenic level up to 10 µg/L in natural water. This significant arsenic concentration reduction was obtained in 5 days. In conclusion, it was found that water hyacinth is the best plant to reduce arsenic levels in natural water.

Keywords: arsenic, natural water, plant species, rhizofiltration, synthetic solutions

Procedia PDF Downloads 110
1342 Nuclear Materials and Nuclear Security in India: A Brief Overview

Authors: Debalina Ghoshal

Abstract:

Nuclear security is the ‘prevention and detection of, and response to unauthorised removal, sabotage, unauthorised access, illegal transfer or other malicious acts involving nuclear or radiological material or their associated facilities.’ Ever since the end of Cold War, nuclear materials security has remained a concern for global security. However, with the increase in terrorist attacks not just in India especially, security of nuclear materials remains a priority. Therefore, India has made continued efforts to tighten its security on nuclear materials to prevent nuclear theft and radiological terrorism. Nuclear security is different from nuclear safety. Physical security is also a serious concern and India had been careful of the physical security of its nuclear materials. This is more so important since India is expanding its nuclear power capability to generate electricity for economic development. As India targets 60,000 MW of electricity production by 2030, it has a range of reactors to help it achieve its goal. These include indigenous Pressurised Heavy Water Reactors, now standardized at 700 MW per reactor Light Water Reactors, and the indigenous Fast Breeder Reactors that can generate more fuel for the future and enable the country to utilise its abundant thorium resource. Nuclear materials security can be enhanced through two important ways. One is through proliferation resistant technologies and diplomatic efforts to take non proliferation initiatives. The other is by developing technical means to prevent any leakage in nuclear materials in the hands of asymmetric organisations. New Delhi has already implemented IAEA Safeguards on their civilian nuclear installations. Moreover, the IAEA Additional Protocol has also been ratified by India in order to enhance its transparency of nuclear material and strengthen nuclear security. India is a party to the IAEA Conventions on Nuclear Safety and Security, and in particular the 1980 Convention on the Physical Protection of Nuclear Material and its amendment in 2005, Code of Conduct in Safety and Security of Radioactive Sources, 2006 which enables the country to provide for the highest international standards on nuclear and radiological safety and security. India's nuclear security approach is driven by five key components: Governance, Nuclear Security Practice and Culture, Institutions, Technology and International Cooperation. However, there is still scope for further improvements to strengthen nuclear materials and nuclear security. The NTI Report, ‘India’s improvement reflects its first contribution to the IAEA Nuclear Security Fund etc. in the future, India’s nuclear materials security conditions could be further improved by strengthening its laws and regulations for security and control of materials, particularly for control and accounting of materials, mitigating the insider threat, and for the physical security of materials during transport. India’s nuclear materials security conditions also remain adversely affected due to its continued increase in its quantities of nuclear material, and high levels of corruption among public officials.’ This paper would study briefly the progress made by India in nuclear and nuclear material security and the step ahead for India to further strengthen this.

Keywords: India, nuclear security, nuclear materials, non proliferation

Procedia PDF Downloads 331
1341 Single Cell Sorter Driven by Resonance Vibration of Cell Culture Substrate

Authors: Misa Nakao, Yuta Kurashina, Chikahiro Imashiro, Kenjiro Takemura

Abstract:

The Research Goal: With the growing demand for regenerative medicine, an effective mass cell culture process is required. In a repetitive subculture process for proliferating cells, preparing single cell suspension which does not contain any cell aggregates is highly required because cell aggregates often raise various undesirable phenomena, e.g., apoptosis and decrease of cell proliferation. Since cell aggregates often occur in cell suspension during conventional subculture processes, this study proposes a single cell sorter driven by a resonance vibration of a cell culture substrate. The Method and the Result: The single cell sorter is simply composed of a cell culture substrate and a glass pipe vertically placed against the cell culture substrate with a certain gap corresponding to a cell diameter. The cell culture substrate is made of biocompatible stainless steel with a piezoelectric ceramic disk glued to the bottom side. Applying AC voltage to the piezoelectric ceramic disk, an out-of-plane resonance vibration with a single nodal circle of the cell culture substrate can be excited at 5.5 kHz. By doing so, acoustic radiation force is emitted, and then cell suspension containing only single cells is pumped into the pipe and collected. This single cell sorter is effective to collect single cells selectively in spite of its quite simple structure. We collected C2C12 myoblast cell suspension by the single cell sorter with the vibration amplitude of 12 µmp-p and evaluated the ratio of single cells in number against the entire cells in the suspension. Additionally, we cultured the collected cells for 72 hrs and measured the number of cells after the cultivation in order to evaluate their proliferation. As a control sample, we also collected cell suspension by conventional pipetting, and evaluated the ratio of single cells and the number of cells after the 72-hour cultivation. The ratio of single cells in the cell suspension collected by the single cell sorter was 98.2%. This ratio was 9.6% higher than that collected by conventional pipetting (statistically significant). Moreover, the number of cells cultured for 72 hrs after the collection by the single cell sorter yielded statistically more cells than that collected by pipetting, resulting in a 13.6% increase in proliferated cells. These results suggest that the cell suspension collected by the single cell sorter driven by the resonance vibration hardly contains cell aggregates whose diameter is larger than the gap between the cell culture substrate and the pipe. Consequently, the cell suspension collected by the single cell sorter maintains high cell proliferation. Conclusions: In this study, we developed a single cell sorter capable of sorting and pumping single cells by a resonance vibration of a cell culture substrate. The experimental results show the single cell sorter collects single cell suspension which hardly contains cell aggregates. Furthermore, the collected cells show higher proliferation than that of cells collected by conventional pipetting. This means the resonance vibration of the cell culture substrate can benefit us with the increase in efficiency of mass cell culture process for clinical applications.

Keywords: acoustic radiation force, cell proliferation, regenerative medicine, resonance vibration, single cell sorter

Procedia PDF Downloads 244
1340 Lessons from Patients Expired due to Severe Head Injuries Treated in Intensive Care Unit of Lady Reading Hospital Peshawar

Authors: Mumtaz Ali, Hamzullah Khan, Khalid Khanzada, Shahid Ayub, Aurangzeb Wazir

Abstract:

Objective: To analyse the death of patients treated in neuro-surgical ICU for severe head injuries from different perspectives. The evaluation of the data so obtained to help improve the health care delivery to this group of patients in ICU. Study Design: It is a descriptive study based on retrospective analysis of patients presenting to neuro-surgical ICU in Lady Reading Hospital, Peshawar. Study Duration: It covered the period between 1st January 2009 to 31st December 2009. Material and Methods: The Clinical record of all the patients presenting with the clinical radiological and surgical features of severe head injuries, who expired in neuro-surgical ICU was collected. A separate proforma which mentioned age, sex, time of arrival and death, causes of head injuries, the radiological features, the clinical parameters, the surgical and non surgical treatment given was used. The average duration of stay and the demographic and domiciliary representation of these patients was noted. The record was analyzed accordingly for discussion and recommendations. Results: Out of the total 112 (n-112) patients who expired in one year in the neuro-surgical ICU the young adults made up the majority 64 (57.14%) followed by children, 34 (30.35%) and then the elderly age group: 10 (8.92%). Road traffic accidents were the major cause of presentation, 75 (66.96%) followed by history of fall; 23 (20.53%) and then the fire arm injuries; 13 (11.60%). The predominant CT scan features of these patients on presentation was cerebral edema, and midline shift (diffuse neuronal injuries). 46 (41.07%) followed by cerebral contusions. 28 (25%). The correctable surgical causes were present only in 18 patients (16.07%) and the majority 94 (83.92%) were given conservative management. Of the 69 (n=69) patients in which CT scan was repeated; 62 (89.85%) showed worsening of the initial CT scan abnormalities while in 7 cases (10.14%) the features were static. Among the non surgical cases both ventilatory therapy in 7 (6.25%) and tracheostomy in 39 (34.82%) failed to change the outcome. The maximum stay in the neuro ICU leading upto the death was 48 hours in 35 (31.25%) cases followed by 31 (27.67%) cases in 24 hours; 24 (21.42%) in one week and 16 (14.28%) in 72 hours. Only 6 (5.35%) patients survived more than a week. Patients were received from almost all the districts of NWFP except. The Hazara division. There were some Afghan refugees as well. Conclusion: Mortality following the head injuries is alarmingly high despite repeated claims about the professional and administrative improvement. Even places like ICU could not change the out come according to the desired aims and objectives in the present set up. A rethinking is needed both at the individual and institutional level among the concerned quarters with a clear aim at the more scientific grounds. Only then one can achieve the desired results.

Keywords: Glasgow Coma Scale, pediatrics, geriatrics, Peshawar

Procedia PDF Downloads 333
1339 Role of Indigenous Peoples in Climate Change

Authors: Neelam Kadyan, Pratima Ranga, Yogender

Abstract:

Indigenous people are the One who are affected by the climate change the most, although there have contributed little to its causes. This is largely a result of their historic dependence on local biological diversity, ecosystem services and cultural landscapes as a source of their sustenance and well-being. Comprising only four percent of the world’s population they utilize 22 percent of the world’s land surface. Despite their high exposure-sensitivity indigenous peoples and local communities are actively responding to changing climatic conditions and have demonstrated their resourcefulness and resilience in the face of climate change. Traditional Indigenous territories encompass up to 22 percent of the world’s land surface and they coincide with areas that hold 80 percent of the planet’s biodiversity. Also, the greatest diversity of indigenous groups coincides with the world’s largest tropical forest wilderness areas in the Americas (including Amazon), Africa, and Asia, and 11 percent of world forest lands are legally owned by Indigenous Peoples and communities. This convergence of biodiversity-significant areas and indigenous territories presents an enormous opportunity to expand efforts to conserve biodiversity beyond parks, which tend to benefit from most of the funding for biodiversity conservation. Tapping on Ancestral Knowledge Indigenous Peoples are carriers of ancestral knowledge and wisdom about this biodiversity. Their effective participation in biodiversity conservation programs as experts in protecting and managing biodiversity and natural resources would result in more comprehensive and cost effective conservation and management of biodiversity worldwide. Addressing the Climate Change Agenda Indigenous Peoples has played a key role in climate change mitigation and adaptation. The territories of indigenous groups who have been given the rights to their lands have been better conserved than the adjacent lands (i.e., Brazil, Colombia, Nicaragua, etc.). Preserving large extensions of forests would not only support the climate change objectives, but it would respect the rights of Indigenous Peoples and conserve biodiversity as well. A climate change agenda fully involving Indigenous Peoples has many more benefits than if only government and/or the private sector are involved. Indigenous peoples are some of the most vulnerable groups to the negative effects of climate change. Also, they are a source of knowledge to the many solutions that will be needed to avoid or ameliorate those effects. For example, ancestral territories often provide excellent examples of a landscape design that can resist the negatives effects of climate change. Over the millennia, Indigenous Peoples have developed adaptation models to climate change. They have also developed genetic varieties of medicinal and useful plants and animal breeds with a wider natural range of resistance to climatic and ecological variability.

Keywords: ancestral knowledge, cost effective conservation, management, indigenous peoples, climate change

Procedia PDF Downloads 661
1338 Determine Causal Factors Affecting the Responsiveness and Productivity of Non-Governmental Universities

Authors: Davoud Maleki

Abstract:

Today, education and investment in human capital is a long-term investment without which the economy will be stagnant Stayed. Higher education represents a type of investment in human resources by providing and improving knowledge, skills and Attitudes help economic development. Providing efficient human resources by increasing the efficiency and productivity of people and on the other hand with Expanding the boundaries of knowledge and technology and promoting technology such as the responsibility of training human resources and increasing productivity and efficiency in High specialized levels are the responsibility of universities. Therefore, the university plays an infrastructural role in economic development and growth because education by creating skills and expertise in people and improving their ability.In recent decades, Iran's higher education system has been faced with many problems, therefore, scholars have looked for it is to identify and validate the causal factors affecting the responsiveness and productivity of non-governmental universities. The data in the qualitative part is the result of semi-structured interviews with 25 senior and middle managers working in the units It was Islamic Azad University of Tehran province, which was selected by theoretical sampling method. In data analysis, stepwise method and Analytical techniques of Strauss and Corbin (1992) were used. After determining the central category (answering for the sake of the beneficiaries) and using it in order to bring the categories, expressions and ideas that express the relationships between the main categories and In the end, six main categories were identified as causal factors affecting the university's responsiveness and productivity.They are: 1- Scientism 2- Human resources 3- Creating motivation in the university 4- Development based on needs assessment 5- Teaching process and Learning 6- University quality evaluation. In order to validate the response model obtained from the qualitative stage, a questionnaire The questionnaire was prepared and the answers of 146 students of Master's degree and Doctorate of Islamic Azad University located in Tehran province were received. Quantitative data in the form of descriptive data analysis, first and second stage factor analysis using SPSS and Amos23 software were analyzed. The findings of the research indicated the relationship between the central category and the causal factors affecting the response The results of the model test in the quantitative stage confirmed the generality of the conceptual model.

Keywords: accountability, productivity, non-governmental, universities, foundation data theory

Procedia PDF Downloads 37
1337 Effect of Packaging Material and Water-Based Solutions on Performance of Radio Frequency Identification for Food Packaging Applications

Authors: Amelia Frickey, Timothy (TJ) Sheridan, Angelica Rossi, Bahar Aliakbarian

Abstract:

The growth of large food supply chains demanded improved end-to-end traceability of food products, which has led to companies being increasingly interested in using smart technologies such as Radio Frequency Identification (RFID)-enabled packaging to track items. As technology is being widely used, there are several technological or economic issues that should be overcome to facilitate the adoption of this track-and-trace technology. One of the technological challenges of RFID technology is its sensitivity to different environmental form factors, including packaging materials and the content of the packaging. Although researchers have assessed the performance loss due to the proximity of water and aqueous solutions, there is still the need to further investigate the impacts of food products on the reading range of RFID tags. However, to the best of our knowledge, there are not enough studies to determine the correlation between RFID tag performance and food beverages properties. The goal of this project was to investigate the effect of the solution properties (pH and conductivity) and different packaging materials filled with food-like water-based solutions on the performance of an RFID tag. Three commercially available ultra high-frequency RFID tags were placed on three different bottles and filled with different concentrations of water-based solutions, including sodium chloride, citric acid, sucrose, and ethanol. Transparent glass, Polyethylneterephtalate (PET), and Tetrapak® were used as the packaging materials commonly used in the beverage industries. Tag readability (Theoretical Read Range, TRR) and sensitivity (Power on Tag Forward, PoF) were determined using an anechoic chamber. First, the best place to attach the tag for each packaging material was investigated using empty and water-filled bottles. Then, the bottles were filled with the food-like solutions and tested with the three different tags and the PoF and TRR at the fixed frequency of 915MHz. In parallel, the pH and conductivity of solutions were measured. The best-performing tag was then selected to test the bottles filled with wine, orange, and apple juice. Despite various solutions altering the performance of each tag, the change in tag performance had no correlation with the pH or conductivity of the solution. Additionally, packaging material played a significant role in tag performance. Each tag tested performed optimally under different conditions. This study is the first part of comprehensive research to determine the regression model for the prediction of tag performance behavior based on the packaging material and the content. More investigations, including more tags and food products, are needed to be able to develop a robust regression model. The results of this study can be used by RFID tag manufacturers to design suitable tags for specific products with similar properties.

Keywords: smart food packaging, supply chain management, food waste, radio frequency identification

Procedia PDF Downloads 96
1336 Degradation Kinetics of Cardiovascular Implants Employing Full Blood and Extra-Corporeal Circulation Principles: Mimicking the Human Circulation In vitro

Authors: Sara R. Knigge, Sugat R. Tuladhar, Hans-Klaus HöFfler, Tobias Schilling, Tim Kaufeld, Axel Haverich

Abstract:

Tissue engineered (TE) heart valves based on degradable electrospun fiber scaffold represent a promising approach to overcome the known limitations of mechanical or biological prostheses. But the mechanical stress in the high-pressure system of the human circulation is a severe challenge for the delicate materials. Hence, the prediction of the scaffolds` in vivo degradation kinetics must be as accurate as possible to prevent fatal events in future animal or even clinical trials. Therefore, this study investigates whether long-term testing in full blood provides more meaningful results regarding the degradation behavior than conventional tests in simulated body fluids (SBF) or Phosphate Buffered Saline (PBS). Fiber mats were produced from a polycaprolactone (PCL)/tetrafluoroethylene solution by electrospinning. The morphology of the fiber mats was characterized via scanning electron microscopy (SEM). A maximum physiological degradation environment utilizing a test set-up with porcine full blood was established. The set-up consists of a reaction vessel, an oxygenator unit, and a roller pump. The blood parameters (pO2, pCO2, temperature, and pH) were monitored with an online test system. All tests were also carried out in the test circuit with SBF and PBS to compare conventional degradation media with the novel full blood setting. The polymer's degradation is quantified by SEM picture analysis, differential scanning calorimetry (DSC), and Raman spectroscopy. Tensile and cyclic loading tests were performed to evaluate the mechanical integrity of the scaffold. Preliminary results indicate that PCL degraded slower in full blood than in SBF and PBS. The uptake of water is more pronounced in the full blood group. Also, PCL preserved its mechanical integrity longer when degraded in full blood. Protein absorption increased during the degradation process. Red blood cells, platelets, and their aggregates adhered on the PCL. Presumably, the degradation led to a more hydrophilic polymeric surface which promoted the protein adsorption and the blood cell adhesion. Testing degradable implants in full blood allows for developing more reliable scaffold materials in the future. Material tests in small and large animal trials thereby can be focused on testing candidates that have proven to function well in an in-vivo-like setting.

Keywords: Electrospun scaffold, full blood degradation test, long-term polymer degradation, tissue engineered aortic heart valve

Procedia PDF Downloads 129
1335 Molecular Characterization of Arginine Sensing Response in Unravelling Host-Pathogen Interactions in Leishmania

Authors: Evanka Madan, Madhu Puri, Dan Zilberstein, Rohini Muthuswami, Rentala Madhubala

Abstract:

The extensive interaction between the host and pathogen metabolic networks decidedly shapes the outcome of infection. Utilization of arginine by the host and pathogen is critical for determining the outcome of pathogenic infection. Infections with L. donovani, an intracellular parasite, will lead to an extensive competition of arginine between the host and the parasite donovani infection. One of the major amino acid (AA) sensing signaling pathways in mammalian cells are the mammalian target of rapamycin complex I (mTORC1) pathway. mTORC1, as a sensor of nutrient, controls numerous metabolic pathways. Arginine is critical for mTORC1 activation. SLC38A9 is the arginine sensor for the mTORC1, being activated during arginine sufficiency. L. donovani transport arginine via a high-affinity transporter (LdAAP3) that is rapidly up-regulated by arginine deficiency response (ADR) in intracellular amastigotes. This study, to author’s best knowledge, investigates the interaction between two arginine sensing systems that act in the same compartment, the lysosome. One is important for macrophage defense, and the other is essential for pathogen virulence. We hypothesize that the latter modulates lysosome arginine to prevent host defense response. The work presented here identifies an upstream regulatory role of LdAAP3 in regulating the expression of SLC38A9-mTORC1 pathway, and consequently, their function in L. donovani infected THP-1 cells cultured in 0.1 mM and 1.5 mM arginine. It was found that in physiological levels of arginine (0.1 mM), infecting THP-1 with Leishmania leads to increased levels of SLC38A9 and mTORC1 via an increase in the expression of RagA. However, the reversal was observed with LdAAP3 mutants, reflecting the positive regulatory role of LdAAP3 on the host SLC38A9. At the molecular level, upon infection, mTORC1 and RagA were found to be activated at the surface of phagolysosomes which was found to form a complex with phagolysosomal localized SLC38A9. To reveal the relevance of SLC38A9 under physiological levels of arginine, endogenous SLC38A9 was depleted and a substantial reduction in the expression of host mTORC1, its downstream active substrate, p-P70S6K1 and parasite LdAAP3, was observed, thereby showing that silencing SLC38A9 suppresses ADR. In brief, to author’s best knowledge, these results reveal an upstream regulatory role of LdAAP3 in manipulating SLC38A9 arginine sensing in host macrophages. Our study indicates that intra-macrophage survival of L. donovani depends on the availability and transport of extracellular arginine. An understanding of the sensing pathway of both parasite and host will open a new perspective on the molecular mechanism of host-parasite interaction and consequently, as a treatment for Leishmaniasis.

Keywords: arginine sensing, LdAAP3, L. donovani, mTORC1, SLC38A9, THP-1

Procedia PDF Downloads 102
1334 Fluoride Contamination and Effects on Crops in North 24 Parganas, West Bengal, India

Authors: Rajkumar Ghosh

Abstract:

Fluoride contamination in water and its subsequent impact on agricultural practices is a growing concern in various regions worldwide, including North 24 Parganas, West Bengal, India. This study aimed to investigate the extent of fluoride contamination in the region's water sources and evaluate its effects on crop production and quality. A comprehensive survey of water sources, including wells, ponds, and rivers, was conducted to assess the fluoride levels in North 24 Parganas. Water samples were collected and analyzed using standard methods, and the fluoride concentration was determined. The findings revealed significant fluoride contamination in the water sources, surpassing the permissible limits recommended by national and international standards. To assess the effects of fluoride contamination on crops, field experiments were carried out in selected agricultural areas. Various crops commonly cultivated in the region, such as paddy, wheat, vegetables, and fruits, were examined for their growth, yield, and nutritional quality parameters. Additionally, soil samples were collected from the study sites to analyse the fluoride levels and their potential impact on soil health. The results demonstrated the adverse effects of fluoride contamination on crop growth and yield. Reduced plant height, stunted root development, decreased biomass accumulation, and diminished crop productivity were observed in fluoride-affected areas compared to uncontaminated control sites. Furthermore, the nutritional composition of crops, including micronutrients and mineral content, was significantly altered under high fluoride exposure, leading to potential health risks for consumers. The study also assessed the impact of fluoride on soil quality and found a negative correlation between fluoride concentration and soil health indicators, such as pH, organic matter content, and nutrient availability. These findings emphasize the need for sustainable soil management practices to mitigate the harmful effects of fluoride contamination and maintain agricultural productivity. Overall, this study highlights the alarming issue of fluoride contamination in water sources and its detrimental effects on crop production and quality in North 24 Parganas, West Bengal, India. The findings underscore the urgency for implementing appropriate water treatment measures, promoting awareness among farmers and local communities, and adopting sustainable agricultural practices to mitigate fluoride contamination and safeguard the region's agricultural ecosystem.

Keywords: agricultural ecosystem, water treatment, sustainable agricultural, fluoride contamination

Procedia PDF Downloads 58
1333 HRCT of the Chest and the Role of Artificial Intelligence in the Evaluation of Patients with COVID-19

Authors: Parisa Mansour

Abstract:

Introduction: Early diagnosis of coronavirus disease (COVID-19) is extremely important to isolate and treat patients in time, thus preventing the spread of the disease, improving prognosis and reducing mortality. High-resolution computed tomography (HRCT) chest imaging and artificial intelligence (AI)-based analysis of HRCT chest images can play a central role in the treatment of patients with COVID-19. Objective: To investigate different chest HRCT findings in different stages of COVID-19 pneumonia and to evaluate the potential role of artificial intelligence in the quantitative assessment of lung parenchymal involvement in COVID-19 pneumonia. Materials and Methods: This retrospective observational study was conducted between May 1, 2020 and August 13, 2020. The study included 2169 patients with COVID-19 who underwent chest HRCT. HRCT images showed the presence and distribution of lesions such as: ground glass opacity (GGO), compaction, and any special patterns such as septal thickening, inverted halo, mark, etc. HRCT findings of the breast at different stages of the disease (early: andlt) 5 days, intermediate: 6-10 days and late stage: >10 days). A CT severity score (CTSS) was calculated based on the extent of lung involvement on HRCT, which was then correlated with clinical disease severity. Use of artificial intelligence; Analysis of CT pneumonia and quot; An algorithm was used to quantify the extent of pulmonary involvement by calculating the percentage of pulmonary opacity (PO) and gross opacity (PHO). Depending on the type of variables, statistically significant tests such as chi-square, analysis of variance (ANOVA) and post hoc tests were applied when appropriate. Results: Radiological findings were observed in HRCT chest in 1438 patients. A typical pattern of COVID-19 pneumonia, i.e., bilateral peripheral GGO with or without consolidation, was observed in 846 patients. About 294 asymptomatic patients were radiologically positive. Chest HRCT in the early stages of the disease mostly showed GGO. The late stage was indicated by such features as retinal enlargement, thickening and the presence of fibrous bands. Approximately 91.3% of cases with a CTSS = 7 were asymptomatic or clinically mild, while 81.2% of cases with a score = 15 were clinically severe. Mean PO and PHO (30.1 ± 28.0 and 8.4 ± 10.4, respectively) were significantly higher in the clinically severe categories. Conclusion: Because COVID-19 pneumonia progresses rapidly, radiologists and physicians should become familiar with typical TC chest findings to treat patients early, ultimately improving prognosis and reducing mortality. Artificial intelligence can be a valuable tool in treating patients with COVID-19.

Keywords: chest, HRCT, covid-19, artificial intelligence, chest HRCT

Procedia PDF Downloads 48
1332 Chatbots and the Future of Globalization: Implications of Businesses and Consumers

Authors: Shoury Gupta

Abstract:

Chatbots are a rapidly growing technological trend that has revolutionized the way businesses interact with their customers. With the advancements in artificial intelligence, chatbots can now mimic human-like conversations and provide instant and efficient responses to customer inquiries. In this research paper, we aim to explore the implications of chatbots on the future of globalization for both businesses and consumers. The paper begins by providing an overview of the current state of chatbots in the global market and their growth potential in the future. The focus is on how chatbots have become a valuable tool for businesses looking to expand their global reach, especially in areas with high population density and language barriers. With chatbots, businesses can engage with customers in different languages and provide 24/7 customer service support, creating a more accessible and convenient customer experience. The paper then examines the impact of chatbots on cross-cultural communication and how they can help bridge communication gaps between businesses and consumers from different cultural backgrounds. Chatbots can potentially facilitate cross-cultural communication by offering real-time translations, voice recognition, and other innovative features that can help users communicate effectively across different languages and cultures. By providing more accessible and inclusive communication channels, chatbots can help businesses reach new markets and expand their customer base, making them more competitive in the global market. However, the paper also acknowledges that there are potential drawbacks associated with chatbots. For instance, chatbots may not be able to address complex customer inquiries that require human input. Additionally, chatbots may perpetuate biases if they are programmed with certain stereotypes or assumptions about different cultures. These drawbacks may have significant implications for businesses and consumers alike. To explore the implications of chatbots on the future of globalization in greater detail, the paper provides a thorough review of existing literature and case studies. The review covers topics such as the benefits of chatbots for businesses and consumers, the potential drawbacks of chatbots, and how businesses can mitigate any risks associated with chatbot use. The paper also discusses the ethical considerations associated with chatbot use, such as privacy concerns and the need to ensure that chatbots do not discriminate against certain groups of people. The ethical implications of chatbots are particularly important given the potential for chatbots to be used in sensitive areas such as healthcare and financial services. Overall, this research paper provides a comprehensive analysis of chatbots and their implications for the future of globalization. By exploring both the potential benefits and drawbacks of chatbot use, the paper aims to provide insights into how businesses and consumers can leverage this technology to achieve greater global reach and improve cross-cultural communication. Ultimately, the paper concludes that chatbots have the potential to be a powerful tool for businesses looking to expand their global footprint and improve their customer experience, but that care must be taken to mitigate any risks associated with their use.

Keywords: chatbots, conversational AI, globalization, businesses

Procedia PDF Downloads 78
1331 Mental Health Monitoring System as an Effort for Prevention and Handling of Psychological Problems in Students

Authors: Arif Tri Setyanto, Aditya Nanda Priyatama, Nugraha Arif Karyanta, Fadjri Kirana A., Afia Fitriani, Rini Setyowati, Moh.Abdul Hakim

Abstract:

The Basic Health Research Report by the Ministry of Health (2018) shows an increase in the prevalence of mental health disorders in the adolescent and early adult age ranges. Supporting this finding, data on the psychological examination of the student health service unit at one State University recorded 115 cases of moderate and severe health problems in the period 2016 - 2019. More specifically, the highest number of cases was experienced by clients in the age range of 21-23 years or equivalent, with the mid-semester stage towards the end. Based on the distribution of cases experienced and the disorder becomes a psychological problem experienced by students. A total of 29% or the equivalent of 33 students experienced anxiety disorders, 25% or 29 students experienced problems ranging from mild to severe, as well as other classifications of disorders experienced, including adjustment disorders, family problems, academics, mood disorders, self-concept disorders, personality disorders, cognitive disorders, and others such as trauma and sexual disorders. Various mental health disorders have a significant impact on the academic life of students, such as low GPA, exceeding the limit in college, dropping out, disruption of social life on campus, to suicide. Based on literature reviews and best practices from universities in various countries, one of the effective ways to prevent and treat student mental health disorders is to implement a mental health monitoring system in universities. This study uses a participatory action research approach, with a sample of 423 from a total population of 32,112 students. The scale used in this study is the Beck Depression Inventory (BDI) to measure depression and the Taylor Minnesota Anxiety Scale (TMAS) to measure anxiety levels. This study aims to (1) develop a digital-based health monitoring system for students' mental health situations in the mental health category. , dangers, or those who have mental disorders, especially indications of symptoms of depression and anxiety disorders, and (2) implementing a mental health monitoring system in universities at the beginning and end of each semester. The results of the analysis show that from 423 respondents, the main problems faced by all coursework, such as thesis and academic assignments. Based on the scoring and categorization of the Beck Depression Inventory (BDI), 191 students experienced symptoms of depression. A total of 24.35%, or 103 students experienced mild depression, 14.42% (61 students) had moderate depression, and 6.38% (27 students) experienced severe or extreme depression. Furthermore, as many as 80.38% (340 students) experienced anxiety in the high category. This article will review this review of the student mental health service system on campus.

Keywords: monitoring system, mental health, psychological problems, students

Procedia PDF Downloads 90
1330 Light-Controlled Gene Expression in Yeast

Authors: Peter. M. Kusen, Georg Wandrey, Christopher Probst, Dietrich Kohlheyer, Jochen Buchs, Jorg Pietruszkau

Abstract:

Light as a stimulus provides the capability to develop regulation techniques for customizable gene expression. A great advantage is the extremely flexible and accurate dosing that can be performed in a non invasive and sterile manner even for high throughput technologies. Therefore, light regulation in a multiwell microbioreactor system was realized providing the opportunity to control gene expression with outstanding complexity. A light-regulated gene expression system in Saccharomyces cerevisiae was designed applying the strategy of caged compounds. These compounds are photo-labile protected and therefore biologically inactive regulator molecules which can be reactivated by irradiation with certain light conditions. The “caging” of a repressor molecule which is consumed after deprotection was essential to create a flexible expression system. Thereby, gene expression could be temporally repressed by irradiation and subsequent release of the active repressor molecule. Afterwards, the repressor molecule is consumed by the yeast cells leading to reactivation of gene expression. A yeast strain harboring a construct with the corresponding repressible promoter in combination with a fluorescent marker protein was applied in a Photo-BioLector platform which allows individual irradiation as well as online fluorescence and growth detection. This device was used to precisely control the repression duration by adjusting the amount of released repressor via different irradiation times. With the presented screening platform the regulation of complex expression procedures was achieved by combination of several repression/derepression intervals. In particular, a stepwise increase of temporally-constant expression levels was demonstrated which could be used to study concentration dependent effects on cell functions. Also linear expression rates with variable slopes could be shown representing a possible solution for challenging protein productions, whereby excessive production rates lead to misfolding or intoxication. Finally, the very flexible regulation enabled accurate control over the expression induction, although we used a repressible promoter. Summing up, the continuous online regulation of gene expression has the potential to synchronize gene expression levels to optimize metabolic flux, artificial enzyme cascades, growth rates for co cultivations and many other applications addicted to complex expression regulation. The developed light-regulated expression platform represents an innovative screening approach to find optimization potential for production processes.

Keywords: caged-compounds, gene expression regulation, optogenetics, photo-labile protecting group

Procedia PDF Downloads 304
1329 Developing and Shake Table Testing of Semi-Active Hydraulic Damper as Active Interaction Control Device

Authors: Ming-Hsiang Shih, Wen-Pei Sung, Shih-Heng Tung

Abstract:

Semi-active control system for structure under excitation of earthquake provides with the characteristics of being adaptable and requiring low energy. DSHD (Displacement Semi-Active Hydraulic Damper) was developed by our research team. Shake table test results of this DSHD installed in full scale test structure demonstrated that this device brought its energy-dissipating performance into full play for test structure under excitation of earthquake. The objective of this research is to develop a new AIC (Active Interaction Control Device) and apply shake table test to perform its dissipation of energy capability. This new proposed AIC is converting an improved DSHD (Displacement Semi-Active Hydraulic Damper) to AIC with the addition of an accumulator. The main concept of this energy-dissipating AIC is to apply the interaction function of affiliated structure (sub-structure) and protected structure (main structure) to transfer the input seismic force into sub-structure to reduce the structural deformation of main structure. This concept is tested using full-scale multi-degree of freedoms test structure, installed with this proposed AIC subjected to external forces of various magnitudes, for examining the shock absorption influence of predictive control, stiffness of sub-structure, synchronous control, non-synchronous control and insufficient control position. The test results confirm: (1) this developed device is capable of diminishing the structural displacement and acceleration response effectively; (2) the shock absorption of low precision of semi-active control method did twice as much seismic proof efficacy as that of passive control method; (3) active control method may not exert a negative influence of amplifying acceleration response of structure; (4) this AIC comes into being time-delay problem. It is the same problem of ordinary active control method. The proposed predictive control method can overcome this defect; (5) condition switch is an important characteristics of control type. The test results show that synchronism control is very easy to control and avoid stirring high frequency response. This laboratory results confirm that the device developed in this research is capable of applying the mutual interaction between the subordinate structure and the main structure to be protected is capable of transforming the quake energy applied to the main structure to the subordinate structure so that the objective of minimizing the deformation of main structural can be achieved.

Keywords: DSHD (Displacement Semi-Active Hydraulic Damper), AIC (Active Interaction Control Device), shake table test, full scale structure test, sub-structure, main-structure

Procedia PDF Downloads 502
1328 AAV-Mediated Human Α-Synuclein Expression in a Rat Model of Parkinson's Disease –Further Characterization of PD Phenotype, Fine Motor Functional Effects as Well as Neurochemical and Neuropathological Changes over Time

Authors: R. Pussinen, V. Jankovic, U. Herzberg, M. Cerrada-Gimenez, T. Huhtala, A. Nurmi, T. Ahtoniemi

Abstract:

Targeted over-expression of human α-synuclein using viral-vector mediated gene delivery into the substantia nigra of rats and non-human primates has been reported to lead to dopaminergic cell loss and the formation of α-synuclein aggregates reminiscent of Lewy bodies. We have previously shown how AAV-mediated expression of α-synuclein is seen in the chronic phenotype of the rats over 16 week follow-up period. In the context of these findings, we attempted to further characterize this long term PD related functional and motor deficits as well as neurochemical and neuropathological changes in AAV-mediated α-synuclein transfection model in rats during chronic follow-up period. Different titers of recombinant AAV expressing human α-synuclein (A53T) were stereotaxically injected unilaterally into substantia nigra of Wistar rats. Rats were allowed to recover for 3 weeks prior to initial baseline behavioral testing with rotational asymmetry test, stepping test and cylinder test. A similar behavioral test battery was applied again at weeks 5, 9,12 and 15. In addition to traditionally used rat PD model tests, MotoRater test system, a high speed kinematic gait performance monitoring was applied during the follow-up period. Evaluation focused on animal gait between groups. Tremor analysis was performed on weeks 9, 12 and 15. In addition to behavioral end-points, neurochemical evaluation of dopamine and its metabolites were evaluated in striatum. Furthermore, integrity of the dopamine active transport (DAT) system was evaluated by using 123I- β-CIT and SPECT/CT imaging on weeks 3, 8 and 12 after AAV- α-synuclein transfection. Histopathology was examined from end-point samples at 3 or 12 weeks after AAV- α-synuclein transfection to evaluate dopaminergic cell viability and microglial (Iba-1) activation status in substantia nigra by using stereological analysis techniques. This study focused on the characterization and validation of previously published AAV- α-synuclein transfection model in rats but with the addition of novel end-points. We present the long term phenotype of AAV- α-synuclein transfected rats with traditionally used behavioral tests but also by using novel fine motor analysis techniques and tremor analysis which provide new insight to unilateral effects of AAV α-synuclein transfection. We also present data about neurochemical and neuropathological end-points for the dopaminergic system in the model and how well they correlate with behavioral phenotype.

Keywords: adeno-associated virus, alphasynuclein, animal model, Parkinson’s disease

Procedia PDF Downloads 279
1327 Development of an Appropriate Method for the Determination of Multiple Mycotoxins in Pork Processing Products by UHPLC-TCFLD

Authors: Jason Gica, Yi-Hsieng Samuel Wu, Deng-Jye Yang, Yi-Chen Chen

Abstract:

Mycotoxins, harmful secondary metabolites produced by certain fungi species, pose significant risks to animals and humans worldwide. Their stable properties lead to contamination during grain harvesting, transportation, and storage, as well as in processed food products. The prevalence of mycotoxin contamination has attracted significant attention due to its adverse impact on food safety and global trade. The secondary contamination pathway from animal products has been identified as an important route of exposure, posing health risks for livestock and humans consuming contaminated products. Pork, one of the highly consumed meat products in Taiwan according to the National Food Consumption Database, plays a critical role in the nation's diet and economy. Given its substantial consumption, pork processing products are a significant component of the food supply chain and a potential source of mycotoxin contamination. This study is paramount for formulating effective regulations and strategies to mitigate mycotoxin-related risks in the food supply chain. By establishing a reliable analytical method, this research contributes to safeguarding public health and enhancing the quality of pork processing products. The findings will serve as valuable guidance for policymakers, food industries, and consumers to ensure a safer food supply chain in the face of emerging mycotoxin challenges. An innovative and efficient analytical approach is proposed using Ultra-High Performance Liquid Chromatography coupled with Temperature Control Fluorescence Detector Light (UHPLC-TCFLD) to determine multiple mycotoxins in pork meat samples due to its exceptional capacity to detect multiple mycotoxins at the lowest levels of concentration, making it highly sensitive and reliable for comprehensive mycotoxin analysis. Additionally, its ability to simultaneously detect multiple mycotoxins in a single run significantly reduces the time and resources required for analysis, making it a cost-effective solution for monitoring mycotoxin contamination in pork processing products. The research aims to optimize the efficient mycotoxin QuEChERs extraction method and rigorously validate its accuracy and precision. The results will provide crucial insights into mycotoxin levels in pork processing products.

Keywords: multiple-mycotoxin analysis, pork processing products, QuEChERs, UHPLC-TCFLD, validation

Procedia PDF Downloads 43
1326 Raman Tweezers Spectroscopy Study of Size Dependent Silver Nanoparticles Toxicity on Erythrocytes

Authors: Surekha Barkur, Aseefhali Bankapur, Santhosh Chidangil

Abstract:

Raman Tweezers technique has become prevalent in single cell studies. This technique combines Raman spectroscopy which gives information about molecular vibrations, with optical tweezers which use a tightly focused laser beam for trapping the single cells. Thus Raman Tweezers enabled researchers analyze single cells and explore different applications. The applications of Raman Tweezers include studying blood cells, monitoring blood-related disorders, silver nanoparticle-induced stress, etc. There is increased interest in the toxic effect of nanoparticles with an increase in the various applications of nanoparticles. The interaction of these nanoparticles with the cells may vary with their size. We have studied the effect of silver nanoparticles of sizes 10nm, 40nm, and 100nm on erythrocytes using Raman Tweezers technique. Our aim was to investigate the size dependence of the nanoparticle effect on RBCs. We used 785nm laser (Starbright Diode Laser, Torsana Laser Tech, Denmark) for both trapping and Raman spectroscopic studies. 100 x oil immersion objectives with high numerical aperture (NA 1.3) is used to focus the laser beam into a sample cell. The back-scattered light is collected using the same microscope objective and focused into the spectrometer (Horiba Jobin Vyon iHR320 with 1200grooves/mm grating blazed at 750nm). Liquid nitrogen cooled CCD (Symphony CCD-1024x256-OPEN-1LS) was used for signal detection. Blood was drawn from healthy volunteers in vacutainer tubes and centrifuged to separate the blood components. 1.5 ml of silver nanoparticles was washed twice with distilled water leaving 0.1 ml silver nanoparticles in the bottom of the vial. The concentration of silver nanoparticles is 0.02mg/ml so the 0.03mg of nanoparticles will be present in the 0.1 ml nanoparticles obtained. The 25 ul of RBCs were diluted in 2 ml of PBS solution and then treated with 50 ul (0.015mg) of nanoparticles and incubated in CO2 incubator. Raman spectroscopic measurements were done after 24 hours and 48 hours of incubation. All the spectra were recorded with 10mW laser power (785nm diode laser), 60s of accumulation time and 2 accumulations. Major changes were observed in the peaks 565 cm-1, 1211 cm-1, 1224 cm-1, 1371 cm-1, 1638 cm-1. A decrease in intensity of 565 cm-1, increase in 1211 cm-1 with a reduction in 1224 cm-1, increase in intensity of 1371 cm-1 also peak disappearing at 1635 cm-1 indicates deoxygenation of hemoglobin. Nanoparticles with higher size were showing maximum spectral changes. Lesser changes observed in case of 10nm nanoparticle-treated erythrocyte spectra.

Keywords: erythrocytes, nanoparticle-induced toxicity, Raman tweezers, silver nanoparticles

Procedia PDF Downloads 274
1325 Developing Methodology of Constructing the Unified Action Plan for External and Internal Risks in University

Authors: Keiko Tamura, Munenari Inoguchi, Michiyo Tsuji

Abstract:

When disasters occur, in order to raise the speed of each decision making and response, it is common that delegation of authority is carried out. This tendency is particularly evident when the department or branch of the organization are separated by the physical distance from the main body; however, there are some issues to think about. If the department or branch is too dependent on the head office in the usual condition, they might feel lost in the disaster response operation when they are face to the situation. Avoiding this problem, an organization should decide how to delegate the authority and also who accept the responsibility for what before the disaster. This paper will discuss about the method which presents an approach for executing the delegation of authority process, implementing authorities, management by objectives, and preparedness plans and agreement. The paper will introduce the examples of efforts for the three research centers of Niigata University, Japan to arrange organizations capable of taking necessary actions for disaster response. Each center has a quality all its own. One is the center for carrying out the research in order to conserve the crested ibis (or Toki birds in Japanese), the endangered species. The another is the marine biological laboratory. The third one is very unique because of the old growth forests maintained as the experimental field. Those research centers are in the Sado Island, located off the coast of Niigata Prefecture, is Japan's second largest island after Okinawa and is known for possessing a rich history and culture. It takes 65 minutes jetfoil (high-speed ferry) ride to get to Sado Island from the mainland. The three centers are expected to be easily isolated at the time of a disaster. A sense of urgency encourages 3 centers in the process of organizational restructuring for enhancing resilience. The research team from the risk management headquarters offer those procedures; Step 1: Offer the hazard scenario based on the scientific evidence, Step 2: Design a risk management organization for disaster response function, Step 3: Conduct the participatory approach to make consensus about the overarching objectives, Step 4: Construct the unified operational action plan for 3 centers, Step 5: Simulate how to respond in each phase based on the understanding the various phases of the timeline of a disaster. Step 6: Document results to measure performance and facilitate corrective action. This paper shows the result of verifying the output and effects.

Keywords: delegation of authority, disaster response, risk management, unified command

Procedia PDF Downloads 107
1324 Carbon Capture and Storage Using Porous-Based Aerogel Materials

Authors: Rima Alfaraj, Abeer Alarawi, Murtadha AlTammar

Abstract:

The global energy landscape heavily relies on the oil and gas industry, which faces the critical challenge of reducing its carbon footprint. To address this issue, the integration of advanced materials like aerogels has emerged as a promising solution to enhance sustainability and environmental performance within the industry. This study thoroughly examines the application of aerogel-based technologies in the oil and gas sector, focusing particularly on their role in carbon capture and storage (CCS) initiatives. Aerogels, known for their exceptional properties, such as high surface area, low density, and customizable pore structure, have garnered attention for their potential in various CCS strategies. The review delves into various fabrication techniques utilized in producing aerogel materials, including sol-gel, supercritical drying, and freeze-drying methods, to assess their suitability for specific industry applications. Beyond fabrication, the practicality of aerogel materials in critical areas such as flow assurance, enhanced oil recovery, and thermal insulation is explored. The analysis spans a wide range of applications, from potential use in pipelines and equipment to subsea installations, offering valuable insights into the real-world implementation of aerogels in the oil and gas sector. The paper also investigates the adsorption and storage capabilities of aerogel-based sorbents, showcasing their effectiveness in capturing and storing carbon dioxide (CO₂) molecules. Optimization of pore size distribution and surface chemistry is examined to enhance the affinity and selectivity of aerogels towards CO₂, thereby improving the efficiency and capacity of CCS systems. Additionally, the study explores the potential of aerogel-based membranes for separating and purifying CO₂ from oil and gas streams, emphasizing their role in the carbon capture and utilization (CCU) value chain in the industry. Emerging trends and future perspectives in integrating aerogel-based technologies within the oil and gas sector are also discussed, including the development of hybrid aerogel composites and advanced functional components to further enhance material performance and versatility. By synthesizing the latest advancements and future directions in aerogel used for CCS applications in the oil and gas industry, this review offers a comprehensive understanding of how these innovative materials can aid in transitioning towards a more sustainable and environmentally conscious energy landscape. The insights provided can assist in strategic decision-making, drive technology development, and foster collaborations among academia, industry, and policymakers to promote the widespread adoption of aerogel-based solutions in the oil and gas sector.

Keywords: CCS, porous, carbon capture, oil and gas, sustainability

Procedia PDF Downloads 9
1323 Seeking Compatibility between Green Infrastructure and Recentralization: The Case of Greater Toronto Area

Authors: Sara Saboonian, Pierre Filion

Abstract:

There are two distinct planning approaches attempting to transform the North American suburb so as to reduce its adverse environmental impacts. The first one, the recentralization approach, proposes intensification, multi-functionality and more reliance on public transit and walking. It thus offers an alternative to the prevailing low-density, spatial specialization and automobile dependence of the North American suburb. The second approach concentrates instead on the provision of green infrastructure, which rely on natural systems rather than on highly engineered solutions to deal with the infrastructure needs of suburban areas. There are tensions between these two approaches as recentralization generally overlooks green infrastructure, which can be space consuming (as in the case of water retention systems), and thus conflicts with the intensification goals of recentralization. The research investigates three Canadian planned suburban centres in the Greater Toronto Area, where recentralization is the current planning practice, despite rising awareness of the benefits of green infrastructure. Methods include reviewing the literature on green infrastructure planning, a critical analysis of the Ontario provincial plans for recentralization, surveying residents’ preferences regarding alternative suburban development models, and interviewing officials who deal with the local planning of the three centres. The case studies expose the difficulties in creating planned suburban centres that accommodate green infrastructure while adhering to recentralization principles. Until now, planners have been mostly focussed on recentralization at the expense of green infrastructure. In this context, the frequent lack of compatibility between recentralization and the space requirements of green infrastructure explains the limited presence of such infrastructures in planned suburban centres. Finally, while much attention has been given in the planning discourse to the economic and lifestyle benefits of recentralization, much less has been made of the wide range of advantages of green infrastructure, which explains limited public mobilization over the development of green infrastructure networks. The paper will concentrate on ways of combining recentralization with green infrastructure strategies and identify the aspects of the two approaches that are most compatible with each other. The outcome of such blending will marry high density, public-transit oriented developments, which generate walkability and street-level animation, with the presence of green space, naturalized settings and reliance on renewable energy. The paper will advance a planning framework that will fuse green infrastructure with recentralization, thus ensuring the achievement of higher density and reduced reliance on the car along with the provision of critical ecosystem services throughout cities. This will support and enhance the objectives of both green infrastructure and recentralization.

Keywords: environmental-based planning, green infrastructure, multi-functionality, recentralization

Procedia PDF Downloads 117
1322 Development and Characterization of Novel Topical Formulation Containing Niacinamide

Authors: Sevdenur Onger, Ali Asram Sagiroglu

Abstract:

Hyperpigmentation is a cosmetically unappealing skin problem caused by an overabundance of melanin in the skin. Its pathophysiology is caused by melanocytes being exposed to paracrine melanogenic stimuli, which can upregulate melanogenesis-related enzymes (such as tyrosinase) and cause melanosome formation. Tyrosinase is linked to the development of melanosomes biochemically, and it is the main target of hyperpigmentation treatment. therefore, decreasing tyrosinase activity to reduce melanosomes has become the main target of hyperpigmentation treatment. Niacinamide (NA) is a natural chemical found in a variety of plants that is used as a skin-whitening ingredient in cosmetic formulations. NA decreases melanogenesis in the skin by inhibiting melanosome transfer from melanocytes to covering keratinocytes. Furthermore, NA protects the skin from reactive oxygen species and acts as a main barrier with the skin, reducing moisture loss by increasing ceramide and fatty acid synthesis. However, it is very difficult for hydrophilic compounds such as NA to penetrate deep into the skin. Furthermore, because of the nicotinic acid in NA, it is an irritant. As a result, we've concentrated on strategies to increase NA skin permeability while avoiding its irritating impacts. Since nanotechnology can affect drug penetration behavior by controlling the release and increasing the period of permanence on the skin, it can be a useful technique in the development of whitening formulations. Liposomes have become increasingly popular in the cosmetics industry in recent years due to benefits such as their lack of toxicity, high penetration ability in living skin layers, ability to increase skin moisture by forming a thin layer on the skin surface, and suitability for large-scale production. Therefore, liposomes containing NA were developed for this study. Different formulations were prepared by varying the amount of phospholipid and cholesterol and examined in terms of particle sizes, polydispersity index (PDI) and pH values. The pH values of the produced formulations were determined to be suitable with the pH value of the skin. Particle sizes were determined to be smaller than 250 nm and the particles were found to be of homogeneous size in the formulation (pdi<0.30). Despite the important advantages of liposomal systems, they have low viscosity and stability for topical use. For these reasons, in this study, liposomal cream formulations have been prepared for easy topical application of liposomal systems. As a result, liposomal cream formulations containing NA have been successfully prepared and characterized. Following the in-vitro release and ex-vivo diffusion studies to be conducted in the continuation of the study, it is planned to test the formulation that gives the most appropriate result on the volunteers after obtaining the approval of the ethics committee.

Keywords: delivery systems, hyperpigmentation, liposome, niacinamide

Procedia PDF Downloads 98
1321 Exploratory Tests of Crude Bacteriocins from Autochthonous Lactic Acid Bacteria against Food-Borne Pathogens and Spoilage Bacteria

Authors: M. Naimi, M. B. Khaled

Abstract:

The aim of the present work was to test in vitro inhibition of food pathogens and spoilage bacteria by crude bacteriocins from autochthonous lactic acid bacteria. Thirty autochthonous lactic acid bacteria isolated previously, belonging to the genera: Lactobacillus, Carnobacterium, Lactococcus, Vagococcus, Streptococcus, and Pediococcus, have been screened by an agar spot test and a well diffusion assay against Gram-positive and Gram-negative harmful bacteria: Bacillus cereus, Bacillus subtilis ATCC 6633, Escherichia coli ATCC 8739, Salmonella typhimurium ATCC 14028, Staphylococcus aureus ATCC 6538, and Pseudomonas aeruginosa under conditions means to reduce lactic acid and hydrogen peroxide effect to select bacteria with high bacteriocinogenic potential. Furthermore, crude bacteriocins semiquantification and heat sensitivity to different temperatures (80, 95, 110°C, and 121°C) were performed. Another exploratory test concerning the response of St. aureus ATCC 6538 to the presence of crude bacteriocins was realized. It has been observed by the agar spot test that fifteen candidates were active toward Gram-positive targets strains. The secondary screening demonstrated an antagonistic activity oriented only against St. aureus ATCC 6538, leading to the selection of five isolates: Lm14, Lm21, Lm23, Lm24, and Lm25 with a larger inhibition zone compared to the others. The ANOVA statistical analysis reveals a small variation of repeatability: Lm21: 0.56%, Lm23: 0%, Lm25: 1.67%, Lm14: 1.88%, Lm24: 2.14%. Conversely, slight variation was reported in terms of inhibition diameters: 9.58± 0.40, 9.83± 0.46, and 10.16± 0.24 8.5 ± 0.40 10 mm for, Lm21, Lm23, Lm25, Lm14and Lm24, indicating that the observed potential showed a heterogeneous distribution (BMS = 0.383, WMS = 0.117). The repeatability coefficient calculated displayed 7.35%. As for the bacteriocins semiquantification, the five samples exhibited production amounts about 4.16 for Lm21, Lm23, Lm25 and 2.08 AU/ml for Lm14, Lm24. Concerning the sensitivity the crude bacteriocins were fully insensitive to heat inactivation, until 121°C, they preserved the same inhibition diameter. As to, kinetic of growth , the µmax showed reductions in pathogens load for Lm21, Lm23, Lm25, Lm14, Lm24 of about 42.92%, 84.12%, 88.55%, 54.95%, 29.97% in the second trails. Inversely, this pathogen growth after five hours displayed differences of 79.45%, 12.64%, 11.82%, 87.88%, 85.66% in the second trails, compared to the control. This study showed potential inhibition to the growth of this food pathogen, suggesting the possibility to improve the hygienic food quality.

Keywords: exploratory test, lactic acid bacteria, crude bacteriocins, spoilage, pathogens

Procedia PDF Downloads 196
1320 Approach to Honey Volatiles' Profiling by Gas Chromatography and Mass Spectrometry

Authors: Igor Jerkovic

Abstract:

Biodiversity of flora provides many different nectar sources for the bees. Unifloral honeys possess distinctive flavours, mainly derived from their nectar sources (characteristic volatile organic components (VOCs)). Specific or nonspecific VOCs (chemical markers) could be used for unifloral honey characterisation as addition to the melissopalynologycal analysis. The main honey volatiles belong, in general, to three principal categories: terpenes, norisoprenoids, and benzene derivatives. Some of these substances have been described as characteristics of the floral source, and other compounds, like several alcohols, branched aldehydes, and furan derivatives, may be related to the microbial purity of honey processing and storage conditions. Selection of the extraction method for the honey volatiles profiling should consider that heating of the honey produce different artefacts and therefore conventional methods of VOCs isolation (such as hydrodistillation) cannot be applied for the honey. Two-way approach for the isolation of the honey VOCs was applied using headspace solid-phase microextraction (HS-SPME) and ultrasonic solvent extraction (USE). The extracts were analysed by gas chromatography and mass spectrometry (GC-MS). HS-SPME (with the fibers of different polarity such as polydimethylsiloxane/ divinylbenzene (PDMS/DVB) or divinylbenzene/carboxene/ polydimethylsiloxane (DVB/CAR/PDMS)) enabled isolation of high volatile headspace VOCs of the honey samples. Among them, some characteristic or specific compounds can be found such as 3,4-dihydro-3-oxoedulan (in Centaurea cyanus L. honey) or 1H-indole, methyl anthranilate, and cis-jasmone (in Citrus unshiu Marc. honey). USE with different solvents (mainly dichloromethane or the mixture pentane : diethyl ether 1 : 2 v/v) enabled isolation of less volatile and semi-volatile VOCs of the honey samples. Characteristic compounds from C. unshiu honey extracts were caffeine, 1H-indole, 1,3-dihydro-2H-indol-2-one, methyl anthranilate, and phenylacetonitrile. Sometimes, the selection of solvent sequence was useful for more complete profiling such as sequence I: pentane → diethyl ether or sequence II: pentane → pentane/diethyl ether (1:2, v/v) → dichloromethane). The extracts with diethyl ether contained hydroquinone and 4-hydroxybenzoic acid as the major compounds, while (E)-4-(r-1’,t-2’,c-4’-trihydroxy-2’,6’,6’-trimethylcyclo-hexyl)but-3-en-2-one predominated in dichloromethane extracts of Allium ursinum L. honey. With this two-way approach, it was possible to obtain a more detailed insight into the honey volatile and semi-volatile compounds and to minimize the risks of compound discrimination due to their partial extraction that is of significant importance for the complete honey profiling and identification of the chemical biomarkers that can complement the pollen analysis.

Keywords: honey chemical biomarkers, honey volatile compounds profiling, headspace solid-phase microextraction (HS-SPME), ultrasonic solvent extraction (USE)

Procedia PDF Downloads 183
1319 Tests for Zero Inflation in Count Data with Measurement Error in Covariates

Authors: Man-Yu Wong, Siyu Zhou, Zhiqiang Cao

Abstract:

In quality of life, health service utilization is an important determinant of medical resource expenditures on Colorectal cancer (CRC) care, a better understanding of the increased utilization of health services is essential for optimizing the allocation of healthcare resources to services and thus for enhancing the service quality, especially for high expenditure on CRC care like Hong Kong region. In assessing the association between the health-related quality of life (HRQOL) and health service utilization in patients with colorectal neoplasm, count data models can be used, which account for over dispersion or extra zero counts. In our data, the HRQOL evaluation is a self-reported measure obtained from a questionnaire completed by the patients, misreports and variations in the data are inevitable. Besides, there are more zero counts from the observed number of clinical consultations (observed frequency of zero counts = 206) than those from a Poisson distribution with mean equal to 1.33 (expected frequency of zero counts = 156). This suggests that excess of zero counts may exist. Therefore, we study tests for detecting zero-inflation in models with measurement error in covariates. Method: Under classical measurement error model, the approximate likelihood function for zero-inflation Poisson regression model can be obtained, then Approximate Maximum Likelihood Estimation(AMLE) can be derived accordingly, which is consistent and asymptotically normally distributed. By calculating score function and Fisher information based on AMLE, a score test is proposed to detect zero-inflation effect in ZIP model with measurement error. The proposed test follows asymptotically standard normal distribution under H0, and it is consistent with the test proposed for zero-inflation effect when there is no measurement error. Results: Simulation results show that empirical power of our proposed test is the highest among existing tests for zero-inflation in ZIP model with measurement error. In real data analysis, with or without considering measurement error in covariates, existing tests, and our proposed test all imply H0 should be rejected with P-value less than 0.001, i.e., zero-inflation effect is very significant, ZIP model is superior to Poisson model for analyzing this data. However, if measurement error in covariates is not considered, only one covariate is significant; if measurement error in covariates is considered, only another covariate is significant. Moreover, the direction of coefficient estimations for these two covariates is different in ZIP regression model with or without considering measurement error. Conclusion: In our study, compared to Poisson model, ZIP model should be chosen when assessing the association between condition-specific HRQOL and health service utilization in patients with colorectal neoplasm. and models taking measurement error into account will result in statistically more reliable and precise information.

Keywords: count data, measurement error, score test, zero inflation

Procedia PDF Downloads 267
1318 Generation of Roof Design Spectra Directly from Uniform Hazard Spectra

Authors: Amin Asgarian, Ghyslaine McClure

Abstract:

Proper seismic evaluation of Non-Structural Components (NSCs) mandates an accurate estimation of floor seismic demands (i.e. acceleration and displacement demands). Most of the current international codes incorporate empirical equations to calculate equivalent static seismic force for which NSCs and their anchorage system must be designed. These equations, in general, are functions of component mass and peak seismic acceleration to which NSCs are subjected to during the earthquake. However, recent studies have shown that these recommendations are suffered from several shortcomings such as neglecting the higher mode effect, tuning effect, NSCs damping effect, etc. which cause underestimation of the component seismic acceleration demand. This work is aimed to circumvent the aforementioned shortcomings of code provisions as well as improving them by proposing a simplified, practical, and yet accurate approach to generate acceleration Floor Design Spectra (FDS) directly from corresponding Uniform Hazard Spectra (UHS) (i.e. design spectra for structural components). A database of 27 Reinforced Concrete (RC) buildings in which Ambient Vibration Measurements (AVM) have been conducted. The database comprises 12 low-rise, 10 medium-rise, and 5 high-rise buildings all located in Montréal, Canada and designated as post-disaster buildings or emergency shelters. The buildings are subjected to a set of 20 compatible seismic records and Floor Response Spectra (FRS) in terms of pseudo acceleration are derived using the proposed approach for every floor of the building in both horizontal directions considering 4 different damping ratios of NSCs (i.e. 2, 5, 10, and 20% viscous damping). Several effective parameters on NSCs response are evaluated statistically. These parameters comprise NSCs damping ratios, tuning of NSCs natural period with one of the natural periods of supporting structure, higher modes of supporting structures, and location of NSCs. The entire spectral region is divided into three distinct segments namely short-period, fundamental period, and long period region. The derived roof floor response spectra for NSCs with 5% damping are compared with the 5% damping UHS and procedure are proposed to generate roof FDS for NSCs with 5% damping directly from 5% damped UHS in each spectral region. The generated FDS is a powerful, practical, and accurate tool for seismic design and assessment of acceleration-sensitive NSCs particularly in existing post-critical buildings which have to remain functional even after the earthquake and cannot tolerate any damage to NSCs.

Keywords: earthquake engineering, operational and functional components (OFCs), operational modal analysis (OMA), seismic assessment and design

Procedia PDF Downloads 223
1317 A World Map of Seabed Sediment Based on 50 Years of Knowledge

Authors: T. Garlan, I. Gabelotaud, S. Lucas, E. Marchès

Abstract:

Production of a global sedimentological seabed map has been initiated in 1995 to provide the necessary tool for searches of aircraft and boats lost at sea, to give sedimentary information for nautical charts, and to provide input data for acoustic propagation modelling. This original approach had already been initiated one century ago when the French hydrographic service and the University of Nancy had produced maps of the distribution of marine sediments of the French coasts and then sediment maps of the continental shelves of Europe and North America. The current map of the sediment of oceans presented was initiated with a UNESCO's general map of the deep ocean floor. This map was adapted using a unique sediment classification to present all types of sediments: from beaches to the deep seabed and from glacial deposits to tropical sediments. In order to allow good visualization and to be adapted to the different applications, only the granularity of sediments is represented. The published seabed maps are studied, if they present an interest, the nature of the seabed is extracted from them, the sediment classification is transcribed and the resulted map is integrated in the world map. Data come also from interpretations of Multibeam Echo Sounder (MES) imagery of large hydrographic surveys of deep-ocean. These allow a very high-quality mapping of areas that until then were represented as homogeneous. The third and principal source of data comes from the integration of regional maps produced specifically for this project. These regional maps are carried out using all the bathymetric and sedimentary data of a region. This step makes it possible to produce a regional synthesis map, with the realization of generalizations in the case of over-precise data. 86 regional maps of the Atlantic Ocean, the Mediterranean Sea, and the Indian Ocean have been produced and integrated into the world sedimentary map. This work is permanent and permits a digital version every two years, with the integration of some new maps. This article describes the choices made in terms of sediment classification, the scale of source data and the zonation of the variability of the quality. This map is the final step in a system comprising the Shom Sedimentary Database, enriched by more than one million punctual and surface items of data, and four series of coastal seabed maps at 1:10,000, 1:50,000, 1:200,000 and 1:1,000,000. This step by step approach makes it possible to take into account the progresses in knowledge made in the field of seabed characterization during the last decades. Thus, the arrival of new classification systems for seafloor has improved the recent seabed maps, and the compilation of these new maps with those previously published allows a gradual enrichment of the world sedimentary map. But there is still a lot of work to enhance some regions, which are still based on data acquired more than half a century ago.

Keywords: marine sedimentology, seabed map, sediment classification, world ocean

Procedia PDF Downloads 217