Search results for: smart camera networks
3495 Delivering Inclusive Growth through Information and Communication Technology: The Miracle of Internet of Everything
Authors: Olawale Johnson
Abstract:
The cry and agitation for the creation of equal opportunities is one of the major reasons behind the social menace countries of the world experience. As the poor, continue to demand for the dividends of economic growth, countries of the world are in a state of dilemma because, despite impressive growth figures, the poor are still far below the empowerment line. However, evidence from the Asian Tigers has proven that with the adoption and efficient utilization of information technology, a growth miracle is not far-fetched. With the mind-boggling pace of technological innovation, the need to ensure that the innovative products are all connected has become vital. Technologies that did not exist a few years ago have become vital equipment used to underlie every aspect of our economy from medicine to banking to sports. The need to connect things sensors, actuators and smart systems with the aim of ensuring person-to-object, object-to-object communications has promoted the need of internet of things. As developing countries struggle with delivering inclusiveness, the Internet of Everything is perceived to be the miracle that will deliver this in no time. This paper examines how the Asian Tigers have been able to promote inclusive growth through the Internet of Everything.Keywords: inclusive growth, internet of everything, innovation, embedded systems and smart technologies
Procedia PDF Downloads 3203494 Using Machine Learning to Classify Different Body Parts and Determine Healthiness
Authors: Zachary Pan
Abstract:
Our general mission is to solve the problem of classifying images into different body part types and deciding if each of them is healthy or not. However, for now, we will determine healthiness for only one-sixth of the body parts, specifically the chest. We will detect pneumonia in X-ray scans of those chest images. With this type of AI, doctors can use it as a second opinion when they are taking CT or X-ray scans of their patients. Another ad-vantage of using this machine learning classifier is that it has no human weaknesses like fatigue. The overall ap-proach to this problem is to split the problem into two parts: first, classify the image, then determine if it is healthy. In order to classify the image into a specific body part class, the body parts dataset must be split into test and training sets. We can then use many models, like neural networks or logistic regression models, and fit them using the training set. Now, using the test set, we can obtain a realistic accuracy the models will have on images in the real world since these testing images have never been seen by the models before. In order to increase this testing accuracy, we can also apply many complex algorithms to the models, like multiplicative weight update. For the second part of the problem, to determine if the body part is healthy, we can have another dataset consisting of healthy and non-healthy images of the specific body part and once again split that into the test and training sets. We then use another neural network to train on those training set images and use the testing set to figure out its accuracy. We will do this process only for the chest images. A major conclusion reached is that convolutional neural networks are the most reliable and accurate at image classification. In classifying the images, the logistic regression model, the neural network, neural networks with multiplicative weight update, neural networks with the black box algorithm, and the convolutional neural network achieved 96.83 percent accuracy, 97.33 percent accuracy, 97.83 percent accuracy, 96.67 percent accuracy, and 98.83 percent accuracy, respectively. On the other hand, the overall accuracy of the model that de-termines if the images are healthy or not is around 78.37 percent accuracy.Keywords: body part, healthcare, machine learning, neural networks
Procedia PDF Downloads 1033493 Performance Evaluation of Refinement Method for Wideband Two-Beams Formation
Authors: C. Bunsanit
Abstract:
This paper presents the refinement method for two beams formation of wideband smart antenna. The refinement method for weighting coefficients is based on Fully Spatial Signal Processing by taking Inverse Discrete Fourier Transform (IDFT), and its simulation results are presented using MATLAB. The radiation pattern is created by multiplying the incoming signal with real weights and then summing them together. These real weighting coefficients are computed by IDFT method; however, the range of weight values is relatively wide. Therefore, for reducing this range, the refinement method is used. The radiation pattern concerns with five input parameters to control. These parameters are maximum weighting coefficient, wideband signal, direction of mainbeam, beamwidth, and maximum of minor lobe level. Comparison of the obtained simulation results between using refinement method and taking only IDFT shows that the refinement method works well for wideband two beams formation.Keywords: fully spatial signal processing, beam forming, refinement method, smart antenna, weighting coefficient, wideband
Procedia PDF Downloads 2263492 A Feasibility Study of Crowdsourcing Data Collection for Facility Maintenance Management
Authors: Mohamed Bin Alhaj, Hexu Liu, Mohammed Sulaiman, Osama Abudayyeh
Abstract:
An effective facility maintenance management (FMM) system plays a crucial role in improving the quality of services and maintaining the facility in good condition. Current FMM heavily relies on the quality of the data collection function of the FMM systems, at times resulting in inefficient FMM decision-making. The new technology-based crowdsourcing provides great potential to improve the current FMM practices, especially in terms of timeliness and quality of data. This research aims to investigate the feasibility of using new technology-driven crowdsourcing for FMM and highlight its opportunities and challenges. A survey was carried out to understand the human, data, system, geospatial, and automation characteristics of crowdsourcing for an educational campus FMM via social networks. The survey results were analyzed to reveal the challenges and recommendations for the implementation of crowdsourcing for FMM. This research contributes to the body of knowledge by synthesizing the challenges and opportunities of using crowdsourcing for facility maintenance and providing a road map for applying crowdsourcing technology in FMM. In future work, a conceptual framework will be proposed to support data-driven FMM using social networks.Keywords: crowdsourcing, facility maintenance management, social networks
Procedia PDF Downloads 1733491 NANCY: Combining Adversarial Networks with Cycle-Consistency for Robust Multi-Modal Image Registration
Authors: Mirjana Ruppel, Rajendra Persad, Amit Bahl, Sanja Dogramadzi, Chris Melhuish, Lyndon Smith
Abstract:
Multimodal image registration is a profoundly complex task which is why deep learning has been used widely to address it in recent years. However, two main challenges remain: Firstly, the lack of ground truth data calls for an unsupervised learning approach, which leads to the second challenge of defining a feasible loss function that can compare two images of different modalities to judge their level of alignment. To avoid this issue altogether we implement a generative adversarial network consisting of two registration networks GAB, GBA and two discrimination networks DA, DB connected by spatial transformation layers. GAB learns to generate a deformation field which registers an image of the modality B to an image of the modality A. To do that, it uses the feedback of the discriminator DB which is learning to judge the quality of alignment of the registered image B. GBA and DA learn a mapping from modality A to modality B. Additionally, a cycle-consistency loss is implemented. For this, both registration networks are employed twice, therefore resulting in images ˆA, ˆB which were registered to ˜B, ˜A which were registered to the initial image pair A, B. Thus the resulting and initial images of the same modality can be easily compared. A dataset of liver CT and MRI was used to evaluate the quality of our approach and to compare it against learning and non-learning based registration algorithms. Our approach leads to dice scores of up to 0.80 ± 0.01 and is therefore comparable to and slightly more successful than algorithms like SimpleElastix and VoxelMorph.Keywords: cycle consistency, deformable multimodal image registration, deep learning, GAN
Procedia PDF Downloads 1313490 Smart Lean Manufacturing in the Context of Industry 4.0: A Case Study
Authors: M. Ramadan, B. Salah
Abstract:
This paper introduces a framework to digitalize lean manufacturing tools to enhance smart lean-based manufacturing environments or Lean 4.0 manufacturing systems. The paper discusses the integration between lean tools and the powerful features of recent real-time data capturing systems with the help of Information and Communication Technologies (ICT) to develop an intelligent real-time monitoring and controlling system of production operations concerning lean targets. This integration is represented in the Lean 4.0 system called Dynamic Value Stream Mapping (DVSM). Moreover, the paper introduces the practice of Radio Frequency Identification (RFID) and ICT to smartly support lean tools and practices during daily production runs to keep the lean system alive and effective. This work introduces a practical description of how the lean method tools 5S, standardized work, and poka-yoke can be digitalized and smartly monitored and controlled through DVSM. A framework of the three tools has been discussed and put into practice in a German switchgear manufacturer.Keywords: lean manufacturing, Industry 4.0, radio frequency identification, value stream mapping
Procedia PDF Downloads 2293489 Simulation Of A Renal Phantom Using the MAG 3
Authors: Ati Moncef
Abstract:
We describe in this paper the results of a phantom of dynamics renal with MAG3. Our phantom consisted of (tow shaped of kidneys, 1 liver). These phantoms were scanned with static and dynamic protocols and compared with clinical data. in a normal conditions we use our phantoms it's possible to acquire a renal images when we can be compared with clinical scintigraphy. In conclusion, Renal phantom also can use in the quality control of a renal scintigraphy.Keywords: Renal scintigraphy, MAG3, Nuclear medicine, Gamma Camera.
Procedia PDF Downloads 4013488 A Study on Using Network Coding for Packet Transmissions in Wireless Sensor Networks
Authors: Rei-Heng Cheng, Wen-Pinn Fang
Abstract:
A wireless sensor network (WSN) is composed by a large number of sensors and one or a few base stations, where the sensor is responsible for detecting specific event information, which is sent back to the base station(s). However, how to save electricity consumption to extend the network lifetime is a problem that cannot be ignored in the wireless sensor networks. Since the sensor network is used to monitor a region or specific events, how the information can be reliably sent back to the base station is surly important. Network coding technique is often used to enhance the reliability of the network transmission. When a node needs to send out M data packets, it encodes these data with redundant data and sends out totally M + R packets. If the receiver can get any M packets out from these M + R packets, it can decode and get the original M data packets. To transmit redundant packets will certainly result in the excess energy consumption. This paper will explore relationship between the quality of wireless transmission and the number of redundant packets. Hopefully, each sensor can overhear the nearby transmissions, learn the wireless transmission quality around it, and dynamically determine the number of redundant packets used in network coding.Keywords: energy consumption, network coding, transmission reliability, wireless sensor networks
Procedia PDF Downloads 3913487 Urban Retrofitting Application Based on Social-Media to Model the Malioboro Smart Central Business Design through Statistical Regression Approach
Authors: Muhammad Hardyan Prastyanto, Aisah Azhari Marwangi, Yulinda Rizky Pratiwi
Abstract:
Globalization has become a driving force for the current technological developments. The presence of the Virtual Space provides opportunities for people to self-actualization through access to a wider world, quickly and easily. Cities that are part of the existence of life, witness the history of civilization over time, also has been the major object to upgrading on technological sector. A smart city is one where the government and citizenry are using the best available means, including ICT, to achieve their shared goals. This often includes economic development, environmental sustainability, and improved quality of life for citizens. Thus theory is the basis for research of this study. This study aimed to know the implementation of the Urban Retrofitting at Malioboro area based on Information and Communication Technologies. The method of this study is by reviewing the effectiveness of the E-commerce uses as a major system to identification the Malioboro Smart Central Business District. By using a significance level of 5 %, it can be concluded that addresses have a significant influence on the ratings obtained, namely regarding the location of the hotel establishment. But despite the use of the website does not have a significant influence on the rating of the hotel, using the website still has influence significantly on the rating, because the p -value (Sig.) of the variable website is not so much different from the significance level determined by the researcher. In the interpretation, if a hotel is located on the Pasar Kembang streets and not to use the website, so the hotel is likely to have a rating of the constant value which is 3.183. However, if a hotel located on the Sosrowijayan streets, so the hotel rating will be increased by 0,302. Then if a hotel has been using a website, so the hotel rating will increase by 0,264. It is possible to conclude the effectiveness of ICT’s (Website) uses and location to identification the urban retrofitting through increasing of building rating in Malioboro Central Business District.Keywords: urban retrofitting, e-commerce, information and communication technology, statistic regression, SCBD, Malioboro
Procedia PDF Downloads 2993486 Prediction of Bodyweight of Cattle by Artificial Neural Networks Using Digital Images
Authors: Yalçın Bozkurt
Abstract:
Prediction models were developed for accurate prediction of bodyweight (BW) by using Digital Images of beef cattle body dimensions by Artificial Neural Networks (ANN). For this purpose, the animal data were collected at a private slaughter house and the digital images and the weights of each live animal were taken just before they were slaughtered and the body dimensions such as digital wither height (DJWH), digital body length (DJBL), digital body depth (DJBD), digital hip width (DJHW), digital hip height (DJHH) and digital pin bone length (DJPL) were determined from the images, using the data with 1069 observations for each traits. Then, prediction models were developed by ANN. Digital body measurements were analysed by ANN for body prediction and R2 values of DJBL, DJWH, DJHW, DJBD, DJHH and DJPL were approximately 94.32, 91.31, 80.70, 83.61, 89.45 and 70.56 % respectively. It can be concluded that in management situations where BW cannot be measured it can be predicted accurately by measuring DJBL and DJWH alone or both DJBD and even DJHH and different models may be needed to predict BW in different feeding and environmental conditions and breedsKeywords: artificial neural networks, bodyweight, cattle, digital body measurements
Procedia PDF Downloads 3723485 Worst-Case Load Shedding in Electric Power Networks
Authors: Fu Lin
Abstract:
We consider the worst-case load-shedding problem in electric power networks where a number of transmission lines are to be taken out of service. The objective is to identify a prespecified number of line outages that lead to the maximum interruption of power generation and load at the transmission level, subject to the active power-flow model, the load and generation capacity of the buses, and the phase-angle limit across the transmission lines. For this nonlinear model with binary constraints, we show that all decision variables are separable except for the nonlinear power-flow equations. We develop an iterative decomposition algorithm, which converts the worst-case load shedding problem into a sequence of small subproblems. We show that the subproblems are either convex problems that can be solved efficiently or nonconvex problems that have closed-form solutions. Consequently, our approach is scalable for large networks. Furthermore, we prove the convergence of our algorithm to a critical point, and the objective value is guaranteed to decrease throughout the iterations. Numerical experiments with IEEE test cases demonstrate the effectiveness of the developed approach.Keywords: load shedding, power system, proximal alternating linearization method, vulnerability analysis
Procedia PDF Downloads 1403484 Artificial Intelligence and Distributed System Computing: Application and Practice in Real Life
Authors: Lai Junzhe, Wang Lihao, Burra Venkata Durga Kumar
Abstract:
In recent years, due to today's global technological advances, big data and artificial intelligence technologies have been widely used in various industries and fields, playing an important role in reducing costs and increasing efficiency. Among them, artificial intelligence has derived another branch in its own continuous progress and the continuous development of computer personnel, namely distributed artificial intelligence computing systems. Distributed AI is a method for solving complex learning, decision-making, and planning problems, characterized by the ability to take advantage of large-scale computation and the spatial distribution of resources, and accordingly, it can handle problems with large data sets. Nowadays, distributed AI is widely used in military, medical, and human daily life and brings great convenience and efficient operation to life. In this paper, we will discuss three areas of distributed AI computing systems in vision processing, blockchain, and smart home to introduce the performance of distributed systems and the role of AI in distributed systems.Keywords: distributed system, artificial intelligence, blockchain, IoT, visual information processing, smart home
Procedia PDF Downloads 1133483 Analysis of Decentralized on Demand Cross Layer in Cognitive Radio Ad Hoc Network
Authors: A. Sri Janani, K. Immanuel Arokia James
Abstract:
Cognitive radio ad hoc networks different unlicensed users may acquire different available channel sets. This non-uniform spectrum availability imposes special design challenges for broadcasting in CR ad hoc networks. Cognitive radio automatically detects available channels in wireless spectrum. This is a form of dynamic spectrum management. Cross-layer optimization is proposed, using this can allow far away secondary users can also involve into channel work. So it can increase the throughput and it will overcome the collision and time delay.Keywords: cognitive radio, cross layer optimization, CR mesh network, heterogeneous spectrum, mesh topology, random routing optimization technique
Procedia PDF Downloads 3593482 Electricity Price Forecasting: A Comparative Analysis with Shallow-ANN and DNN
Authors: Fazıl Gökgöz, Fahrettin Filiz
Abstract:
Electricity prices have sophisticated features such as high volatility, nonlinearity and high frequency that make forecasting quite difficult. Electricity price has a volatile and non-random character so that, it is possible to identify the patterns based on the historical data. Intelligent decision-making requires accurate price forecasting for market traders, retailers, and generation companies. So far, many shallow-ANN (artificial neural networks) models have been published in the literature and showed adequate forecasting results. During the last years, neural networks with many hidden layers, which are referred to as DNN (deep neural networks) have been using in the machine learning community. The goal of this study is to investigate electricity price forecasting performance of the shallow-ANN and DNN models for the Turkish day-ahead electricity market. The forecasting accuracy of the models has been evaluated with publicly available data from the Turkish day-ahead electricity market. Both shallow-ANN and DNN approach would give successful result in forecasting problems. Historical load, price and weather temperature data are used as the input variables for the models. The data set includes power consumption measurements gathered between January 2016 and December 2017 with one-hour resolution. In this regard, forecasting studies have been carried out comparatively with shallow-ANN and DNN models for Turkish electricity markets in the related time period. The main contribution of this study is the investigation of different shallow-ANN and DNN models in the field of electricity price forecast. All models are compared regarding their MAE (Mean Absolute Error) and MSE (Mean Square) results. DNN models give better forecasting performance compare to shallow-ANN. Best five MAE results for DNN models are 0.346, 0.372, 0.392, 0,402 and 0.409.Keywords: deep learning, artificial neural networks, energy price forecasting, turkey
Procedia PDF Downloads 2923481 Liquid Crystal Elastomers as Light-Driven Star-Shaped Microgripper
Authors: Indraj Singh, Xuan Lee, Yu-Chieh Cheng
Abstract:
Scientists are very keen on biomimetic research that mimics biological species to micro-robotic devices with the novel functionalities and accessibility. The source of inspiration is the complexity, sophistication, and intelligence of the biological systems. In this work, we design a light-driven star-shaped microgripper, an autonomous soft device which can change the shape under the external stimulus such as light. The design is based on light-responsive Liquid Crystal Elastomers which fabricated onto the polymer coated aligned substrate. The change in shape, controlled by the anisotropicity and the molecular orientation of the Liquid Crystal Elastomer, based on the external stimulus. This artificial star-shaped microgripper is capable of autonomous closure and capable to grab the objects in response to an external stimulus. This external stimulus-responsive materials design, based on soft active smart materials, provides a new approach to autonomous, self-regulating optical systems.Keywords: liquid crystal elastomers, microgripper, smart materials, robotics
Procedia PDF Downloads 1403480 Comparison of Power Consumption of WiFi Inbuilt Internet of Things Device with Bluetooth Low Energy
Authors: Darshana Thomas, Edward Wilkie, James Irvine
Abstract:
The Internet of things (IoT) is currently a highly researched topic, especially within the context of the smart home. These are small sensors that are capable of gathering data and transmitting it to a server. The majority of smart home products use protocols such as ZigBee or Bluetooth Low Energy (BLE). As these small sensors are increasing in number, the need to implement these with much more capable and ubiquitous transmission technology is necessary. The high power consumption is the reason that holds these small sensors back from using other protocols such as the most ubiquitous form of communication, WiFi. Comparing the power consumption of existing transmission technologies to one with WiFi inbuilt, would provide a better understanding for choosing between these technologies. We have developed a small IoT device with WiFi capability and proven that it is much more efficient than the first protocol, 433 MHz. We extend our work in this paper and compare WiFi power consumption with the other most widely used protocol BLE. The experimental results in this paper would conclude whether the developed prototype is capable in terms of power consumption to replace the existing protocol BLE with WiFi.Keywords: bluetooth, internet of things (IoT), power consumption, WiFi
Procedia PDF Downloads 2673479 Radio-Frequency Identification (RFID) Based Smart Helmet for Coal Miners
Authors: Waheeda Jabbar, Ali Gul, Rida Noor, Sania Kurd, Saba Gulzar
Abstract:
Hundreds of miners die from mining accidents each year due to poisonous gases found underground mining areas. This paper proposed an idea to protect the precious lives of mining workers. A supervising system is designed which is based on ZigBee wireless technique along with the smart protective helmets to detect real-time surveillance and it gives early warnings on presence of different poisonous gases in order to save mineworkers from any danger caused by these poisonous gases. A wireless sensor network is established using ZigBee wireless technique by integrating sensors on the helmet, apart from this helmet have embedded heartbeat sensor to detect the pulse rate and be aware of the physical or mental strength of a mineworker to increase the potential safety. Radio frequency identification (RFID) technology is used to find the location of workers. A ZigBee based base station is set-upped to control the communication. The idea is implemented and results are verified through experiment.Keywords: Arduino, gas sensor (MQ7), RFID, wireless ZigBee
Procedia PDF Downloads 4553478 MIMO Radar-Based System for Structural Health Monitoring and Geophysical Applications
Authors: Davide D’Aria, Paolo Falcone, Luigi Maggi, Aldo Cero, Giovanni Amoroso
Abstract:
The paper presents a methodology for real-time structural health monitoring and geophysical applications. The key elements of the system are a high performance MIMO RADAR sensor, an optical camera and a dedicated set of software algorithms encompassing interferometry, tomography and photogrammetry. The MIMO Radar sensor proposed in this work, provides an extremely high sensitivity to displacements making the system able to react to tiny deformations (up to tens of microns) with a time scale which spans from milliseconds to hours. The MIMO feature of the system makes the system capable of providing a set of two-dimensional images of the observed scene, each mapped on the azimuth-range directions with noticeably resolution in both the dimensions and with an outstanding repetition rate. The back-scattered energy, which is distributed in the 3D space, is projected on a 2D plane, where each pixel has as coordinates the Line-Of-Sight distance and the cross-range azimuthal angle. At the same time, the high performing processing unit allows to sense the observed scene with remarkable refresh periods (up to milliseconds), thus opening the way for combined static and dynamic structural health monitoring. Thanks to the smart TX/RX antenna array layout, the MIMO data can be processed through a tomographic approach to reconstruct the three-dimensional map of the observed scene. This 3D point cloud is then accurately mapped on a 2D digital optical image through photogrammetric techniques, allowing for easy and straightforward interpretations of the measurements. Once the three-dimensional image is reconstructed, a 'repeat-pass' interferometric approach is exploited to provide the user of the system with high frequency three-dimensional motion/vibration estimation of each point of the reconstructed image. At this stage, the methodology leverages consolidated atmospheric correction algorithms to provide reliable displacement and vibration measurements.Keywords: interferometry, MIMO RADAR, SAR, tomography
Procedia PDF Downloads 1953477 Perception-Oriented Model Driven Development for Designing Data Acquisition Process in Wireless Sensor Networks
Authors: K. Indra Gandhi
Abstract:
Wireless Sensor Networks (WSNs) have always been characterized for application-specific sensing, relaying and collection of information for further analysis. However, software development was not considered as a separate entity in this process of data collection which has posed severe limitations on the software development for WSN. Software development for WSN is a complex process since the components involved are data-driven, network-driven and application-driven in nature. This implies that there is a tremendous need for the separation of concern from the software development perspective. A layered approach for developing data acquisition design based on Model Driven Development (MDD) has been proposed as the sensed data collection process itself varies depending upon the application taken into consideration. This work focuses on the layered view of the data acquisition process so as to ease the software point of development. A metamodel has been proposed that enables reusability and realization of the software development as an adaptable component for WSN systems. Further, observing users perception indicates that proposed model helps in improving the programmer's productivity by realizing the collaborative system involved.Keywords: data acquisition, model-driven development, separation of concern, wireless sensor networks
Procedia PDF Downloads 4343476 Multimodal Direct Neural Network Positron Emission Tomography Reconstruction
Authors: William Whiteley, Jens Gregor
Abstract:
In recent developments of direct neural network based positron emission tomography (PET) reconstruction, two prominent architectures have emerged for converting measurement data into images: 1) networks that contain fully-connected layers; and 2) networks that primarily use a convolutional encoder-decoder architecture. In this paper, we present a multi-modal direct PET reconstruction method called MDPET, which is a hybrid approach that combines the advantages of both types of networks. MDPET processes raw data in the form of sinograms and histo-images in concert with attenuation maps to produce high quality multi-slice PET images (e.g., 8x440x440). MDPET is trained on a large whole-body patient data set and evaluated both quantitatively and qualitatively against target images reconstructed with the standard PET reconstruction benchmark of iterative ordered subsets expectation maximization. The results show that MDPET outperforms the best previously published direct neural network methods in measures of bias, signal-to-noise ratio, mean absolute error, and structural similarity.Keywords: deep learning, image reconstruction, machine learning, neural network, positron emission tomography
Procedia PDF Downloads 1113475 Quantitative Analysis of Presence, Consciousness, Subconsciousness, and Unconsciousness
Authors: Hooshmand Kalayeh
Abstract:
The human brain consists of reptilian, mammalian, and thinking brain. And mind consists of conscious, subconscious, and unconscious parallel neural-net programs. The primary objective of this paper is to propose a methodology for quantitative analysis of neural-nets associated with these mental activities in the neocortex. The secondary objective of this paper is to suggest a methodology for quantitative analysis of presence; the proposed methodologies can be used as a first-step to measure, monitor, and understand consciousness and presence. This methodology is based on Neural-Networks (NN), number of neuron in each NN associated with consciousness, subconsciouness, and unconsciousness, and number of neurons in neocortex. It is assumed that the number of neurons in each NN is correlated with the associated area and volume. Therefore, online and offline visualization techniques can be used to identify these neural-networks, and online and offline measurement methods can be used to measure areas and volumes associated with these NNs. So, instead of the number of neurons in each NN, the associated area or volume also can be used in the proposed methodology. This quantitative analysis and associated online and offline measurements and visualizations of different Neural-Networks enable us to rewire the connections in our brain for a more balanced living.Keywords: brain, mind, consciousness, presence, sub-consciousness, unconsciousness, skills, concentrations, attention
Procedia PDF Downloads 3143474 Environmental Assessment of Roll-to-Roll Printed Smart Label
Authors: M. Torres, A. Moulay, M. Zhuldybina, M. Rozel, N. D. Trinh, C. Bois
Abstract:
Printed electronics are a fast-growing market as their applications cover a large range of industrial needs, their production cost is low, and the additive printing techniques consume less materials than subtractive manufacturing methods used in traditional electronics. With the growing demand for printed electronics, there are concerns about their harmful and irreversible contribution to the environment. Indeed, it is estimated that 80% of the environmental load of a product is determined by the choices made at the conception stage. Therefore, examination through a life cycle approach at the developing stage of a novel product is the best way to identify potential environmental issues and make proactive decisions. Life cycle analysis (LCA) is a comprehensive scientific method to assess the environmental impacts of a product in its different stages of life: extraction of raw materials, manufacture and distribution, use, and end-of-life. Impacts and major hotspots are identified and evaluated through a broad range of environmental impact categories of the ReCiPe (H) middle point method. At the conception stage, the LCA is a tool that provides an environmental point of view on the choice of materials and processes and weights-in on the balance between performance materials and eco-friendly materials. Using the life cycle approach, the current work aims to provide a cradle-to-grave life cycle assessment of a roll-to-roll hybrid printed smart label designed for the food cold chain. Furthermore, this presentation will present the environmental impact of metallic conductive inks, a comparison with promising conductive polymers, evaluation of energy vs. performance of industrial printing processes, a full assessment of the impact from the smart label applied on a cellulosic-based substrate during the recycling process and the possible recovery of precious metals and rare earth elements.Keywords: Eco-design, label, life cycle assessment, printed electronics
Procedia PDF Downloads 1633473 New Approach for Minimizing Wavelength Fragmentation in Wavelength-Routed WDM Networks
Authors: Sami Baraketi, Jean Marie Garcia, Olivier Brun
Abstract:
Wavelength Division Multiplexing (WDM) is the dominant transport technology used in numerous high capacity backbone networks, based on optical infrastructures. Given the importance of costs (CapEx and OpEx) associated to these networks, resource management is becoming increasingly important, especially how the optical circuits, called “lightpaths”, are routed throughout the network. This requires the use of efficient algorithms which provide routing strategies with the lowest cost. We focus on the lightpath routing and wavelength assignment problem, known as the RWA problem, while optimizing wavelength fragmentation over the network. Wavelength fragmentation poses a serious challenge for network operators since it leads to the misuse of the wavelength spectrum, and then to the refusal of new lightpath requests. In this paper, we first establish a new Integer Linear Program (ILP) for the problem based on a node-link formulation. This formulation is based on a multilayer approach where the original network is decomposed into several network layers, each corresponding to a wavelength. Furthermore, we propose an efficient heuristic for the problem based on a greedy algorithm followed by a post-treatment procedure. The obtained results show that the optimal solution is often reached. We also compare our results with those of other RWA heuristic methods.Keywords: WDM, lightpath, RWA, wavelength fragmentation, optimization, linear programming, heuristic
Procedia PDF Downloads 5273472 Systematic Literature Review and Bibliometric Analysis of Interorganizational Employee Mobility Determinants
Authors: Iva Zdrilić, Petra Došenović Bonča, Darija Aleksić
Abstract:
Since the boundaryless career, with its emphasis on cross-employer movements, was introduced as a new paradigm of career development, inter-organizational employee mobility has been increasing. Although this phenomenon may have positive implications for individual careers and destination organizations, the consequences for the source organizations losing workers are less clear. The aim of this paper is thus to develop a comprehensive typology of possible inter-organizational employee mobility determinants. Since the most common classification differentiates between mobility determinants at different levels (i.e., economic, organizational, and individual), this paper focuses on building a comprehensive multi-level typology of inter-organizational mobility determinants across diverse sectors and industries. By using a structured literature review approach and bibliometric analysis, the paper reveals both intricate relationships between different mobility determinants and the complexity of inter-organizational networks and social ties. The latter appears as both a mobility determinant (at the organizational and individual level) and a mobility effect. Indeed, inter-organizational employee mobility leads to the formation of networks between source and destination organizations. These networks are practically based on the social ties between mobile employees and their colleagues and, in this way, they close the "inter-organizational employee mobility - inter-organizational network/ties" circle. The paper contributes to the career development literature by uncovering hitherto underexplored diverse determinants of intra- and inter-sectoral mobility as well as the conflicting results of the existing studies on some factors (e.g., inter-organizational networks and/or social ties) that appear both as a mobility determinant and a mobility effect.Keywords: inter-organizational mobility, social ties, inter-organizational network, knowledge transfer
Procedia PDF Downloads 1163471 Social Media Mining with R. Twitter Analyses
Authors: Diana Codat
Abstract:
Tweets' analysis is part of text mining. Each document is a written text. It's possible to apply the usual text search techniques, in particular by switching to the bag-of-words representation. But the tweets induce peculiarities. Some may enrich the analysis. Thus, their length is calibrated (at least as far as public messages are concerned), special characters make it possible to identify authors (@) and themes (#), the tweet and retweet mechanisms make it possible to follow the diffusion of the information. Conversely, other characteristics may disrupt the analyzes. Because space is limited, authors often use abbreviations, emoticons to express feelings, and they do not pay much attention to spelling. All this creates noise that can complicate the task. The tweets carry a lot of potentially interesting information. Their exploitation is one of the main axes of the analysis of the social networks. We show how to access Twitter-related messages. We will initiate a study of the properties of the tweets, and we will follow up on the exploitation of the content of the messages. We will work under R with the package 'twitteR'. The study of tweets is a strong focus of analysis of social networks because Twitter has become an important vector of communication. This example shows that it is easy to initiate an analysis from data extracted directly online. The data preparation phase is of great importance.Keywords: data mining, language R, social networks, Twitter
Procedia PDF Downloads 1843470 The Network Effect on Green Information on Taiwan Social Network Sites
Authors: Pi Hsia Liang
Abstract:
The rise of Facebook, Twitter, and other social networks significantly changes in interconnections between people, enhancing the process of information dissemination and amplify the influence of that information. Therefore, to develop informational efficiency or signaling equilibrium type of information environment among social networks, without adverse selection effects, becomes an important issue. Thus, someone may post a piece of intentional information in relation to personal interest for trying to create marginal influence. Therefore, economists are seeking to establish theories of informational efficiency under social network environment in order to resolve adverse selection issues. Reputation could be one of the important factors in the process of creating informational efficiency. Additionally, investors how to process green information, or information of corporate social responsibility is a very important study. This study essentially employs experimental study for examining how investors use stock relevant green information in Facebook and various Taiwan local networks. Facebook, and blogs of Money DJ, Technews and cnYES, respectively, are the primary sites for this examination that also allow to differentiate effects between Facebook and other local social networks. Questionnaire is developed for such an experimental testing. Note that questionnaire allows this study to group, for example, decision frequency and length of time duration focusing on social networks that are used for discriminating investor type and competence of informed investor. This study selects 500 investors that can be separated into two respective 250 samples as the control group and 250 samples in such an experimental. The quantity of sample investor sufficiently results in statistic significance of this experimental study. The empirical results of this study can be used for explaining how financial information in relation to corporate social responsibility would be disseminated in social websites. Therefore, we can lead to better interpretation of price/earnings relationship type of study and empirical studies of green information usefulness or informational efficiency Note that the above mentioned empirical studies did not exist any social network and annual report of corporate social responsibility. This study expects to find the results that both network degree and network cluster significantly affected green information dissemination frequency. In other words, investors with more connections and with high clustered connections might exert a greater influence on their green information dissemination process. The preferred users of financial social networks could make better stock decision that could amplify effects of green information. In addition, Facebook would be more influential than other local Taiwan financial social networks, although Facebook is not a specialized financial social network. In other words, the popularity and reputation effects of Facebook significantly contribute to usefulness of green information and influence of green information. Third, it has a better chance to find rumor or cheating information in local Taiwan financial social networks than Facebook. In other words, Facebook possesses reputation effect, or a better informational efficiency. Or, even though Taiwan local financial social networks have marginal informational effects on stock price, because of shortage of informational efficiency or monitoring system, information could be a tool for those whom owning superior information.Keywords: network effect on financial services, informational efficiency theory, social networks, social websites
Procedia PDF Downloads 2463469 Performance Assessment of Carrier Aggregation-Based Indoor Mobile Networks
Authors: Viktor R. Stoynov, Zlatka V. Valkova-Jarvis
Abstract:
The intelligent management and optimisation of radio resource technologies will lead to a considerable improvement in the overall performance in Next Generation Networks (NGNs). Carrier Aggregation (CA) technology, also known as Spectrum Aggregation, enables more efficient use of the available spectrum by combining multiple Component Carriers (CCs) in a virtual wideband channel. LTE-A (Long Term Evolution–Advanced) CA technology can combine multiple adjacent or separate CCs in the same band or in different bands. In this way, increased data rates and dynamic load balancing can be achieved, resulting in a more reliable and efficient operation of mobile networks and the enabling of high bandwidth mobile services. In this paper, several distinct CA deployment strategies for the utilisation of spectrum bands are compared in indoor-outdoor scenarios, simulated via the recently-developed Realistic Indoor Environment Generator (RIEG). We analyse the performance of the User Equipment (UE) by integrating the average throughput, the level of fairness of radio resource allocation, and other parameters, into one summative assessment termed a Comparative Factor (CF). In addition, comparison of non-CA and CA indoor mobile networks is carried out under different load conditions: varying numbers and positions of UEs. The experimental results demonstrate that the CA technology can improve network performance, especially in the case of indoor scenarios. Additionally, we show that an increase of carrier frequency does not necessarily lead to improved CF values, due to high wall-penetration losses. The performance of users under bad-channel conditions, often located in the periphery of the cells, can be improved by intelligent CA location. Furthermore, a combination of such a deployment and effective radio resource allocation management with respect to user-fairness plays a crucial role in improving the performance of LTE-A networks.Keywords: comparative factor, carrier aggregation, indoor mobile network, resource allocation
Procedia PDF Downloads 1783468 Intelligent Transport System: Classification of Traffic Signs Using Deep Neural Networks in Real Time
Authors: Anukriti Kumar, Tanmay Singh, Dinesh Kumar Vishwakarma
Abstract:
Traffic control has been one of the most common and irritating problems since the time automobiles have hit the roads. Problems like traffic congestion have led to a significant time burden around the world and one significant solution to these problems can be the proper implementation of the Intelligent Transport System (ITS). It involves the integration of various tools like smart sensors, artificial intelligence, position technologies and mobile data services to manage traffic flow, reduce congestion and enhance driver's ability to avoid accidents during adverse weather. Road and traffic signs’ recognition is an emerging field of research in ITS. Classification problem of traffic signs needs to be solved as it is a major step in our journey towards building semi-autonomous/autonomous driving systems. The purpose of this work focuses on implementing an approach to solve the problem of traffic sign classification by developing a Convolutional Neural Network (CNN) classifier using the GTSRB (German Traffic Sign Recognition Benchmark) dataset. Rather than using hand-crafted features, our model addresses the concern of exploding huge parameters and data method augmentations. Our model achieved an accuracy of around 97.6% which is comparable to various state-of-the-art architectures.Keywords: multiclass classification, convolution neural network, OpenCV
Procedia PDF Downloads 1763467 BOFSC: A Blockchain Based Decentralized Framework to Ensure the Transparency of Organic Food Supply Chain
Authors: Mifta Ul Jannat, Raju Ahmed, Al Mamun, Jannatul Ferdaus, Ritu Costa, Milon Biswas
Abstract:
Blockchain is an internet-based invention that is coveted in the permanent, scumbled record for its capacity to openly accept, record, and distribute transactions. In a traditional supply chain, there are no trustworthy participants for an organic product. Yet blockchain engineering may provide confidence, transparency, and traceability. Blockchain varies in how companies get real, checked, and lasting information from their supply chain and lock in customers. In an arrangement of cryptographic squares, Blockchain digitizes each connection by sparing it. No one person may alter the documents, and any alteration within the agreement is clear to all. The coming to the record is tamper proof and unchanging, offering a complete history of the object’s life cycle and minimizing opening for extorting. The primary aim of this analysis is to identify the underlying problem that the customer faces. In this post, we will minimize the allocation of fraud data through the ’Smart Contract’ and include a certificate of quality assurance.Keywords: blockchain technology, food supply chain, Ethereum, smart contract, quality assurance, trustability, security, transparency
Procedia PDF Downloads 1533466 An Efficient Tool for Mitigating Voltage Unbalance with Reactive Power Control of Distributed Grid-Connected Photovoltaic Systems
Authors: Malinwo Estone Ayikpa
Abstract:
With the rapid increase of grid-connected PV systems over the last decades, genuine challenges have arisen for engineers and professionals of energy field in the planning and operation of existing distribution networks with the integration of new generation sources. However, the conventional distribution network, in its design was not expected to receive other generation outside the main power supply. The tools generally used to analyze the networks become inefficient and cannot take into account all the constraints related to the operation of grid-connected PV systems. Some of these constraints are voltage control difficulty, reverse power flow, and especially voltage unbalance which could be due to the poor distribution of single-phase PV systems in the network. In order to analyze the impact of the connection of small and large number of PV systems to the distribution networks, this paper presents an efficient optimization tool that minimizes voltage unbalance in three-phase distribution networks with active and reactive power injections from the allocation of single-phase and three-phase PV plants. Reactive power can be generated or absorbed using the available capacity and the adjustable power factor of the inverter. Good reduction of voltage unbalance can be achieved by reactive power control of the PV systems. The presented tool is based on the three-phase current injection method and the PV systems are modeled via an equivalent circuit. The primal-dual interior point method is used to obtain the optimal operating points for the systems.Keywords: Photovoltaic system, Primal-dual interior point method, Three-phase optimal power flow, Voltage unbalance
Procedia PDF Downloads 332