The Industry 4.0 National Technology Platform was established under the leadership of the Institute for Computer Science and Control (SZTAKI), Hungarian Academy of Sciences, with the participation of research institutions, companies, universities and professional organizations having premises in Hungary, and with the full support and commitment of the Government of Hungary, and specifically that of the Ministry of National Economy.
The HydroCobotics H2020 cascade consortium is coordinated by industrial robotics company Hepenix kft., while hydroponics company Green Drops Farm kft. is also a member. Within ELKH SZTAKI, the project is coordinated by Research Laboratory on Engineering & Management Intelligence research fellow Imre Paniti. The research and development project starts on the 1st of January, 2021 and presumably concludes on the 30th of September, 2021.
Knowledge graphs have become the most important tools of sharing and connecting research or industry information and contacts. These graphs enable, in a flexible way, the access, use and publication of data as a distributed system.
The DIGITbrain project aims to enable customised industrial products and to facilitate cost-effective distributed and localised production for manufacturing SMEs, by means of leveraging edge-, cloud- and HPC-based modelling, simulation, optimisation, analytics, and machine learning tools and by means of augmenting the concept of digital twin with a memorising capacity towards a) recording the provenance and boosting the cognition of the industrial product over its full lifecycle, and b) empowering the network of DIHs to implement the smart business model "Manufacturing as a Service".
Higher education has to keep pace with the global market needs for the necessary ICT(Information and Communications Technology) skills and the overall understanding of the complexity of industries in the 21st century. Global market companies have to effectively deal with the constant evolution of products, processes and production systems (and all in parallel) that can be more easily monitored, developed and up-graded using digital applications based on the concept of digital twin and taking advantage of Virtual Reality (VR) and Augmented Reality (AR) simulations.
The EOSC-hub project creates the integration and management system of the future European Open Science Cloud that delivers a catalogue of services, software and data from the EGI Federation, EUDAT CDI, INDIGO-DataCloud and major research e-infrastructures. This integration and management system (the Hub) builds on mature processes, policies and tools from the leading European federated e-Infrastructures to cover the whole life-cycle of services, from planning to delivery.
The goal of the project is to reduce the death rate of neonatal and prenatal infants, and increase their chances of living via developing vision based non-contact body devices for monitoring physiological signals like pulse rate, breath rate, blood oxygenation, activity, and body temperature using remote photoplethysmographic methods.
The project aims to process the data of novel 3D sensors (e.g. Microsoft Kinect, Lidar, MRI, CT) available in a wide range of application fields and to fuse them with 2D image modalities to build saliency models, which are able to automatically and efficiently emphasize visually dominant regions. Such models not only tighten the region of interest for further image processing steps, but facilitate and increase the efficiency of segmentation in different application fields with available 3D sensor data, e.g.
Numerous automotive and small aircraft companies have announced promising new applications in the field of autonomous vehicles. Alongside self-driving cars, in the near future small-size micro aerial vehicles could be used for goods delivery (Amazon Prime Air, DHL, Alibaba, Matternet, Swiss Post), in healthcare (Matternet, Flirtey, Wingtra, RedLine), to carry out various inspection and surveillance tasks (SenseFly, Skycatch), or can be deployed at accidents as remote-controlled first aid/responder devices (Drone Aventures, Microdrones).
The aim of the project is to develop an image fusion and processing method that uses images of cameras with different modalities to track various objects, taking into account the needs of border surveillance end-users.