A b s t r a c t s
of Keynote Speakers

Patrick J.C. Aerts
Netherlands eScience Center (NLeSC), Amsterdam, The Netherlands

Strengthening the European eScience scene: PLAN-E,
A Platform of National eScience Centers in Europe


Abstract:

The gradual shift in focus from e-infrastructures towards their efficient deplorability leads to new ways of doing science: problem driven, multi-disciplinary, holistic in its deployment of e-infrastructures and software/tools, such as the latest fruits from Computer Science, and – in this particular time frame – often data intensive. This shift in focus and this new way of doing science is embodied in the term escience and in the organizations in support of this development.

The Netherlands eScience Center is solely devoted to this development, but there are many other institutes and centers in The Netherlands, Europe and elsewhere that face the same transition and are in a phase of delivering support, developing services, new tools, new educational directions to enable scientists to accelerate discovery. Therefore new social network platforms are being set up (ePLAN and PLAN-E) to share experiences, best practices, career challenges, curriculum innovation and more.

One of the shared topics is "data stewardship and software sustainability" for which a conceptual approach is being designed, involving researchers at disciplinary level.


Peter Brezany
Brno University of Technology, Czech Republic,
and University of Vienna, Austria

Optimized Management of BIG Data produced in Collaborative Life-Science and Health Care Research Environments

Abstract:

The life-science and health care research environments offer an abundance of new opportunities for improvement their efficiency and productivity using big data and analytics in collaborative research processes. This talks introduces research results we achieved in two recent research projects, ABA and SPES, respectively.
1) Dataspace-Based Support Platform for Breath Gas Analysis (ABA). We have addressed the development and utilization of an information infrastructure called the ABA-Cloud [1] based on the novel scientific dataspace paradigm, which has aimed at providing advanced data management, automated data collection, data analysis and experimentation within the domain of breath gas analysis. The breath gas analysis scientific community is continuously developing new analytical methods and collecting pilot data, for instance to identify marker compounds for various diseases, like lung cancer, negative side effects of chemotherapies, etc.
2) Support Patients through E-Service Solutions (SPES). Based on the results of the ABA project, we focused on the development of the indoor and outdoor support for dementia patients, non-invasive methods of their brain stimulation, capturing and analysis of the produced data and querying the analysis results [2]. The talk will also discuss the extension of the developed technology to other research domains.

[1] I. Elsayed, et al. ABA-Cloud: Support for Collaborative Breath Research. Journal of Breath Research, 7 (2013) 026007 (13 pp). Online at stacks.iop.org/JBR/7/026007.
[2] P. Brezany, et al. Optimized Management of BIG Data Produced in Brain Disorder Rehabilitation. To appear in the book "BIG Data Optimization", Springer, March 2016.


Peter Coveney
Department of Chemistry, University College London, UK

Using Prometheus and PL-Grid to Predict the Properties of Nanomaterials and Drug Ranking in Personalised Medicine

Abstract:

We shall discuss two ambitious projects in computational science which are currently underway, making use of Prometheus and PL-Grid resources. One of these projects is concerned with a grand challenge in materials science — namely, the theoretical prediction of large scale materials properties based directly on the chemical composition of their ingredients. We describe our virtual laboratory for multiscale modelling and simulation in distributed high performance computing environments which makes this possible, through the use of sophisticated software services which handle a very demanding workflow.

The second project is concerned with personalised medicine. The potential impact centres around our ability to marshall substantial quantities of patient data and to use them to perform predictive modelling and simulation in order to deliver therapies and to enhance clinical decision making, on time scales which are far shorter than those usually considered in the context of academic research. Secure access to personal data, as well as to powerful computational resources, are essential.

We shall discuss the underlying e-infrastructure requirements, including data, compute and networks, and describe how we propose to use Prometheus & PL-Grid to support this work.


Tiziana Ferrari
EGI.eu, Amsterdam, The Netherlands

EGI: Status, Challenges and Opportunities in the Digital Single Market Era

Abstract:

Nowadays, research is increasingly and in many cases exclusively data driven. Knowledge of how to use tools to manipulate research data, and the availability of e-infrastructures to support them, are foundational.

New communities of practice are forming around interests in digital tools, computing facilities and data repositories. By making infrastructure services, community engagement and training inseparable, existing communities can be empowered by new ways of doing research, and can grow around interdisciplinary data that can be easily accessed and manipulated via innovative tools as never done before.

Key enablers of this process are openness, participation, collaboration, sharing and reuse, and the definition of standards and the enforcement of their adoption within the community. In some disciplinary areas communities succeeded in defining and adopting a community governance of the production and sustainable sharing of data and applications.

The European Commission has identified the completion of the Digital Single Market (DSM) as one of its 10 political priorities and much effort is being done in aligning strategies and initiatives along the DSM targets.

This presentation provides an overview of the services and the initiatives that EGI is supporting to promote open science and the technical roadmap and challenges that will drive its future evolution.


Fabrizio Gagliardi
Barcelona SuperComputing Centre, Spain
EU-ACM Europe Chair, Switzerland
GS Scientific Institute, Italy

RDA, Research Data Alliance, a Worldwide Effort to Promote Open Sharing and Preservation of Scientific Data

Abstract:

The mission statement of RDA is: Researchers and innovators openly sharing data across technologies, disciplines, and countries to address the grand challenges of society. The Research Data Alliance aims to building the social and technical bridges that enable open sharing of data and their preservation.  The talk will review the progress so far and discuss current challenges and future plans.


Piotr Gawron
Institute of Theoretical and Applied Informatics, PAN, Gliwice, Poland

Tensor Networks a Language for Classical and Quantum Informatics

Abstract:

Tensor network represents tensors (data cubes) and multi-linear relations between them. In quantum information tensors form a natural way of representing quantum states and operations. Multi-linear relations between tensors provide an extension of matrix analysis, that simplifies greatly reasoning about quantum objects. Tensor networks have found use in many applications in classical informatics. Those networks can efficiently describe large data and relations between them. The talk will cover elements of tensor networks with applications to quantum informatics and data compression.


Matti Heikkurinen
Ludwig-Maximilians-Universität München, Germany

Trends in Environmental Computing

Abstract:

Environmental computing is a relatively novel term describing an integrated approach to producing a comprehensive view of different environmental processes and their interactions and impact. On technical level this means linking tools and approaches from several disciplines together to create a multi-model, multi-data ensembles that can be used in solving different planning, policy or operational challenges. However, in order to effectively support the strategic or operational action, it is important to pay close attention to standards and processes surrounding the technical solution. For example, consistent metadata describing the capabilities and limitations of the models is a crucial feature of environmental computing systems. Similarly, consistent data curation practices and sustainability of the archive services across organizations is similarly essential. The multi-model approach also inherently removes the original researchers from direct contact with the execution of the simulations and analysis tasks, and makes metadata and other supporting material crucial in linking the research and policy/operational activities.
Technical interoperability and consistency of the procedures are further complicated by the environmental computing being a distributed system: data sources (especially sensors) require distributed approach, and amounts of data in different repositories make transferring all the data to single location impossible.


Rod Hose
University of Sheffield, UK

The Virtual Physiological Human: Where Now and Where Next?

Abstract:

The Virtual Physiological Human is an initiative that was supported under the Framework 7 programme, with the aim of developing sophisticated and integrated computational models of human systems and of reaching impact in the industrial and clinical environments. In this talk I will give examples of the types of models and work flows that have been developed and outline their current translational status. I will review the requirement for hpc, grid and cloud computing in the context of exploitation of VPH technology. In the industrial context of device design there is a clear requirement for operation of the most sophisticated models, with associated hpc demands. For the most efficient and effective analysis in a clinical environment there is a great potential for the use of Reduced Order Models, which will often be underpinned by large-scale off-line computing. I will outline some of the remaining challenges, which include the representation of variation and uncertainty in VPH models, and the integration of a wide range of heterogeneous data from clinical and other sources to inform the physiological envelope over which the models should operate.


Michael Lees
Computational Science Lab, University of Amsterdam, The Netherlands

Building, Calibrating and Validating Models of Human Crowds: A Complex Systems Approach.

Abstract:

In this talk I will describe an approach to the modelling and simulation of human crowds. By developing accurate models of human crowds the hope is to be able to predict and perhaps manage mass gatherings of people, or enable safer architectural design. There have been numerous attempts at modelling crowds over the past two decades, here I will highlight one approach which uses agent-based modelling. While agent-based modelling offers a very natural approach for modelling human crowds it does introduce significant challenges. Firstly, how to build a model capable of capturing human-like decisions. Secondly, how to calibrate models against real-world data and finally how to validate the model output. This talk will cover work in these areas and finally discuss ways in which these findings might be used to help make large human gatherings safer.


Philippe Trautmann
HPC & POD Sales Director, Europe, Hewlett-Packard

Is HPC/HPCDA Moving from a Capex to an Opex Model? An HP Vision

Abstract:

During the last 25 years HPC has been considered as a fixed asset. Companies and organizations running HPC workloads gradually realize that moving their HPC infrastructure and operations from a fixed asset to an operating expense model has several advantages.  Despite the technical constraints imposed by the nature of the HPC workloads, there is a business model that today provides companies and organizations a way to move a part or all of their HPC infrastructure to the cloud.  Large European HPC users are currently moving their HPC workloads to the cloud, whatever the cloud implementation could be. In cooperation with Intel HP has a capability to provide a variety of solutions in this space, with adapted technical and financial models. We will provide some examples of companies which have made this move, and will also provide some hints and guidelines to support this model.


Gregory V. Wilson
Software Carpentry Foundation

Software Carpentry: Lessons Learned

Abstract:

Since its re-launch in 2010, Software Carpentry has become a world-wide organization: between January 2013 and July 2015, its volunteer instructors delivered over 400 workshops to over 12,000 researchers in two dozen countries. Along the way, it has developed an instructor training program that introduces scientists to evidence-based teaching practices, and shown that development practices originating in open source software can successfully be applied to the development of educational materials. This talk will explore the lessons we have learned along the way and discuss some of the experiments we would like to try next.


Hai Zhuge
Aston University, UK

Mapping Big Data into Knowledge Space

Abstract:

Big data have attracted great attention in science, industry and society. However, its nature and the fundamental challenge have not been recognised. Big data research is developing with the evolving scientific paradigm, the fourth industrial revolution, and transformational innovation of technologies. This lecture tries to answer the following questions: What is the nature of big data? What kind of infrastructure is requested to support not only big data management and analysis but also knowledge discovery and management? What is the fundamental challenge of big data computing? Can we find an approach to mapping big data into knowledge space?

Organizers:

AGH University of Science and Technology
AGH University of Science
and Technology
ACC Cyfronet AGH
ACC Cyfronet AGH
Department of Computer Science AGH

Department of Computer
Science AGH