LCB 98-xx

Models of Networked Analysis at Regional Centres for LHC Experiments

(MONARC)

PROJECT EXECUTION PLAN

Version of 10th. September 1998



Prepared by

M. Aderholz (MPI), Katsuya Amako (KEK), Eva Arderiu Ribera (CERN), L. Barone (Roma), G. Battistona (Milano), Julian Bunn (Caltech/CERN), Joel Butler (FNAL), M. Campanella (Milano), Paolo Capiluppi (Bologna/INFN), M. Dameri (Genova), D. Diacono (Bari), A. di Mattia (Roma), U. Gasparini (Padova), Fabrizio Gagliardi (CERN), Irwin Gaines (FNAL), Philippe Galvez (Caltech), C. Grandi (Bologna), Frank Harris (Oxford/CERN), Koen Holtman (CERN), Veikko Karimaki (Helsinki), J. Klem (Helsinki), M. Leltchouk (Columbia), P. Lubrano (Perugia), L. Luminari (Roma), Michele Michelotto (Padua), Ian McArthur (Oxford), Harvey Newman (Caltech), Steve O'Neale (Birmingham), Bianca Osculati (Genova), Laura Perini (Milano), Jim Pinfold (Alberta), Les Robertson (CERN), Ruth Pordes (FNAL), Simona Rolli (Tufts), Takashi Sasaki (KEK), L. Servoli (Perugia), R.D. Schaffer (Orsay), Massimo Sgaravatto (Padova), Terry Schalk (BaBar), Jamie Shiers (CERN), Lucia Silvestris (Bari), K. Sliwa (Tufts), C. Stanescu (Roma), Tim Smith (CERN/IT), Krzysztof Sliwa (Tufts), Stig Topp-Jorgentsen (Oxford), Christoph von Praun (CERN), Enzo Valente (INFN), Ian Willers (CERN), Rick Wilkinson (Caltech), David O. Williams (CERN)


20 September 1998

Executive Summary

The MONARC Project will attempt to determine which classes of Computing Models are feasible for the LHC Experiments. The boundary conditions for the models will be the network capacity and data handling resources likely to be available at the start of, and during LHC running.

The main deliverable from the Project will be a set of example "baseline" models. The Project will also help to define Regional Centre architectures and functionality, the physics analysis process for LHC Experiments, and guidelines for retaining feasibility over the course of running. The results will be made available in time for use in the Experiments’ Computing Technical Design Reports which are due in 2002.

The approach taken in the Project is to develop and execute simulations of the various candidate distributed computing systems. The granularity of the simulations will be adjusted according to the detail required from the results. The models will be iteratively tuned in the light of experience. Simulation of the diverse tasks that are part of the spectrum of computing in HEP will be undertaken. For this, a simulation and modelling tool kit will be developed which will enable a full understanding of the impact of the varying degrees of network saturation on the models.

Chapter 1: Introduction

The LHC experiments have envisaged Computing Models (CM) involving hundreds of physicists doing analysis at institutions around the world. CMS and ATLAS also are considering the use of Regional Centres, each of which could complement the functionality of the CERN Centre. They are intended to facilitate the access to the data, with more efficient and cost-effective data delivery to the groups in each world region, using national networks of greater capacity than may be available on intercontinental links.

The LHC Models encompass a complex set of wide-area, regional and local-area networks, a heterogeneous set of compute- and data-servers, and a yet-to-be determined set of priorities for group-oriented and individuals' demands for remote data. Distributed systems of this scope and complexity do not exist yet, although systems of a similar size to those foreseen for the LHC experiments are predicted to come into operation by around 2005 at large corporations.

In order to proceed with the planning and design of the LHC Computing Models, and to correctly dimension the capacity of the networks and the size and characteristics of Regional Centres, it is essential to conduct a systematic study of these distributed systems. The MONARC project therefore intends to simulate and study network-distributed computing architectures, data access and data management systems that are major components of the CM, and the ways in which the components interact across networks. MONARC will bring together the efforts and relevant expertise from the LHC experiments and their R&D projects, as well as from current or imminent experiments already engaged in building distributed systems for computing, data access, simulation and analysis.


The primary goals of this project are:

As a result of this study, MONARC will deliver a set of tools for simulating candidate Computing Models of the experiments, and a set of common guidelines to allow the experiments to formulate their final Models.

Distributed databases are a crucial aspect of the studies. The RD45 project has developed considerable expertise in the field of Object Oriented Database Management Systems (ODBMS). MONARC will benefit from this experience and cooperate with RD45 in the specific areas where the work of the two projects overlaps. MONARC will investigate questions which are largely complementary to RD45, such as network performance and prioritization of traffic for a variety of applications that must coexist and share the network resources.

 

Chapter 2: Objectives

A set of common modelling and simulation tools will be developed in MONARC. These tools will be integrated in an environment which will enable the LHC experiments to realistically evaluate and optimize their physics analysis procedures, and their Computing Models, basing them on distributed data and computing architectures. Tools to realistically estimate the network bandwidth required in a given Computing Model will be developed. The parameters that are necessary and sufficient to characterize the Computing Model and its performance will be identified. The methods and tools to measure the Model's performance, and to detect bottlenecks, will be designed and developed, and also tested in prototypes. This work will be done with as much as possible co-operation with the present LHC R&D Projects, and current or imminent experiments.

The principal objective is to determine a set of feasible models, and to provide a set of guidelines which the experiments could use to build their respective Computing Models.

The subsidiary objectives are:

 

Chapter 3: Workplan

3.1 Scope

MONARC aims to study analysis models and architectures suitable for LHC experiments, in order to contribute to their Computing Models in time for the updated version of CTP (CTPR) scheduled around end 1999.

3.2 Interaction with other projects

The project involves collaboration not only from the LHC experiments, but also from other HEP experiments preparing to run in shortly (such as BaBar and CDF); these are going to develop expertise in many of the fields of interest for LHC. Channels for regular contact will be set up as soon as the project is approved, taking advantage of collaborators who are members also of the collaborations of interest.

The project has to interact also with other teams: with RD45, with CERN Technology Tracking Team, and with the CERN HPSS team, as explained in the section of this document devoted to "Interactions with other Projects".

The project requires a sizeable manpower contribution from CERN:

3.3 Working methods

The working methods to be employed in MONARC are largely dictated by aspects of the Project’s structure:

MONARC working methods will embody the following principles:

The principles and requirements stated above lead naturally to a collaboration structure based on Working Groups and on a Steering Group (composed by the Chairpersons of the Working Groups together with the Spokesman and the Project Leader). The Steering Group assures the integration of the different tasks that are pursued separately in the various Working Groups and Institutes after the decomposition of the activities.

 

3.4 Working Groups

The Working Groups are:

    1. Simulation:
    2. Computing Model Architecture
    3. Analysis and Network Process Design
    4. Implement TestBeds and Measure Critical Parameters.
    5. Steering Group

The interplay between the different Working Groups and their activities and the need to coordinate the decomposition/integration steps in the various iteration cycles, require that a detailed schedule be established which coordinates the timing of the relevant tasks. This schedule is therefore explained in the "Phases" section. A summary of the Milestones and the Gantt chart will be provided in the Schedule Chapter, after detailing the Tasks and SubTasks.

3.5 Phases of the Project

MONARC’s workplan is divided into two main phases, the first lasting 9 months, the second lasting 6 months. The phases are described in detail in Chapter 10.

 

3.6 Scope

3.7 Starting Conditions

Chapter 4: Task Definitions

4.1 Overview

The tasks are loosely defined in time order. Interdependency is quite strong between all the tasks. Although some can be run in parallel, constant feedback is needed between almost all of them.

  1. Simulation:
  1. Computing Model Architecture
  1. Analysis and Network Process Design
  1. Implement TestBeds and Measure Critical Parameters.
  1. Steering Group
    1. Parameters
    2. Models
    3. Metrics
    4. Experimental results and set-up
    5. Models and reccomendations
    6. Tools

4.2 Task 1: Simulation and Modelling

Realistic simulation and extensive modelling of the a priori possible distributed computing systems are the most important tasks in the first phase of the project. The goal is to be able to reliably model the behaviour of network given the assumed physical structure of a given computer system and the complicated resource use patterns, including third party traffic and the inherently stochastic manner in which hundreds of the physicists would access the data in their analyses of the LHC experiments. The network load, hardware and networking costs and the performance of a range of possible computer systems, measured by their ability to provide the physicists with the requested data in the required time are the main metrics in enabling to evaluate the results of a scan of multidimensional space of parameters which characterize the models. The goal is to narrow down a region in this parameter space in which viable models can be chosen by any of the LHC-era experiments.

FTE

Affiliation

1.2

Tufts

0.3

Caltech

 

 

1.5

TOTAL

Table 4.2.1 Effort allocated to the Simulation and Modelling Task

The planned research can be divided into the following sub-tasks, where the activities are expected to follow an iterative approach, with a complete cycle of the design-development-modelling-validation steps for every model studied and every simulation tool used.

4.2.1 SubTask: Survey of existing modelling tools.

Ideally one would like to use a small number (preferrably 2-3) tools to be able to cross-check the results. There exist a number of packages, for example, SoDA, MODNET, SES, COMNET, Ptolemy, Simple++, PARASOL; some of which have been looked at by various groups within MONARC. This work should be pursued vigorously, with a recommendation by the end of 1998, or so. SoDA, a simulation environment developed by Christoph von Praun at CERN IT, is at present the leading candidate to become one of the chosen packages.

4.2.2 SubTask: Conceptual design of the models which MONARC will explore.

It is essential to decide on the appropriate level of modelling complexity. This work should start immediately, as it must run in parallel with the task of development of the modelling tools. We anticipate that the models will include the details of data transfers from the disk to CPU, hardware configurations, complex network connections with varying availability of bandwidth depending on geographical topology, time and varying level of quality of service implementations. It is also essential to develop, as good as possible, models for data access patterns and analysis patterns. Here, the importance of input from physicists involved in experiments being part of MONARC cannot be underestimated. Also, experience from current, or near future, large statistics HEP experiments should be examined.

4.2.3 SubTask: Development of modelling packages or a combining existing tools.

Significant development work might be required to extend the existing SoDA class libraries in order to be able to describe the models which will be simulated. The goal is to have an advanced set of simulation tools by Spring of 1999.

4.2.4 SubTask: Modelling.

This sub-task will involve detailed simulations performed with the adopted set of tools and agreed upon different sets of input parameter values, in order to explore meaningfully the multidimensional parameter space of variables which describe the computer system models. The goal is to have be able to deliver to the experiments first reliable results by the Summer of 1999.
4.2.5 SubTask: Validation.

This important step is anticipated to be taking place in parallel with the first implementation of the models. We expect preparations for this sub-task to begin almost immediately. Designing the test-bed measurements with which one can verify the results of model simulations, at first simple and then more complex, is of paramount importance. There may be overlap here with designing of the test-bed measurements with which to improve our knowledge of a number of parameters which are needed as input to models of the computing systems.

4.2.6 Resources

The manpower needed to complete the simulation and modelling task is expected to vary with the project's progress. Initially, a significant number of people will have to be involved in the design stage of the simulation tools, as well as in defining the analyses and data access patterns which must be satisfied by the computing systems which will be explored in this project.

We expect of the order of 5 FTE be involved in this stage, although it may take many physical persons who will contribute only a fraction of their time. It is obvious that strong ties to the experiments will be very important in this initial stage.

We anticipate a need for 3 FTE, preferably 3 software professionals, to help with the design and development of the simulation tools. One person should be located at CERN, while the other two are anticipated to work in outside institutions. The search for a CERN-based modeller has already begun. Tufts University will co-ordinate the simulation and modelling work within the US-ATLAS groups, and they started the search for a full-time modeller, as well..

In the modelling phase we anticipate the need of about 4-5 FTE, with 3 modellers involved, and additional manpower coming from several people each contributing with a fraction of their time. Those people will most likely be the same group as those involved in the design stage.

Validation will be carried out by about 2-3 FTE. Again, we anticipate some overlap with the group of people who participated in the design stage, as well as with the researchers involved in test-bed measurements within this tasks project.

 

4.2.7 Milestones and Schedules

4.3 Task 2: Computing Model Architecture

This task addresses the design of the overall computing system for the Experiments. Components of the task include:

 

FTE

Affiliation

0.3

Caltech

0.3

Milano/INFN

0.1

CERN

0.4

Oxford

0.3

FNAL

0.2

Roma/INFN

0.3

Tufts

 

 

1.9

TOTAL

Table 4.3.1 Effort allocated to the Computing Model Architecture Task

 

4.4 Task 3: Analysis and Network Process Design.

Analysis Process is the ultimate goal of every experimental enterprise in the field of HEP while addressing the "Computing" design. This task of the project will address the definition of a few different approaches to the schemes of "way of doing analysis" among the many possibilities afforded by new (and perhaps unforeseen) computing technologies.

The task is a clear "user driven" one, since the Physicist Analyser has a large degree of freedom in his/her choices to get relevant results exploiting the better or more efficient (from his/her point of view) methods and using every resource he/she can catch. Nevertheless some procedures and priorities have to be fixed (even if they will change during the time of the experiment) in order to get reasonable analysis response in a reasonable competitive time.

The task will select some different scenarios for the Analysis Process, taking into account:

In addition, this task will address the definition of a set of values able to characterize the Analysis Processes in strict connection with the different Computing Model Architectures, being the correlation very strong.

In summary, the Task will identify possible Analysis Processes defining where the Raw data reside, how the Reconstructed Objects will be produced and stay, how and where the selection of relevant data for the analyses will be accessed, and finally where and how the Physicists will go through his/her analyses.

All of that will have to be coded and parameterized for in the Simulation.

A very simple, and probably unrealistic, example can better explain the task. Given a PByte/year of data, an Analysis Group goes through the Tag Database (produced quasi-online) selecting 1% of reconstructed objects created by the "Offline Reconstruction". They need to go through the full 1% sample once per month, producing an Analysis Object sample (reduced in number of events and in size, e.g.) which contains relevant analysis information, refined by experience and results from previous steps. Each component of the Group goes through the selected Analysis Objects by his/her own, with his/her own timing, trying to extract Results and/or personal sub-samples.

The timings, workaround, consistence, residence, storage, CPU needs and efficiency of such an example will be addressed by the present task which will try to schematically describe some different (and affordable) possibilities.

All the Analysis Processes chosen as possible and then to be simulated, either initially or later on during the Project, will have to be sketched in clear diagrams of serial and/or cycling steps.

 

FTE

Affiliation

0.4

Bologna/INFN

0.3

Caltech

0.3

Milano/INFN

0.5

FNAL

0.3

Tufts

 

 

1.7

TOTAL

Table 4.4.1 Effort allocated to the Analysis and Network Process Design Task

 

The task will be performed trough the following Work Packages or Subtasks which will embrace some or all of the Phases described in the Workplan Chapter:

 

4.4.1 SubTask : Identify Silly Models Early On by Using Backs of Envelopes

The aim of the task is to exclude a priori Analysis Models that lead to unbelievable and/or clearly unaffordable resources, even if projected to year 2005. E. g. an Analysis Process that will require either a local desktop storage of hundredth of Terabytes or a network bandwidth of more that a Gigabit/sec dedicated to every analyzing Physicist is clearly to be ruled out for the purpose of the Project.

  1. Required Resources.
  1. Task deliverables: A set of Analysis Processes, yet not complete by definition, having such strong requirements to result unfeasible. The Processes have to be expressed in terms of flow diagrams and in correlation with the Computing Model Architectures.

Task Milestones: Two milestones are settled.
The First during Phase 1A of WorkPlan : Exclude obvious unfeasible Analysis Processes.
The Second during Phase 1B/1C : Exclude less obvious unfeasible Analysis Processes profiting of the experience gained during the progress of the simulation.

  1. Task Duration/Schedule : According the previous Milestones and the WorkPlan Chapter.

 

4.4.2) SubTask : Identify User Requests.

The aim of the Task is to identify some (eventually different) schemes of "User's" needs while doing Analysis of Data. Some examples of that are the response time of a Database Query, the ability (and willingness) of doing queries locally or remotely (Regionally or Centrally), which tools and methods the End-User will adopt. As an example, the task will identify the methods that better fit the User behavior while doing analysis, trying to fix the range of parameters that will avoid refusal (and divergence) of the Analysis Model.

a) Required Resources:

b) Task Deliverables : A range of Constrained Parameters defining the User Behavior while doing Analysis.

c) Task Milestones : First milestone during Phase 1B for initial simulation. Second milestone during Phase 1C to refine different (some) ranges of parameters. Third milestone during Phase 2 to settle final range of affordable User's requests.

d) Task Duration/Schedule : The task will end with the Project, during Phase 2 (being eventually revisited for Phase 3) having a schedule as stated by the previous milestones.

 

4.4.3 SubTask : Analyze Existing Models.

The aim of the task is to extract information from the current Experiments' Analysis Processes. E. g. the number of concurrent Users doing analysis in present experiments will give some constraint (scaled to LHC) on the range of parameters to be simulated. Many other informations like the number of Analyzing Groups, their natural topology etc. can be investigated and recorded.

a) Required Resources:

b) Task Deliverables : A range of constrained parameters about the Groups of Analysis, the components of the Groups, the Number of Analysis, the number of concurrent Users, the distribution of Physicists in the Groups.

c) Task Milestones : The task has to be completed during Phase 1B at most, with a possible rescue during Phase 2.

d) Task Duration/Schedule : A strong relation is needed with other Experiments, mainly (but not only) those using Hadron Colliders. The duration/schedule is connected to the milestones previously stated and to the evolution of the Processes used by our Colleagues that may suggest us some modification to the project simulation.

 

4.4.4 SubTask : Identify feasible Models to be simulated.

The aim of the task is to identify a set of Analysis Processes that will meet the requirements coming from LHC data processing. The Models will have to be expressed in terms of clear diagrams. Such diagrams have to specify the input, output required parameters and resources, as well as the locations of data and computing power. A strong correlation between this task and the Computing Model Architectures is of course to be taken into account, being the Network one of the major concerns. As an example, where, when and how often the Physicist will access his/her own sample of selected data has to be designed; moreover the behaviors of the Analysis Groups has to be classified.

a) Required Resources:

b) Task Deliverables : A set of feasible Analysis Processes for LHC treatment of data, suitable for simulation according to the Computing Model Architectures defined by the present Project.

c) Task Milestones : The task will spread all over the duration of the Project.
A First milestone is foreseen for Phase 1B when an Analysis Process for first simulation is needed.
A Second milestone is due for Phase 1C when a set of Analysis Processes is needed to understand the possible spread of Models.
A Third milestone is due during Phase 2 in order to refine the Analysis Process possibilities taking into account the constraints given by the simulation performed and by the technology and budget limitations.
Finally another (or more) milestone can be foreseen just before the end of the Project or, eventually, for a possible extension to Phase 3.

d) Task Duration/Schedule : See milestones.

 

4.4.5 SubTask : Elaborate Policies, Priorities and Schedule for different Models.

The aim of the task is to study and then define different schemes of use of the LHC Collaborations Resources (Central and distributed) in order to guarantee that prioritized Analysis will get what they need in due time. Regional Centres access, use of the Network and Database distribution are the key parameters for this task. For example, when an Analysis item (and Group) get scientific interest over the others, the coordinating Analysis Team has to have the methods (and real technical authority) to set priorities on Collaborator's resources and needs a way to understand which schedule could give the result in time. While the skill of Physicists is out of our control, the overall capacity of the Computing System should be manageable.

a) Required Resources:

b) Task Deliverables : A set of rules to access Collaboration's Computing Resources, together with suggested methods to implement them. The rules have to be "tuned" to different Analysis Resources and to different Computing Model Architectures.

c) Task Milestones : The task will end with the Project, since the coordination and management of Resources is an item intrinsically embedded into the Analysis Process.
A First milestone can be foreseen for Phases 1B-1C, when some different schemes of Processes have to be simulated on distributed resources.
A Second milestone is certainly due for Phase 2 when priorities and schedules have to be taken into account to get reasonable simulation results.
Some more and refined milestones will be necessary for the evolution to Phase 3.

d) Task Duration/Schedule : According to previous milestones and to the evolution of distributed database Management Systems during the Project's life.

 

4.4.6 SubTask : Identify Key Parameters to evaluate Simulated Models.

The aim of the task is to establish which parameters are more sensitive for the evaluation of Simulated Analysis Models. This task is a clear hard one, since not only scientific (technological) parameters have to be accounted, but also "political" constraints have to be honored. The present Infrastructure, or the foreseen one, in different participating Countries (Laboratories, Computing Centres, Network, Man Power) have to be carefully considered.
An example can help understanding the task duties. The required money for implementing a specific Analysis Process (onto a given Computing Model Architecture) is only marginal if the ultimate goal of the Experiment at LHC is faced: all of the resources put on game have to have also a national, not only international, feedback.

Apart from that, the task will try to produce a set of sensible parameters that can give a "measurement" of the goodness of a specific Analysis Process. Such parameters are naively identified at present as Costs, Man Power, Flexibility, Reliability, Efficiency and finally Adherence to Physicist willing.

a) Required Resources:

b) Task Deliverables : A set of stringent (technically) and loose (Institution affordable) parameters to evaluate the different simulate Analysis Models.

c) Task Milestones : A First milestone stating a preliminary set of parameters is needed during Phases 1B/1C.
A Second milestone is due during Phase 2 to specify the results of the simulations in a measurable way.
A Third (and probably not definitive) milestone will be requested for Phase 3 of the Project.

d) Task Duration/Schedule : According to previous milestones.

 

4.4.7 SubTask : Coding the different Models.


4.4.8 SubTask : Evaluate the Models.

 

4.5 Task 4: Implement Testbeds and Measure Critical Parameters

In order to accomplish the analysis of computing models and to evaluate the impact of data distribution schemes, testbeds have to be implemented, depending on a set of well defined measurable parameters. Testbeds will be a limited size implementation of computing models as selected in Task 3. Task 4.5 defines the stages of this process.

 

FTE

Affiliation

0.3

Caltech

0.15

Milano/INFN

0.2

Bologna/INFN

0.6

Padua/INFN

0.2

FNAL

0.2

CERN

0.3

Roma/INFN

0.2

Tufts

 

 

2.15

TOTAL

Table 4.5.1 Effort allocated to the Implement Testbeds and Measure Critical Parameters Task

 

The task can be subdivided in following subtasks:

 

4.5.1 SubTask: Define Scope and Configuration of Testbeds

This subtask will implement several use cases based on different configurations for ODBMS distributed computing model, and data access patterns related to different functionalities like reconstruction and analysis.

Define which Testbeds will be implemented in relation to:

 

4.5.2 SubTask: Implementation and Operation of Testbeds

Test beds will involve many institutes that need to be ready for running the appropriate tests. There will be defined procedures for managing the global setup and facilitating test bed execution in remote sites. This global setup includes the configuration and management of hardware and software in all the involved regional centres.

This subtask includes:

 

4.5.3 SubTask Verify Key Parameters of Simulation for Each Testbed

This subtask is in collaboration with the simulation task. All parameters obtained in the simulation phase will be verified in the different test beds. We believe that parameters like network connection can not be properly simulated in the scope of this project. The fact that these test beds are relying on the client/server architecture of an ODBMS it is necessary to study also those configuration parameters related exclusively to the ODBMS. This step will be done in close collaboration with RD45 work.

Study parameters of ODBMS which are involved in the testbeds, i.e. considering overheads, network protocol latencies, etc.

Study parameters related to the network and connection between regional centres.

Manpower: (the number will come from the matrix of responsibility to be done.)

 

4.5.4 Required Resources

Equipment: the number will come from the matrix of responsibility

Software

Travel

Network Cost

 

4.5.5 Deliverables

packages etc..

 

4.5.6 Milestones

Each of these subtasks. Main milestone:

 

4.6 Task: Steering Group

  1. Parameters
  2. Models
  3. Metrics
  4. Experimental results and set-up
  5. Models and reccomendations
  6. Tools

FTE

Affiliation

0.1

Bologna/INFN

0.3

Caltech

 

 

0.4

TOTAL

Table 4.6.1 Effort allocated to the Steering Group Task

Chapter 5: Deliverables

 

Chapter 6: Resources

The resources required for the Project are broken down into the five principal areas of activity:

 

Chapter 7: Schedule

The project’s two principal phases are:

A third phase is now more clearly envisaged as a further R&D project, aimed at prototype designs and test implementations of the Computing Models. It will be the natural continuation of this project; it is however definitely separated from it in terms of deliverables. It should start at the completion of phase 2 and will have duration of a couple of years.

 

7.1 Phase 1

The workplan for Phase 1 is organized in three sub-phases

7.1.1 Phase 1-A, duration 2 months
This is a setup phase and the following activities shall be carried on in parallel in time ( not in isolation ):

are needed for starting the simulation and study their modelling.

7.1.2 Phase 1-B, duration 3 months
This is a startup phase with three activities going on in parallel:

These simulation startup steps should in principle be followed by every group member of the Simulation W.G.

7.1.3 Phase 1-C, duration 4 months
This phase is the first cycle of modeling, to be performed on a set choosen models "MS2-MSn". The modeling steps ( Code, Run, Analyze, Evaluate) will be preceeded by the selection of the range of Models ( analysis processes and system input parameters ), and followed by testbed validation of key results.
In parallel with the Simulation/Modeling and interacting strongly with it, the following activities will be pursued:

At the end of the modeling cycle, the results achieved shall be:

 

7.2 Phase 2

In this phase, two or three cycles of simulation/modelling will be performed. In the following list we summarise the items that will be addressed

At the end of Phase 2, the deliverables of this project will be made available:

The detailed workplan for Phase 2 will be provided later, in a progress report which will be delivered at the conclusion of Phase 1, in June 1999.

Chapter 8: Risk Identification

Chapter 9: Management and Organisational Responsibility

Chapter 10: References

 

Appendix 1: Detailed FTE breakdown (Internal Use Only)

FTE

Name(s)

Affiliation

1.2

Rolli, Sliwa, Somigliana, (modeller)

Tufts

0.3

Bunn, Galvez, Newman, Wilkinson, (DoE app.)

Caltech

 

 

 

1.5

TOTAL

 

Table 4.2.1 Effort allocated to the Simulation and Modelling Task

FTE

Name(s)

Affiliation

0.3

Bunn, Galvez, Newman, Wilkinson, (DoE app.)

Caltech

0.3

Battistoni, Campanella, Perini

Milano/INFN

0.1

Williams

CERN

0.4

Harris, McArthur

Oxford

0.3

Butler, Gaines, Pordes

FNAL

0.2

Barone

Roma/INFN

0.3

Rolli, Sliwa, Somigliana

Tufts

 

 

 

1.9

TOTAL

 

Table 4.3.1 Effort allocated to the Computing Model Architecture Task

FTE

Name(s)

Affiliation

0.4

Capiluppi, Giacomelli, Grandi

Bologna/INFN

0.3

Bunn, Galvez, Newman, Wilkinson, (DoE app.)

Caltech

0.3

Battistoni, Campanella, Perini

Milano/INFN

0.5

Butler, Gaines, Pordes

FNAL

0.3

Rolli, Sliwa, Somigliana

Tufts

 

 

 

1.7

TOTAL

 

Table 4.4.1 Effort allocated to the Analysis and Network Process Design Task

 

FTE

Name(s)

Affiliation

0.3

Bunn, Galvez, Newman, Wilkinson, (DoE app.)

Caltech

0.15

Battistoni, Campanella, Perini

Milano/INFN

0.2

Capiluppi, Giacomelli, Grandi

Bologna/INFN

0.6

Gasparini, Michelotto, Sgaravatto

Padua/INFN

0.2

Butler, Gaines, Pordes

FNAL

0.2

Holtman, Willers

CERN

0.3

Barone

Roma/INFN

0.2

Rolli, Sliwa, Somigliana

Tufts

 

 

 

2.15

TOTAL

 

Table 4.5.1 Effort allocated to the Implement Testbeds and Measure Critical Parameters Task

FTE

Name(s)

Affiliation

0.1

Capiluppi, Giacomelli, Grandi

Bologna/INFN

0.3

Bunn, Galvez, Newman, Wilkinson, (DoE app.)

Caltech

 

 

 

0.4

TOTAL

 

Table 4.6.1 Effort allocated to the Steering Group Task