Mohammed and the Mou.. - Computer Engineering Research Group

Download Report

Transcript Mohammed and the Mou.. - Computer Engineering Research Group

Mohammed and the Mountain

Eric Jul DIKU Department of Computer Science University of Copenhagen 1

Mohammed and Mount Safa

IF THE MOUNTAIN WILL NOT COME TO MOHAMMED, MOHAMMED WILL GO TO THE MOUNTAIN "If one cannot get one's own way, one must adjust to the inevitable. “ The legend goes that when the founder of Islam was asked to give proofs of his teaching, he ordered Mount Safa to come to him. When the mountain did not comply, Mohammed raised his hands toward heaven and said, 'God is merciful. Had it obeyed my words, it would have fallen on us to our destruction. I will therefore go to the mountain and thank God that he has had mercy on a stiff-necked generation.” 2

Data and Computation

Source Destination Data Question: where should computation go?

Near the source? Or near the destination?

Or somewhere in the middle?

Shall the mountain of data go to the destination?

OR Shall the computation be moved toward the source?

3

Problem Statement

At which points should the data be processed?

And how much? (Just filtering? Content-based?) Close to destination: • may be network-inefficient Close to source: • more complicated • may require non-trivial cooperation in network 4

Main Point: Spectrum: Moving Data or Computation

Source Destination Data Spectrum ranging from moving all data to destination to moving all computation to the source Broker Publisher Broker Subscriber Broker 5

Overview

• Mobility of Data and Computation • Grid Computing Example • Mobile Objects in Emerald • Group Communication in Emerald • Mobile Grid Applications – Evil Man • Advice from experiences with dozens of Ph.D. projects 6

Mobility of Data and Computation

What is computation?

Pub-sub systems: just filtering?

Flooding vs. Match-first Flooding: filter at destination – simple, but network inefficient Match-first: network efficient Moving work closer to source as to reduce data transferred.

7

Motivating Example with huge data sets

The Large Hadron Collider at CERN ATLAS project 8

LHC – Large Hadron Collider

Atlas project

All of LHC…

Problem: Huge amounts of data

The problem is the large amount of data (Petabytes!) and the geographical distribution of the scientists wanting to look at / compute upon the data 12

LHC Data Distribution

CERN T2 … T2 T2 T2 T2 T2 T2 T1 T1 T1 T1 T1 T1 … T1 … T1 T1 T1 T1 T1 T1 T1 T1 Data is collected at CERN & distributed to ~12 data centers.

Data is then distributed to sub-centers.

13

Programs and Data Meet Up

Scientists can access the data and perform computation by submitting jobs to the sub-centers, e.g., via a Grid middleware So both computation and data are moved – and meet at Grid centers 14

Techniques to Consider

There are a number of general ideas/schemes to consider: • Compressing Data • Decoupling data and metadata, e.g., video & videoinfo • Content Routing Protocol (Pascal’s talk) • Caching • Multicast • Providing Mobile Objects and Mobile Computation 15

Mobile Objects

Problem: efficient placement of data and computation in a distributed system A solution: an Object-Oriented Language with: – objects – mobility of objects – transparent remote calls – group communication for pub-sub 16

Mobile Objects in Emerald

Developed 1984-1988 at the University of Washington Distributed language and run-time support system Emerald is an OO language: • “Pure” OO like Smalltalk – all data represented as objects (no primitive types) • Algol-family syntax (statements are NOT objects) • Process concept (threads) • Synchronization (Hoare monitors) • Conformity based type system (worth several talks in itself) • Like Java, but simpler 17

Distribution Features

Concept of location: A

node

is merely a machine (within a

semi-closed

network) • Mobility: move X to Y • Attachment allows groups to be moved • Location: loc <- locate X • “Remote” object invocation – transparent! X.f

• Checkpoint: stable version to disk • Node failure: failure handler, unavailability • Immutable objects (instead of primitives) Emerald Spring Cleaning Collector 18

Example: Kilroy

const Kilroy == object Kilroy process var i: Integer <- 0 var myNode: Node <- locate self var myList: Nodelist var remoteNode: Node myList <- myNode.getActiveNodes for (i <- 0; i < myList.upperbound; i <- i+1) remoteNode <- myList(i)$theNode

move

end loop Kilroy to remoteNode end process end Kilroy Emerald Spring Cleaning Collector 19

Example: Call By Move

const CallByMove == object CallByMove process var myData: LocalDataSet var myHomeNode: Node <- locate self var remoteNode: Node X.f[

move

myData]

move

self to X X.f[myData]

move

self to myHomeNode end process end Kilroy Emerald Spring Cleaning Collector 20

Example: Call By Visit

const CallByVisit == object CallByVisit process var myData: LocalDataSet var myHomeNode: Node <- locate self var remoteNode: Node X.f[

visit

myData] move self to X X.f[myData] move self to myHomeNode end process end Kilroy Emerald Spring Cleaning Collector 21

Group Communication

Added group communication to Emerald Introduce a

group

concept and a

group multi-cast call

Calling a function on a group object calls

all

the group members 22

Group Communication

const GroupExample == object GroupExample process var subscriberGroup: Group.of[Subscriber] subscriberGroup.Add[s1] ...

mySubscriberGroup <- subscribeGroup.getGroup

% using a group mySubscriberGroup.notify[data] end process end GroupExample Emerald Spring Cleaning Collector 23

New Problem Area

Switching to an entirely new problem area Problem: moving applications to Grid computing sites Trust code?

Trust Operating System?

Proposed solution – named Evil Man Move entire application

including

OS into a Grid Cluster 26

OS Migration

Desire: move entire running applications OS solution: pick up and move the

ENTIRE OS

First, use a Virtual Machine Monitor (Zen, VMWare) to separate OS from hardware Second, put migration code

inside

OS – it has all the needed functionality already – and get a

Self-migrating OS

27

Laundromat Computing with Evil Man

Evil Man

is based on

:

• Using virtual machines as containers for untrusted code • Using live VM migration to make execution independent of location • Using micro-payments for pay-as-you-go computing Evil Man uses Virtual Machine Migration to move actively executing applications into a Grid Cluster – e.g., to number crush data from the Atlas project at CERN 31

Evil Man

Prototype cluster management system developed at the Danish Center for Grid Computing at the University of Copenhagen One

great

Ph.D. student deserves most of the credit for the work:

Jacob Gorm Hansen

2008 Eurosys Rodger Needham award for Best Systems Ph.D. 32

Virtual Machine Migration

process process file VMM VMM • A VM is self-contained, including open files, sockets, shared memory • Interface to virtual hardware is clearly defined –

and simple

• All dependencies abstracted via fault-resilient network protocols • VMs typically long-lived 39

Why Migration Downtime Matters

•Upsets users of interactive applications such as games •May trigger failure detectors in a distributed system 44

Live Migration Reduces Downtime

•The VM can still be used while it is migrating •Data is transferred in the background, changes sent later 45

Self-Migration

Letting the OS move itself • Pre-copy migration relies on: – TCP/IP for transferring system state – Paging for tracking of write accesses • A VM is self-paging & has a TCP/IP stack • Reduce VMM complexity by performing migration

from within the VM

• No need for networking, threading or crypto in the TCB VM Migration TCP/IP Paging VM TCP/IP Paging VMM Hardware VM TCP/IP Paging 53

An Inspiring Example of Self-Migration

von Münchhausen in the swamp 54

First Iteration

57

Delta Iteration

58

Snapshot/Copy-on-Write Phase

59

Dealing with Network Side-effects

• The copy-on-write phase results in a

network fork

• “Parent” and “child” overlap and diverge • Firewall network traffic during final copy phase • All except migration-traffic is silently dropped in last phase 60

Impact of Migration on Foreground Load

httperf 61

Laundromat Computing

63

Pay-as-you-go Processing

• Laundromats do this already – Accessible to anyone – Pre-paid & pay-as-you-go – Small initial investment • We propose to manage clusters the same way – Micro-payment currency – Pay from first packet – Automatic garbage collection when payment runs out 64

Token Payments

• Initial payment is enclosed in Boot Token • Subsequent payments keep the VM alive as resources are consumed • Use a simple hash-chain for recurring payments – H n (s), H n-1 (s), …, H(s), s • Boot Token signed by trusted broker service • Broker handles authentication 65

Injecting a New VM

• Two-stage boot loader handles different incoming formats – ELF loader for injecting a Linux kernel image – Checkpoint loader for injecting a migrating VM • “Evil Man” service decodes Boot Token “magic ping” • Evil Man is 500 lines of code + network driver 66

Laundromat Summary

• Pros: – Simple and flexible model – Hundreds instead of millions LOC – Built-in payment system – Hypercube forks – Self-scaling applications • Cons: – Needs direct network access – Network load – Problems migrating across firewalls 67

Yet another switch

Some advice from my experience with a score of Ph.D. projects 68

Things to think about

• State your problem • Spread out initially, then focus • Stay close to your focus 69

State your problem

• Be able to answer the question: ”What is the problem you are trying to solve?” -- don’t end up with a ”solution in search of a problem!” • State your problem • Often good with a motivating example • Often good with a narrow, less-general example that also can be used as a milestone – even a killer app 70

Spread out initially, then focus

• Spread out, then narrow & focus (diagram) • Establish a dissertation title when you start to focus • Remember that your thesis focus is a SUBSET of all you have done – your dissertation is NOT a logbog of your work – nor a description of ”see what I have done” (diagram) 71

Stay close to your focus

• Write a draft of your conclusion early – use it for focus • Concentrate on your main contribution (diagram) • Try for a good, narrow example that is implementable and can be a good demo and an excellent milestone 72

Walk away with…

• Spectrum: moving data to computation site or moving computation to the data site • Spectrum underlies many decisions concerning architecture of pub-sub systems • Moving data is well-known; moving computation more tricky • Don’t have a solution in search of a problem • State the problem clearly – and state which part of it you will solve – use examples/killer apps • As you go narrow your focus • Shot for a good contribution 73

Questions?

74

The End

Thank you for

subscribing

 75