greedy algorithms 15-211 fundamental data structures and algorithms margaret reid-miller 25 march...

Post on 19-Jan-2016

220 Views

Category:

Documents

1 Downloads

Preview:

Click to see full reader

TRANSCRIPT

Greedy Algorithms

15-211 Fundamental Data Structures and Algorithms

Margaret Reid-Miller

25 March 2004

2

Announcements

• HW6 Parts 1 & 2 due Today, 11:59pm! All of Hw6 due Monday, April 5, 11:59

• Quiz #2 To be completed within one hour on a

web browser• actually, only a 15-minute quiz

Available Wednesday, 1:30pm Must be completed by Thursday,

April 1, 11:59pm

Greed is Good

4

Example 1: Counting change

• Suppose we want to give out change, using the minimal number of bills and coins.

5

A change-counting algorithm

• An easy algorithm for giving out N cents in change:

Choose the largest bill or coin that is N. Subtract the value of the chosen bill/coin from N,

to get a new value of N. Repeat until a total of N cents has been counted.

• Does this work? I.e., does this really give out the minimal number of coins and bills?

6

Our simple algorithm

• For US currency, this simple algorithm actually works.

• Why do we call this a greedy algorithm?

7

Greedy algorithms

• At every step, a greedy algorithm

makes a locally optimal decision,

with the idea that in the end it all adds up to a globally optimal solution.

• Being optimistic like this usually leads to very simple algorithms (i.e., easy to code).

8

Change counting is greedy

• Makes a locally optimal decision.

Uses the next-largest bill or coin. But once a coin is accepted, it is permanently included in the solution. Once a coin is rejected, it is permanently excluded from the solution.

• To reach a globally optimal solution.

Can you prove it for US currency?

9

But…

• What happens if we have a 12-cent coin?

10

Hill-climbing

• Greedy algorithms are often visualized as “hill-climbing”. Suppose you want to reach the summit,

but can only see 10 yards ahead and behind (due to thick fog).

Which way?

11

Hill-climbing

• Greedy algorithms are often visualized as “hill-climbing”. Suppose you want to reach the summit,

but can only see 10 yards ahead and behind (due to thick fog).

Which way?

12

Hill-climbing, cont’d

• Making the locally-best guess is efficient and easy, but doesn’t always work.

13

Example 2: Fractional Knapsack Problem (FKP)

• You rob a store: find n kinds of items Gold dust. Wheat. Beer.

15

Example 2: Fractional knapsack problem (FKP)

• You rob a store: find n kinds of items Gold dust. Wheat. Beer.

• The total inventory for the i th kind of item:

Weight: wi pounds Value: vi dollars

• Knapsack can hold a maximum of W pounds.

• Q: how much of each kind of item should you take?

(Can take fractional weight)

16

FKP: solution

• Greedy solution 1: Get a bigger knapsack! Build up extra muscles if necessary.

17

FKP: solution

• Greedy solution 1: Get a bigger knapsack! Build up extra muscles if necessary.

But seriously folks…

18

FKP: solution

• Greedy solution 1: Get a bigger knapsack! Build up extra muscles if necessary.

But seriously folks…

• Greedy solution 2: Fill knapsack with “most valuable” item

until all is taken. •Most valuable = vi /wi (dollars per pound)

Then next “most valuable” item, etc. Until knapsack is full.

19

Ingredients of a greedy alg.

1. Optimization problem Of the many feasible solutions, finds the

minimum or maximum solution.

2. Can only proceed in stages no direct solution available

3. Greedy-choice property:A locally optimal (greedy) choice will lead to a globally optimal solution.

4. Optimal substructure:An optimal solution contains within it optimal solutions to subproblems.

Show the proof

20

FKP is greedy

• An optimization problem: Maximize value of loot, subject to

maximum weight W. (constrained optimization)

• Proceeds in stages: Knapsack is filled with one item at a

time.

21

FKP is greedy

• Greedy-choice property: A locally greedy choice will lead to a globally optimal solution.

• Proof:Step 1: Prove that an optimal solution contains the greedy choice.

Step 2: Prove that the greedy choice can always be made first.

22

FKP: Greedy-choice proof: step 1

We want to show that the optimal solution alwayscontains the greedy choice.

• Consider total value, V, of knapsack.

• Knapsack must contain item h: Item h is the item with highest $/lb.

• Why? Because if h is not included, we can replace some other item in knapsack with an equivalent weight of h, and increase V.

• This can continue until knapsack is full, or all of h is taken.

Therefore any optimal solution must include greedy-choice.

23

For item i let wi be the total inventory, vi be the total value, ki be the weight in knapsack.

Let item h be the item with highest $/lb. If kh<wh, and kj>0 for some jh, then replace j with an equal weight of h. Let new total value = V’.

Difference in total value:

since, by definition of h,

Therefore all of item h should be taken.

More rigorously…

∑=

⎟⎟⎠

⎞⎜⎜⎝

⎛=

n

i i

ii w

vkV

1

0≥⎟⎟⎠

⎞⎜⎜⎝

⎛−⎟⎟

⎞⎜⎜⎝

⎛=−′

j

jj

h

hj w

vk

wv

kVV

h

h

j

j

w

v

w

v≤

Assume total optimal value:

24

FKP: Greedy-choice proof: step 2

Next, we want to show that we can always make the greedy choice first.

• If item h is more than what knapsack can hold, then fill knapsack completely with h. No other item gives higher total value.

• Otherwise, knapsack contains all of h and some other item. We can always make h the first choice, without changing total value V.

In either case the greedy-choice can always be made FIRST.

25

More rigorously…

• Case I: wh W Fill knapsack completely with h. No other item gives higher total value.

• Case II: wh < W Let 1st choice be item i, and kth choice be h,

then we can always swap our 1st and kth choices, and total value V remains unchanged.

• Therefore greedy-choice can always be made first.

26

FKP: Optimal substructure

• The optimal substructure property:An optimal solution contains within it optimal solutions to subproblems.

• If we remove weight w of one item i from the optimal load, then the remaining load must be optimal solution using the remaining items.

• The subproblem is the most valuable load of maximum weight W-w from n-1 items and wi - w of item i.

27

FKP: Optimal substructure proof

• We want to show that an optimal solution contains within it optimal solutions to subproblems.

• Consider the most valuable load L, weighing W lbs.

• Remove w pounds of some item i .

• Remaining load L’ must be the most valuable load for a smaller fractional knapsack problem:

Maximum weight is W – w lbs.

Only n-1 items and wi - w lbs. of item i.

• Why? Because otherwise we can find a load, L’’, more valuable than L’, add w pounds of item i, and this will be more valuable than L. (Contradiction!)

28

Example 3: Binary knapsack problem (BKP)

• Variation on FKP. The “Supermarket Shopping Spree”!

• Suppose, instead, that you can only take an item wholly, or not at all (no fractions allowed). Diamond rings. Laptops. Watches.

• Q: How many of each item to take?

• Will the greedy approach still work?

Surprisingly, no.

29

The Binary Knapsack Problem

• You win the Supermarket Shopping Spree contest. You are given a shopping cart with

capacity C. You are allowed to fill it with any

items you want from Giant Eagle. Giant Eagle has items 1, 2, … n,

which have values v1, v2, …, vn, and sizes s1, s2, …, sn.

How do you (efficiently) maximize the value of the items in your cart?

                

31

BKP is not greedy

• The obvious greedy strategy of taking the maximum value item that still fits in the cart does not work.

• Consider: Suppose item i has size si = C and value

vi. It can happen that there are items j and

k with combined size sj+sk C but vj+vk > vi.

32

BKP: Greedy approach fails

item 1 item 2 item 3 knapsack

$60, 10 lbs

$100, 20 lbs

$120, 30 lbs

Maximum weight = 50 lbs

Dollars/pound

Item 1 $6

Item 2 $5

Item 3 $4

BKP has optimal substructure, but not greedy-choice property: optimal solution does not contain greedy choice.

$160 $180 $220

(optimal)

33

Succeeding with greed.

4 ingredients needed:1. Optimization problem

2. Can only proceed in stages

3. Greedy-choice property:A greedy choice will lead to a globally optimal solution.

4. Optimal substructure:An optimal solution contains within it optimal solutions to subproblems.

Data Compression

35

Example 4: Huffman coding

• Compresses data using prefix codes, where no codeword is a prefix of some other code word.

• Huffman codes are optimal prefix codes.

• Huffman algorithm is greedy.

36

Example

37

Example

38

Example

39

Huffman coding is greedy

• Optimization problem: Find a prefix code to compress a file to a minimum size.

• Proceeds in stages: For each iteration “merge” two trees with the minimum

frequency.

• A greedy choice will lead to a globally optimal solution: There exists an optimal prefix code in which the two

characters with minimum frequency had codewords that differ only in the last bit.

• An optimal solution contains optimal solutions to subproblems: Replace the two characters with minimum frequency

with a new character with frequency that is the sum of the frequencies of the replaced characters.

40

Huffman: Greedy-choice property

• Suppose tree T represents an optimal prefix code.

• Modify T so that the two characters with lowest frequencies, x and y, are sibling leaves of maximum depth.

• Swap x and y with leaves that are of maximum depth. Each swap exchanges a lower frequency letter with a higher frequency letter that has a longer code. Thus, each swap does not increase the cost of compression.

41

Huffman: optimal substructure

• If we replace any two characters of sibling leaves in T with a character with same total frequency, then the new tree T’ is an optimal prefix code with the new characters.

B(T) = 45*1+12*3+13*3+5*4+9*4+16*3

z:14

B(T’) = 45*1+12*3+13*3+(5+9)*3+16*3= B(T) - 5 - 9

Minimum Spanning Trees

43

Prim’s algorithm

• Pick a vertex now we have a minimum spanning tree

of the chosen vertex

• Grow the tree until all vertices are included Each step, add the edge (u,v) s.t. the

cost of (u,v) is minimum among all edges where u is in the tree and v is not in the tree

• Prim’s algorithm is greedy.

44

e a b c d

0

Prim’s algorithm

a

ce

d

b2

44

9

6

4

5

3

b d c a

3 4 5

Vertex Parente -b ec ed e

The MST initially consists of the vertex e, and we updatethe distances and parent for its adjacent vertices

Vertex Parente -b -c -d -

45

Prim’s algorithm

a

ce

d

b2

44

9

6

4

5

3

d c a

4 5 9

Vertex Parente -b ec ed ea b

b d c a

3 4 5

Vertex Parente -b ec ed e

46

Vertex Parente -b ec ed ea b

Prim’s algorithm

a

ce

d

b2

44

9

6

4

5

3

a c

2 4

Vertex Parente -b ec dd ea d

d c a

4 5 9

47

Another Approach – Kruskal’s

a

ce

d

b2

44

9

6

4

5

3

• Create a forest of trees from the vertices.• Repeatedly merge trees by adding “safe edges”

until only one tree remains.• A “safe edge” is an edge of minimum weight which

does not create a cycle.

forest: {a}, {b}, {c}, {d}, {e}

Kruskal’s algorithm is greedy.

48

Kruskal’s algorithm

For each edge (u,v) E in increasing order while more than one set remains:

If u and v, belong to different sets a. A = A {(u,v)} b. merge the sets containing u and v

Return A

• Use Union-Find algorithm to efficiently determine if u and v belong to different sets.

49

Kruskal’s algorithm

E = {(a,d), (b,e), (c,d), (a,c), (d,e), (c,e), (b,d), (a,b)}

Forest{a}, {b}, {c}, {d}, {e}{a,d}, {b}, {c}, {e}

T{(a,d)}

a

ce

d

b2

44

9

6

4

5

3

50

Kruskal’s algorithm

E = {(a,d), (b,e), (c,d), (a,c), (d,e), (c,e), (b,d), (a,b)}

Forest{a}, {b}, {c}, {d}, {e}{a,d}, {b}, {c}, {e}{a,d}, {b,e}, {c}

T{(a,d)}{(a,d), (b,e)}

a

ce

d

b2

44

9

6

4

5

3

51

Kruskal’s algorithm

E = {(a,d), (b,e), (c,d), (a,c), (d,e), (c,e), (b,d), (a,b)}

Forest{a}, {b}, {c}, {d}, {e}{a,d}, {b}, {c}, {e}{a,d}, {b,e}, {c}{a,d,c,}, {b,e}

T{(a,d)}{(a,d), (b,e)}{(a,d), (b,e), (c,d)}

a

ce

d

b2

44

9

6

4

5

3

Weighted Single-SourceShortest Path Algorithm

(Dijkstra’s Algorithm)

53

Dijkstra’s Single-Source Shortest Path

s

f

a

b

d

e

c

g

4

2

5

1

11

4

2

2

Given: a directed graph G = (V,E)weight(u,v) 0 for all edges (u,v) E

Find: shortest path from start vertex s to every other vertex in G

Dijkstra’s algorithm is greedy.

54

Dijkstra’s algorithm

s a b c d e f g

0

Visited

s

f

a

b

d

e

c

g

4

2

5

1

11

4

2

2

55

Dijkstra’s algorithm

Visiteds (D = 0)

b c a d e f g

2 4 5

s

f

a

b

d

e

c

g

4

2

5

1

11

4

2

2

56

Dijkstra’s algorithm

Visiteds (D = 0)b (D = 2)s

f

a

b

d

e

c

g

4

2

5

1

11

4

2

2

d c a e f g

3 4 5 6

57

Greedy Algorithms

• Many optimization problems can be solved using a greedy approach. The basic principle is that local optimal

decisions will lead to a global optimal solution.

But the greedy approach may not always lead to an optimal solution for all problems.

The key is knowing which problems will work with this approach and which will not.

• Next time we will look at problems for which it will not work.

top related