- Reminder: This is a double unit! So you should be spending 200 hours on it.
- 30 one hour lectures
- 10 two hour labs
- 150 hours you spend on your own!
- 12 hours a week!!
- Might want to spend one or two hours reading over the lecture notes every week.
- Being a
student is a full time job. 5 units take 100 hours each
over a 12 week period

- 500 / 12 = 42 hours a week.
- As teenagers you should probably be sleeping 8.5 hours a night -- 60 hours a week.
- This leaves you 66 hours a week to do what you like in!
- Even if you spend 3 hours a day doing meals, that leaves 45
other hours --- more than you need to spend on your studies.

- Unlike real jobs, you get to largely pick when you work (e.g. doing some on weekends rather than putting in a full 8 hour day every week day.)
- Unless you did no programming before you came here, first year is far easier than second year.
- All the years are roughly the same if you are struggling with
programming your first year.

- If you can't find time, you are probably having time management issues.
- University
of Bath's Student Time Management Page.

- Suggestions from Chicago (coincidently, where I went to Uni, though a few years ago it was just what came up first / best with Google).
- Definition of default (as in "default values" -- adjective / noun).
- Ask courswork questions on the Moodle forums
- This way tutors only have to answer a question once.
- This way everyone gets the same (and same quality) answer.
- You can get answers from each other
- The best way to learn is to teach.
- Tutors are less likely to be awake at 4am than undergraduates.

- Trees are a good way to think about both exponentiation & about logarithms.
- b = branching factor, d = depth of tree, N = number of elements.

o b

In general, N= b^{0}nodes in level 0

/ \

o o b^{1}nodes in level 1

/ \ / \

o o o o b^{2}nodes in level 2

/ \ / \ / \ / \

o oo oo oo o b^{3}nodes in level 3^{d}+ b^{d-1}+ ... b^{0}.

- As we know, this is dominated
by the largest term, b
^{d}(see lecture 3).

- Things would be different if we were multiplying, not adding.
- Factorial is when you multiply 1 * 2 * 3 * ... N

- So N grows exponentially on the depth of the tree,
- the base of the exponent is the branching factor of the tree.
- On the other hand, a length of the path from the root to the leaves of the tree is logarithmic on N,
- the branching factor is the base of the log too.
- N = O(b
^{d}), d = O(log_{b}(N)). In the case above, b =2, 8 = 2^{3}, 3 = log_{2}(8).

- Search is one of the most basic things you do with a computer --- finding the item you want.
- Data.
- Algorithm.
- Action
- Actually, search is one of the most basic things you do, even without a computer!
- Speeding up search is what computers are for.
- How long does it take you to search for one item in data that isn't sorted?
- Algorithm:
- Start at the beginning
- Go to the end
- Stop if you find an object that's key matches the one you are looking for.
- Analysis (equally likely to be any item)
- Best case: 1
- Worst case: N
- Average case: N/2
- What if its sorted?
- Algorithm:
- Start at the beginning
- Go till you find the place your key should be.
- Analysis (equally likely to be any item)
- Best case: 1
- Worst case: N
- Average case: N/2
- Not a big win! And notice, sorting took us at least n log
n time!

- What if you are looking for K items with the same key?
- Unsorted: K*N
- Sorted: Probably just N+K
- But both are O(N)

- If you have a sorted list, you can do better than to go through from one end to another.
- Think about how you might use a dictionary or a phone book (that doesn't have tabs!)
- Clue: try to think of a divide and conquer algorithm!
- Algorithm
- look in the middle of the list.
- If that value is your key, you're done!

- If your key is less than the current value, only consider the first half of the list, otherwise only consider the right side of the list.
- Go back to 1 with your new (half-sized) list.
- Example: 20, 35, 41, 48, 52, 76, 79, 84: Find the value 79!
- Center is 48 -- too little so go right
- Center of right is 76 -- too little so go right.
- Center of right of right is 79! Done in 3 steps (list size of 8.)
- Divide & Conquer should always give you log n performance.
- In case anyone's not following, I'll do this one in pseudo code
too,
but really you should be able to write pseudo code (& real code!)
by
now from the description of the algorithm.
integer key, sorta[] // what we are looking for? & the sorted array (don't have to be ints...)

boolean found // have we found it?

integer botIx, topIx // bottom and top boundaries of the search

integer midIx // middle of current search

integer locIx // for storing the correct location if we find it

botIx = 0, topIx = sorta.length - 1, locIx = -1, found = false; // initialize

while ((found == false) AND (botIx < topIx) do

midIx = (botIx + topIx) / 2; // returns the floor for ints

if (sorta[midIx] == key) {

found = true;

locIx = midIx; // all done!

}

else if (key < sorta[midIx]) {

topIx = midIx - 1;

}

else { // key > sorta[midIx]

botIx = midIx + 1;

}

} // end of while

return locIx; // will have value -1 if key was never found -- calling routine must handle this case! - This is a lot like the tree search I showed you Tuesday. You are basically using pointer arithmetic to replicate a perfectly-balanced tree.
- Note that we've done divide and conquer
here without recursion, just with a loop & some arithmetic on the
index values. Of course, recursion is simpler: now from the
description of the algorithm.

if (key == self.key) return self;

if (key < self.key)

if (not self.left) return null;

else return self.left.search(key)

if (not self.return) return null;

else return self.right.search(key - what is the expected search time? (all three cases, both versions
above)

- Now
if we are searching a lot more frequently than we are sorting, the cost
of the initial sort may not matter
much,
because the fact we have a sorted list allows us to search in just log
n
time!

- Can we do even better than log n? Yes, we can do constant
time!

- If you just use the key value as an array index, then you can sort in O(n) time and search in O(1) time.
- What if you have more than one value for a single index?
- Could make an array of lists, then search the list for the items you actually want.
- If only a few ever have the same index, this is still O(1), but if only a few indecies shared by all items, becomes O(n).
- That's a good indication you didn't choose a very good key!
- (draw picture of this)

- What if your keys are complicated values, e.g. words?
- Could still make into an integer, e.g. treat the word as a base 26 number.
- But the array would have to be really, really big!
- Most of it would probably be empty.
- Remember, using too much space slows you down too (swapping onto disk.)
- A common solution is a data structure called a hash table.
- Index is an array that's about the right size.
- Need a cunning way to convert the keys into then index arrays.
- This is called the hash function.
- Lots of different functions are used for hash functions.
- Should be something simple / fast so you can look things up quickly.
- Take an example using modular arithmetic.
- the mod is the remainder left over after an integer divide.
- good hash function
- division is pretty fast
- can choose the denominator for the mod to be the size of the array.
- List of values is 3, 10, 12, 20, 23, 27, 30, 31, 39, 42, 44, 45, 49, 53, 57, 60
- Let's try hashing it into an array of size 15 (same size as list)
0

1

2

3

4

5

6

7

8

9

10

11

12

13

14

30, 45, 60

31

3

49

20

23, 53

39

10

12, 27, 42, 57

44

- Analysis
- 10 items located in 1 step = 10

- 3 items in 2 steps = 6
- 2 items in 3 steps = 6
- 1 item in 4 steps
= 4

- So average time is (10 + 6 + 6 + 4 = 26) / 16 = 1.625 steps per retrieval
- Because hashing functions are usually random with respect to the input, you can expect:
- empty cells / wasted space
- collisions (where more than one item tries to get in one cell)
- have to make & search through list. This is called chaining,
and illustrated above.

- wasted time!
- Can use 2-dimensional array, but that wastes way more space, limits number of possible collisions
- Can indicate you should use of new hash function if too many
collisions
(e.g. do mod 20, not mod 15)

- More time efficient if less collisions, more space efficient if more collisions.
- Normally people favor time, make the table sparse (that is, much larger than the data):
- Still saves space over one big array indexed off keys.
- If the table is sparse enough, can put collisions in the next open index.
- Probably the simplest way to handle them.
- This is called probing.

- Let's try hashing it into an array of size 19 this time:
0

1

2

3

4

5

6

7

8

9

10

11

12

13

14

15

16

17

18

57

20 [39]

39

3 [60]

23 [42]

42 44

45

27

60

10

30 [49]

12 [31]

31

49

53

- Analysis
- 12 items located in 1 step = 12
- 3 items in 2 steps
= 6

- 1 item in 4 steps = 4
- 1 item in 6
steps
=
6

- So average time is 28 / 16 = 1.75 steps per retrieval
- You wouldn't really put the information in the square brackets
in the
table, I just did that to let you see where those numbers really wanted
to
be sorted originally. The table really would look like this:
0

1

2

3

4

5

6

7

8

9

10

11

12

13

14

15

16

17

18

57

20

39

3

23

42

44

45

27

60

10

30

12

31

49

53

- Algorithm to search:
- Derive a hash value mod tablesize from the search key, hashvalue.
- Search for key starting at searchIx = hashvalue.
- if sorta[searchIx] == null, then failed -- not in table.
- if sorta[searchIx] == key, then succeeded.
- otherwise, searchIx++, if searchIx > arraysize then searchIx =0 !! (wrap at the end!)
- go back to 6.2.1
- Analysis of Hash Search
- Want to know how many items have to be searched to find a value.
- As shown above, dependent on the load factor, A (really alpha, but HTML is lame.)
- A = number of occupied locations / table size
- On average, probability of clash is 1 / (1 - A).
- Successful search needs to examine (1 + 1 /(1 - A)) / 2 locations.
- So for example, an 80% full table requires about 3 steps, irrespective of how large the table is!
- Unsuccessful search needs to examine (1 + 1 / (1 - A)
^{2}) / 2. - So on an 80% full table, the average for an unsuccessful search is 13 steps, again regardless of table size.
- The normal rule of thumb is never to let a table get more than 80% full.
- Wikipedia has a nice illustration of why not.

- Don't forget about linear & binary sorts!
- Hashing is actually ubiquitous in computer science:
- How do you think your variable names are stored?
- Languages like python & perl let you use strings as array indecies.
- These are actually just hashes, like Java has, but they are
disguised
by the syntax (like java disguises the use of integer division with a
/).

22 February 2011