Differences
This shows you the differences between two versions of the page.
Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
courses:cs211:winter2011:journals:camille:chapter4 [2011/02/21 22:16] – [Section 5: The Minimum Spanning Tree Problem] cobbc | courses:cs211:winter2011:journals:camille:chapter4 [2011/02/21 23:00] (current) – [Section 8: Huffman Codes and Data Compression] cobbc | ||
---|---|---|---|
Line 68: | Line 68: | ||
* **Summary: | * **Summary: | ||
+ | Oh hey the answer to other times that a minimum spanning tree can be useful! (We did this in class ... why didn't I still have to ask that question earlier haha?). Clustering comes up when you've got a collection of things and you're trying to separate them into coherent groups. How do you tell how similar objects are? The distance function - the farther away two objects are, the less similar they are (the closer objects are the more likely they are to be in the same cluster). If you're trying to find a clustering of maximum spacing, then you want the shortest distance between clusters to be as big as possible. They explain the "long way" of doing this, but basically if you want k clusters, you should delete the k biggest edges from the minimum spanning tree. Then those are the shortest distances between the clusters and they' | ||
* **Insights and Questions: | * **Insights and Questions: | ||
+ | They talk about the distance function being one way to cluster, but there must be other things to base it on. Also, in the context of pictures that they mentioned, the pixel stuff they talked about is probably not the best way to cluster pictures into similar groups the way you'd really want to in real life. Interesting stuff. | ||
* **Readability: | * **Readability: | ||
+ | Pretty easy to read. They sort of "went through with" the explanation of the "long way" (i.e. non-minimum-spanning tree explanation) to the solution more than we did before they started talking about the minimum spanning tree. I found that interesting. | ||
+ | ===== Section 8: Huffman Codes and Data Compression ===== | ||
+ | |||
+ | * **Summary: | ||
+ | This problem concerns data compression. In order to represent text in symbols, you have to change them into bits. One way to do this would be to use a fixed number of bits to represent each letter (ex. if you've got 32 symbols to represent, you'd use 5 bits for each letter). But we want to reduce the average number of bits that it takes to represent letters. You can take advantage of nonuniform frequencies in the number of times each letter comes up to do this (ex. the vowels come up a lot more than the letter ' | ||
+ | |||
+ | * **Insights and Questions: | ||
+ | Huffman codes may not always be the best for every set of data, but they are still used for compressing text, aren't they? Because the frequencies of letters in the English language is pretty consistent across different texts. | ||
+ | |||
+ | * **Readability: | ||
+ | This was a long section, but it was pretty easy to read, except the math they introduced with summations was a little hard to figure out. | ||
===== Section 9: Minimum-Cost Arborescences: | ===== Section 9: Minimum-Cost Arborescences: | ||