The book is almost out there. There is code and selected solutions as well. The book is supposed to be in full colour from what I heard. What do you think of the code? There is an interesting interview as well.
Balbir's Blog
I describe all the programming and non-programming stuff I do
Sunday, March 27, 2022
Defer style programming in C
It's a well known trick, but I loved the ability to defer functions in golang. In C, gcc provides an attribute for cleanup. I posted a pull request for the libcgroup. Turns out systemd has something similar, but I do like that the attributes have implicit functions and the overall design seems cleaner. I do not like dereferencing a double void star pointer though.
More details can be found in the pull request
Sunday, November 28, 2021
Test Driven Development
I've been learning rust programming language. The addition of the ability to write tests in-built is quite a welcome change to most programming languages. I found that I spent a lot of time debugging corner cases (bounds) in my algorithms, so I decided to do a test driven approach (not a chaos/random approach), but more targetted towards my weakness with boundary conditions. My favourite small/short/complex algorithm is quicksort.
Here is the code I wrote
- The use of the parameter high in the algorithm is unique, high is always one greater than the last element, so should the contract abstract it and expect a pointer to the last element? Notice the use of high-1 more than once in the code.
- Rust itself has some quirks associated with the Borrow Checker, after taking a mutable reference to the vector, one cannot use elements.len(), since that internally grabs a shared reference (the compiler tells you).
- The while loops use elements[i] <= elements[low], even though elements[low] is constant, using it before the loop causes a move of the element to the temporary variable used to cache it and that causes the Borrow Checker to complain, it might be possible to do some interesting things by having the template type T have a constraint that requires it to implement the Copy trait.
- Typically quicksort partitions the elements at p and do a quicksort with 0, p-1 and p+1 to high, but since the bounds is [low, high), we need to ensure we pass p as oposed to p-1
1| 1|use std::cmp::PartialOrd;use std::cmp::PartialOrd;
2| |
3| |/// Quicksort
4| |/// Simple sort algorithm that is type aware and is unique in that it's test driven
5| |///
6| 49|fn partition<T: PartialOrd>(elements: &mut [T], low: usize, high: usize) -> usize {
7| 49| // Explictly set the pivot to the lowest element
8| 49| let mut i : usize;
9| 49| let mut j : usize;
10| 49|
11| 49| if high <= low {
12| 15| return low;
13| 34| }
14| 34|
15| 34| i = low; j = high -1; // Bounds 1, check
16| |
17| 37| while i < j {
18| 63| while elements[i] <= elements[low] && i < (high-1) { i = i + 1; } // Bounds 2
^45 ^36
19| 65| while elements[j] >= elements[low] && j > low { j = j - 1; } // Bounds 3
^50 ^38
20| |
21| 27| if i >= j {break; } // Bounds 4
^24
22| 3|
23| 3| elements.swap(i, j); // Check 5
24| | }
25| 34| elements.swap(low, j); // Check 6
26| 34| j
27| 49|}
28| |
29| 43|fn __quicksort<T: PartialOrd + std::fmt::Debug>(elements: &mut [T], low: usize, high: usize)
30| 43|{
31| 43| let p = partition(elements, low, high);
32| 43| // p - 1 should not be negative
33| 43| if p > low {
34| 9| __quicksort(elements, low, p);
35| 34| }
36| 43| if p < high {
37| 28| __quicksort(elements, p+1, high);
38| 28| }
^15
39| 43|}
41| 6|fn quicksort<T: PartialOrd + std::fmt::Debug>(elements: &mut [T])
42| 6|{
43| 6| let len = elements.len();
44| 6|
45| 6| __quicksort(elements, 0, len);
46| 6|}
Not bad, but the branch coverage needs improvement. This variant of quicksort is prone to several errors, although it might seem faster at first look. Interestingly, the algorithms book by Jeff Erickson very correctly states
Hoare proposed a more complicated “two-way” partitioning algorithm that has some practical advantages over Lomuto’s algorithm. On the other hand, Hoare’s partitioning algorithm is one of the places off-by-one errors go to die
I do find the Lomuto's algorithm easier to parse and implement. What do you think? Is test-driven iterative development the right way forward for even things like data structures? What's the best way to compute the total number of branches and cover them? Should we search for easier alternatives that perform well and are easier to test?
Update: I made some slice based improvements to the code, with the Lomuto algorithm as well
/// Quicksort
/// Simple sort algorithm that is type aware and is unique in that it's test driven
///
fn partition<T: PartialOrd>(elements: &mut [T]) -> usize {
// Explictly set the pivot to the lowest element
let mut i : usize;
let mut j : usize;
if elements.is_empty() { // Bounds 1
return 0;
}
let high = elements.len();
i = 0;
j = high - 1;
while i < j { // Bounds 2
while elements[i] <= elements[0] && i < (high-1) { i += 1; } // Bounds 3
while elements[j] >= elements[0] && j > 0 { j -= 1; } // Bounds 4
if i >= j {break; } // Bounds 5
elements.swap(i, j); // Check 6
}
elements.swap(0, j); // Check 7
j
}
fn quicksort<T: PartialOrd + std::fmt::Debug>(elements: &mut [T])
{
let low = 0;
let high = elements.len();
if elements.is_empty() {
return;
}
let p = partition(&mut elements[low..high]);
// p - 1 should not be negative
quicksort(&mut elements[low..p]);
quicksort(&mut elements[p+1..high]);
}
///
/// Use the algorithm that has fewer boundary conditions to worry about
fn lomuto_partition<T: std::cmp::PartialOrd>(elements: &mut [T]) -> usize
{
let mut i: usize = 0;
let size: usize = elements.len();
let mut j : usize = 0;
if elements.is_empty() { // bounds 1
return 0;
}
while i < size-1 { // bounds 2
i += 1;
if elements[i] <= elements[0] { // bounds 3
j += 1;
elements.swap(i, j); // bounds 4
}
}
elements.swap(0, j); // bounds 5
j
}fn quicksort<T: PartialOrd + std::fmt::Debug>(elements: &mut [T])
{
let low = 0;
let high = elements.len();
if elements.is_empty() {
return;
}
let p = lomuto_partition(&mut elements[low..high]);
// p - 1 should not be negative
quicksort(&mut elements[low..p]);
quicksort(&mut elements[p+1..high]);
}
Sunday, June 07, 2020
Volume of a unit hyper sphere of radius 1
(%i17) integrate(integrate(integrate(1, z, -sqrt(1-x^2-y^2), sqrt(1-x^2-y^2)),
y, -sqrt(1-x^2), sqrt(1-x^2)), x, -1, 1);
"Is "(x-1)*(x+1)" positive or negative?"For a circle this value is definitely negative and voila we get the answer as $4\pi/3$. Higher dimenisons lead to interesting results
integrate(integrate(integrate(integrate(1, x4, -sqrt(1-x1^2-x2^2-x3^2),
sqrt(1-x1^2-x2^2-x3^2)), x3, -sqrt(1-x1^2-x2^2),
sqrt(1-x1^2-x2^2)), x2, -sqrt(1-x1^2), sqrt(1-x1^2)),
x1, -1, 1);
Saturday, May 30, 2020
Passive reaction: Computer Science has changed/changing
"Courses in theoretical computer science covered finite automata, regular expressions, context-free languages, and computability. In the 1970’s, the study of algorithms was added as an important component of theory. The emphasis was on making computers useful. Today, a fundamental change is taking place and the focus is more on a wealth of applications. There are many reasons for this change. The merging of computing and communications has played an important role. The enhanced ability to observe, collect, and store data in the natural sciences, in commerce, and in other fields calls for a change in our understanding of data and how to handle it in the modern setting. The emergence of the web and social networks as central aspects of daily life presents both opportunities and challenges for theory.
While traditional areas of computer science remain highly important, increasingly researchers of the future will be involved with using computers to understand and extract usable information from massive data arising in applications, not just how to make computers useful on specific well-defined problems. With this in mind we have written this book to cover the theory we expect to be useful in the next 40 years, just as an under standing of automata theory, algorithms, and related topics gave students an advantage in the last 40 years. One of the major changes is an increase in emphasis on probability, statistics, and numerical methods.
...
Modern data in diverse fields such as information processing, search, and machine learning is often advantageously represented as vectors with a large number of components"
Sounds like I am a student of the past that focused more on Automata Theory and algorithms (traditional) in the past. I am yet to catch up with the mentioned emergence and opportunities topics. Time to learn new things, but it's going to be tricky to do it by myself, but I am glad to see someone thinking of the next 40 years.
Sunday, November 03, 2019
Frustration with the state of math in schools
NOTE: I had learnt the formula for squares and cubes without necessarliy mastering how we arrived at the differential, it was algorithm/formula for me to follow
My frustration is with the math of today, some with the teachers where I put my son for some extra classes and he was asked to remember formulae without detailed concepts. I see why those topics are important, but it seems like:
- Teachers who want to clarify concepts and teach nicely are too slow in catching up with topics to be covered
- Teachers catching up with topics and keeping good pace, don't spend enough time on concepts
Sunday, July 09, 2017
Dynamic programming for the binomial coefficient
# # dynamic programming for binomial co-efficient # def dynamic_bc(n, k, indent=0): # # return (n k) using dynamic programming # tail cases are k == 0 or k == 1 # (n 0) = 1 and (n 1) = n for i in range(0, indent): print " ", print "%d:%d" %(n, k) if (k == 0): arr[n][k] = 1 return 1 if (k == 1): arr[n][k] = n return n if (n == k): arr[n][k] = 1 return 1 # (n k) = (n!)/(n-k)!k!, lets split it further # n!/(n-k)!k! = n(n-1)!/(n-k)!k(k-1)! # or (n k) = (n-1 k) + (n -1 k-1) if (arr[n-1][k-1] == 0): arr[n-1][k-1] = dynamic_bc(n-1,k-1,indent+1) if (arr[n-1][k] == 0): arr[n-1][k] = dynamic_bc(n-1, k,indent+1) return arr[n-1][k] + arr[n-1][k-1] def bc(n, k, indent=0): # # tail cases are k == 0 or k == 1 # (n 0) = 1 and (n 1) = n for i in range(0, indent): print " ", print "%d:%d" %(n, k) if (k == 0): return 1 if (k == 1): return n if (n == k): return 1 # (n k) = (n!)/(n-k)!k!, lets split it further # n!/(n-k)!k! = n(n-1)!/(n-k)!k(k-1)! # or (n k) = (n-1 k) + (n -1 k-1) return bc(n-1,k,indent+1) + bc(n-1,k-1,indent+1) number = 6 select = 3 arr = [[0 for x in range(0, number)] for x in range(0, number)] print bc(number, select) print dynamic_bc(number, select)The output for the first call with full recursion
6:3 5:3 4:3 3:3 3:2 2:2 2:1 4:2 3:2 2:2 2:1 3:1 5:2 4:2 3:2 2:2 2:1 3:1 4:1 20The output for For the first call with full recursion
6:3 5:2 4:1 4:2 3:1 3:2 2:1 2:2 5:3 4:3 3:3 20Python supports memoization via functools (wraps, lru_cache, etc). I am using the wrapper (decorator pattern). Using the following pattern makes the programming so transparent
# # dynamic programming for binomial co-efficient # from functools import wraps def dynamic_bc2(func): cache = {} def wrap(*args): if args not in cache: cache[args] = func(*args) return cache[args] return wrap @dynamic_bc2 def bc2(n, k, indent=0): # # return (n k) using dynamic programming # tail cases are k == 0 or k == 1 # (n 0) = 1 and (n 1) = n for i in range(0, indent): print " ", print "%d:%d" %(n, k) if (k == 0): return 1 if (k == 1): return n if (n == k): return 1 # (n k) = (n!)/(n-k)!k!, lets split it further # n!/(n-k)!k! = n(n-1)!/(n-k)!k(k-1)! # or (n k) = (n-1 k) + (n -1 k-1) return bc2(n-1,k-1,indent+1) + bc2(n-1,k,indent+1) number = 6 select = 3 print bc2(number, select)Comparing the output will show the affect of functools. The output with functools is:
6:3 5:2 4:1 4:2 3:1 3:2 2:1 2:2 5:3 4:3 3:3 20
Thursday, July 06, 2017
Iterative combinations algorithm
# # Do iterative combinations generation # From Knuth's observation, we know for (6 3) that # [0, 1, 2] # [0, 1, 3] # .. # [3, 4, 5] # Basically we can see a set of for loops, if we call the columns c1,c2 and c3 then # # for c3 = 2 to n-1 # for c2 = 1 to c3-1 # for c1 = 0 to c2-1 # # The loop gives us what we need # My implementation is based on an index (i) and max (n) # which are used to determine the next value to generate # def iterative_combination(n, k): a = [i for i in range(0, n)] index = k - 1 done = False while (done == False): print a[0:k] while (a[index] == n - (k - index)): # boundary for that index index = index - 1 if (index < 0): done = True break a[index] = a[index] + 1 while (index < k - 1): index = index + 1 if (a[index] == (n - (k - index))): # initialize the next neighbour a[index] = a[index - 1] + 1
Random CS picture
The picture is a counter example of why greedy selection does not work optimally for the set cover problem. If the picture seems not so well done, it's because my asymptote skills are lacking :)
You can probably guess what I'm reading by connecting set cover with combinatorial generation of combinations.
Introduction to Algorithms Fourth Edition (almost out, so is the code)
The book is almost out there . There is code and selected solutions as well. The book is supposed to be in full colour from what I heard....

-
I have been extremely busy to keep this blog upto date. Well, I have been bitten by the "busy" virus. Yes, I have been extremely b...
-
(Photo from http://content-usa.cricinfo.com/indvaus2008/content/current/player/28114.html) Dravid's dismal form continues in test crick...
-
The book is almost out there . There is code and selected solutions as well. The book is supposed to be in full colour from what I heard....