# Inside Class Cruncher

You’ll find here the concepts used to create Class Cruncher, a handy webapp for school administrators.

# The problem

How should you deploy a certain number of teachers across your school’s classes, given rules about class sizes? You may need to use composite classes.

More precisely, given:

• The total number of teachers
• How many children are in each grade
• Min & max class sizes in each grade
• Which composite classes are allowed (eg. grades 1 & 2 but not grades 1 & 3 together)
• The smallest number of children from a grade allowed in a composite class (eg. so you don’t have a single lonely first grader in a class with grade 2).

Then list out the classes that each teacher should take (i.e. how many children from each grade are in each class).

# A solution

Solve this as a mixed-integer linear programming problem, using lpsolve. This uses the simplex and branch-and-bound algorithms to maximise an objective function subject to some constraints. Here we have lots of constraints, and just want a feasible solution, rather than an “optimal” solution, so the objective function is not so important.

## Step 1: Choose your variables

For every potential class j, we’ll have a binary variable y(j) (ie. it must be either 0 or 1), to tell us if it is being used or not.

Because classes can include students from different grades, we will invent the term “subclass” to refer to the number of children from one particular grade in a class. Then our key variables are x(i), the number of children in each potential subclass i.

Let’s relate the classes and subclasses via a matrix C(j,i), with 1 if subclass i is in class j, and 0 otherwise. Since subclasses are only in a single class, for any subclass i, C(j,i) is 1 for only one class j. So C is mostly 0s, with each row having at least one 1, and no columns having more than one 1.

Then the number of children in class j = the sum over i of C(j,i).x(i).

(In matrix notation, Cx = number of children in each class, where x and the right hand side are column vectors.)

C(j,i) also tells us if subclass i is being used: it is given by the sum over j of C(j,i).y(j).

Recall that y(j) tells us if class j is being used. Also, for a given subclass i, C(j,i) is only 1 for a single class j, so the sum is just a single term.

(In matrix notation, C’y = binaries telling if each subclass is used, where C’ is the transpose of C.)

For example: Suppose you have 3 grades (kindergarten, 1 and 2), and can allow for 1 pure kindy class, 2 composite K/1 classes, 2 pure 1st grade and 1 pure 2nd grade class. That’s a total of 6 classes and 8 subclasses. The matrix C is:

—-8 subclasses—–
| 1 . . . . . . .
| . 1 1 . . . . .
6 . . . 1 1 . . .
classes . . . . . 1 . .
| . . . . . . 1 .
| . . . . . . . 1

To simplify the model, we’ll assume potential composite classes cannot optionally only use some of their subclasses – it’s all or nothing.

So in fact, we also need a preliminary step (before solving) to decide on the number of potential classes and subclasses from the input data. We can make a guess at an upper limit, eg. by dividing the number of children in each grade by the maximum number allowed in a class, and rounding up.

## Step 2: Choose your constraints

Our constraints need to achieve these things:

1. Ensure all the children in each grade are in a class (or equivalently, a subclass)
2. Classes sizes must be between the minimum and maximum class size – or zero (since not all potential classes need to be used!)
3. Subclass sizes must be above the minimum subclass size, if the class is being used
4. The number of classes must match (or be less than) the number of available teachers

We also need to relate the x and y variables. In a linear program, all the constraints need to be linear in the variables z. This means you can express them as matrices A, so that A.z is greater than, equal to, or less than, another vector c. Here z is just a column vector containing both x and y.

Take each of the above in turn.

### Ensure all the children in each grade are in a class (or equivalently, a subclass)

We have to express this in terms of the number of children in the subclasses x, because the y’s don’t tell us how many children there are in each class.

In our earlier example, the first point involves 3 constraints, one for each grade. Multiply the below matrix with the column vector x (which has 8 rows).

—-8 subclasses—– sums to
| 1 1 . 1 . . . . = number of kindergarteners
3 grades . . 1 . 1 1 1 . = number of first graders
| . . . . . . . 1 = number of 2nd graders

### Classes sizes must be between the minimum and maximum class size, or zero

If we had to use every potential class (ie. all the y variables were known to be 1), then this is just the constraint Cx >= min_class_size, Cx <= max_class_size (where I’m applying the inequality to each row of each side, so these are 6+6=12 constraints in our example.)

However, Cx >= min_class_size should only apply if y is 1. That’s easy to write: Cx – min_class_size.y >= 0. So the constraint matrix has C (N_classes × N_subclasses or 6×8) on the left and -min_class_size (N_classes × N_classes or 6×6, a diagonal matrix of the minimum class sizes) on the right.

I make the same modification for the max class sizes, although it’s not strictly necessary; this has the effect of forcing y >= 0.

### Subclass sizes must be above the minimum subclass size, if the class is being used

If we had to use every potential class (ie. all the y variables were known to be 1), and therefore every subclass, then this is a very simple set of constraints: x >= min_subclass_size. However, we only want this constraint to apply if the corresponding binary value is 1. That corresponding y is actually the corresponding row of C’y (as we discussed earlier).

There’s a classic trick to only apply constraints based on the value of a binary variable, called the “big M” method.

Write x >= min_subclass_size as (x – min_subclass_size) >= 0, and then replace the right hand side with M(C’y – 1), where M is a big number (eg. 10000). Now when C’y is 1, the constraint applies; when it is 0, it does not. Cool!

### The number of classes must match (or be less than) the number of available teachers

The last constraint is easy – the sum of the y’s must be less than or equal the number of teachers; the x’s don’t come into it.

### Do we need to explicitly constrain y to be binary?

Certainly all our variables need to be integers, not floating point. But y already can’t be bigger than 1, because our big M constraint would fail. And it can’t be less than 0 because of the way we wrote our max class size constraint.

So y is already forced to be binary.

### Matrix representation of the constraints

```A:  (integer vars)      (binary vars)                     b:
-- num_subclasses -- -- num_classes --
(--------------------+-----------------)                  (--------------)
(                    |                 )      |           (       |      )
(                    |                 )      |           (       |      )
(--------------------+-----------------)                  (--------------)
(                    |-l1              )      |           (       |      )
( class_size_matrix  |     (diag.)     )  num_classes >=  (       0      )
(                    |             -lnc)      |           (       |      )
(--------------------+-----------------)                  (--------------)
(                    |-u1              )      |           (       |      )
( class_size_matrix  |     (diag.)     )  num_classes <=  (       0      )
(                    |             -unc)      |           (       |      )
(--------------------+-----------------)                  (--------------)
( 1                  |                 )      |           (       |      )
(     1              |                 )      |           (       |      )
(        ...         |     -M * C'     ) num_sub_classes>=(    ls - M    )
(              1     |                 )      |           (       |      )
(                  1 |                 )      |           (       |      )
(--------------------+-----------------)                  (--------------)
(         0          | 1 1   ...   1 1 ) num_classes =,<= ( max_classes  )
(--------------------+-----------------)                  (--------------)

where l is the min_class_size
u is the max_class_size
ls is the min_subclass_size
C' is the transpose of the class_size_matrix
M is a large number
```

## Step 3: Choose the objective function

For a single answer, I just minimise the sum of all the x(i).

Ideally you could get second-best and other feasible solutions from the solver, but I had trouble doing this. As a work-around, I produce different solutions by changing the weights on each subclass. I raise the weight from 1 to 2 for the subclasses in a given grade or composite class, and repeat.

## Step 4: Express the solution in an understandable way

We recast the resulting x and y values so that people can understand them, eg. in our example with 8 subclasses we might get an answer like

```--------------- x ----------- ------- y -------
[20, 12, 10, 0, 0, 25, 22, 21, 1, 1, 0, 1, 1, 1]
```

This is more easily understood as being 5 classes with the following compositions:

```           K  1st 2nd
class #1: 20,  0,  0
class #2: 12, 10,  0
class #3:  0, 25,  0
class #4:  0, 22,  0
class #5:  0,  0, 21
```

## Step 5: Post-process to humanise the answers

The optimisation produces often solutions with very different numbers of students in each class, eg. two kindergarten classes, one with 10 students, and one with 20. Clearly, a better solution is two classes of 15.

As is often the case, the constraints we thought we needed don't quite fully give us the solutions we expected.

We could tackle this by adding more constraints or changing the objective function, but in this case it is simpler to do some "shuffling" after the optimisation, according to some prescribed rules.

Composite classes make this a bit harder. For another example, given the formatted output (note the composite classes are listed at the end):

```[[10,0,0], [20,0,0], [0,25,0], [0,0,10], [25,5,0], [0,5,25]]
```

We could return:

```[[21,0,0], [21,0,0], [0,22,0], [0,0,20], [13,8,0], [0,5,15]]
```

The rules I came up with took some discovering:

1. For each grade, find the average class size of classes with subclasses from that grade. (In the above example, it's 20, which includes 5 grade 1s in the comp class)
2. For each grade, move that grade's students around to make them as close to this average as possible, eg:
• kindergarten: avg = 20;

`[[20,0,0], [20,0,0], [0,25,0], [0,0,10], [15,5,0], [0,5,25]]`

• grade 1: avg = (25+20+30)/3 = 25; in the example, this cannot be done, since the last class is the problem.
• grade 2: avg = (10+30)/2 = 20,

`[[20,0,0], [20,0,0], [0,25,0], [0,0,20], [15,5,0], [0,5,15]]`

3. Repeat until class sizes only change by 1, eg:
• kindergarten: avg = 20, no shuffling required
• grade 1: avg = (25+20+20)/3 = 21.7;

`[[20,0,0], [20,0,0], [0,22,0], [0,0,20], [15,7,0], [0,6,15]]`

• grade 2: avg = (20+21)/2 = 20.5, no change required
• kindergarten: avg = (20+20+22)/3 = 20.7,

`[[21,0,0], [20,0,0], [0,22,0], [0,0,20], [14,7,0], [0,6,15]]`

• etc

## Step 6: Build it!

Build a web app to take the inputs, perform the optimisation and serve the results!

# Harness your Python style to write good Javascript

Python encourages you to write good code. Javascript does not. How can you harness your Python style to write good Javascript?

## Classes

Let’s try to write this Python code in Javascript.

```class Vehicle(object):
def __init__(self, size):
self.size = size
def __str__(self):
return "Vehicle of size {0}".format(self.size)

class Hovercraft(Vehicle):
def __init__(self, *args, **kwargs):
# initialize just like a vehicle
super(Hovercraft, self).__init__(*args, **kwargs)
# add extra custom init if needed
def hover(self, height):
print("Hovering at height {0}".format(height))

# eg. make a new size 8 hovercraft
h = Hovercraft(8)
# eg. hover at height 12
h.hover(12)
# eg. change size and print
h.size = 9
print(h)
```

Not obvious? The problem is that Javascript doesn’t have an out-of-the-box analogue to classes. Somehow we need to adapt objects and functions to the task.

### Use prototypes

Your first thought might be to define new object types by defining object constructors, which are called with the `new` keyword, and adding methods to the object’s prototype chain. These objects look a lot like classes, don’t they? So you might write something like this:

```// Not the best approach - see below
function Vehicle(size) {
this.size = size;
// or use this and arguments pseudo-arguments
}
Vehicle.prototype.toString = function () {
return "Vehicle of size " + this.size;
};
function Hovercraft(size) {
this.size = size;
}
Hovercraft.prototype = new Vehicle();
Hovercraft.prototype.hover = function (height) {
}

// eg. make a new size 8 hovercraft
h = new Hovercraft(8);
// eg. hover at height 12
h.hover(12);
// eg. change size and print
h.size = 9
h.toString()
```

This has a number of problems:

• Vehicle and Hovercraft have the same constructor, but you can’t reuse it.
• It feels strange to define the class’s methods outside the constructor, with the `prototype` lines.
• There’s no room for private variables or methods.

### Use closures

For a better solution, we need to think laterally – and use Javascript’s closures. I was hesitant to make use of these at first, thinking it would only lead to perverse and impenetrable code. But in fact they are a force for good, as we’ll see.

Put simply, a closure is a function which has variables bound to it. You write the closure as a function inside another function. The trick is that the inner function can refer to any of the outer function’s local variables. (Thanks for this succinct summary StackExchange!)

Here’s a nice example of a closure from a talk on functional programming given by Douglas Crockford, JavaScript architect at PayPal and formerly Yahoo (at 57mins). (In fact the previous example is from this talk too.) In this example, a closure is used to produce an object called `singleton`, which has a private variable, a private function, and two methods. You call the two methods as `singleton.firstMethod(a,b)` and `singleton.secondMethod(c)`:

```var singleton = (function () {
var privateVariable;
function privateFunction(x) {
...privateVariable...
}

return {
firstMethod: function (a, b) {
...privateVariable...
},
secondMethod: function (c) {
...privateFunction()...
}
};
}() );
// note the function is called immediately,
// so the var singleton is its returned value
// the surrounding brackets are just to help the reader
```

So let’s adapt that to our problem:

```function vehicle(size) {
var that = {
size: size,
toString: function () {
return "Vehicle of size " + this.size;
},
};
return that;
}
function hovercraft(size) {
var that = vehicle(size);  // inherit from vehicle
that.hover = function(height) {
return that; // optional - allows chaining
};
return that;
}

// eg. make a new size 8 hovercraft
h = hovercraft(8);
// eg. hover at height 12
h.hover(12);
// eg. both at once using chaining
g = hovercraft(8).hover(12);
// eg. change size and print
h.size = 9
h.toString()
```

In the above, `size` is accessible to the world, just as it is in Python. But Javascript also lets us make it private:

```function vehicle(size) {
// can define private variables here
// instead, here we use the fact that parameters are private
return {
toString: function () {
return "Vehicle of size " + size;
},
};
}
```

In general, to make your own constructor function (eg. `vehicle`, `hovercraft`), you follow this recipe from the talk:

1. Make an object
2. Define (private) variables and functions
3. Augment the object with methods (which have access to the privates above)
4. Return the object

This pattern has a name: the module pattern. You’ll find a great writeup of it, and some ways to use it across multiple js files, in this post by Ben Cherry. He also points out the best way to handle dependencies, and to update existing variables.

### Use chaining

In the example above, I have added the extra feature of “chaining”. In Python you have to put each effect on a separate line, eg.:

```h.size = 9
h.hover(18)
print(h)
```

But in Javascript you can potentially chain it all together into one, like so:

```h.size(9).hover(18).toString()
```

I first discovered the joy of chaining while using Mike Bostock’s super-powerful D3 library. He describes how to do it here – in fact, his description of how to write a reusable chart arrives at the same closure-based solution as we have, with the addition of getters and setters as below.

We made `hover` chain, but `size` do it doesn’t yet, as it’s a variable, not a function. Mike solves this by adding a getter/setter function for each public variable, which we could add like this:

```function vehicle(startSize) {
var size = startSize;
var that = {
toString: function () {
return "Vehicle of size " + size;
},
};
// getter/setter functions
that.size = function(_) {
if (!arguments.length) return size;
size = _;
return that; // Q: would 'return this' work?
};
return that;
}
// eg. this works now
h.size(9).hover(18).toString()
```

Then `h.size()` returns (gets) the hovercraft `h`‘s size, and `h.size(9)` sets the size to 9.

Another benefit: the code never refers to `this`. That’s handy, because I find whenever I refactor code into smaller functions I get tripped up by the meaning of `this` changing.

You may also recognize such getters and setters from jQuery, where eg. `\$("body").text()` returns the page body’s text, and `\$("body").text("eels")` sets the body’s text to “eels”.

Still, as nice as chaining is, needing to add 5 lines of boilerplate code for every variable, with 3 references to the variable name that must be changed each time, is the sort of thing we became programmers to avoid.

To solve this, I am starting to put all the variables with getters and setters into a single object, eg. `xt`:

```function vehicle(startSize) {
var ext = {size: startSize};
if (Object.keys) {
var extKeys = Object.keys(ext);
}
function toString() {
return "Vehicle of size " + ext.size;
}
return {
toString: toString,
get: function(name) {
if (!arguments.length) return ext;
return ext[name];
},
set: function(name, val) {
if (typeof extKeys!=="undefined" && extKeys.indexOf) {
if (extKeys.indexOf(name)>=0) {
ext[name] = val;
} else {
}
} else {
// on browsers without Object.keys or indexOf,
// don't check the name is valid
ext[name] = val;
}
return this;
}
}
}
// eg. these work
h.get("size");
h.set("size",9).hover(18).toString();
```

### Don’t make it a global

Finally, you probably don’t want to have a new global called `vehicle`. It’s better to add it to another module, eg. `RT` (which may or may not already exist), as `RT.vehicle`. It might also depend on other modules, eg. the underscore (_) library. To do this, wrap the whole function in another closure!

```if (typeof _ === 'undefined') {
throw new Error('Vehicle requires underscore')
}
(function(RT, _) {
function vehicle(startSize) {
... // copy from above
}

// attach vehicle to RT
if (typeof RT==="undefined") {
RT = {};
}
RT.vehicle = vehicle;
return RT;
}(typeof RT === "undefined" ? {} : RT, _));
// eg. make a new size 9 vehicle
v = RT.vehicle(9);
```

Too crazy?

### In conclusion

Javascript’s closures give you access to some interesting programming patterns. Foremost among them, it lets you implement Python-style classes, with the added bonus of private variables and functions. And this is not just an academic gimmick that risks complicating your code in the real world: it is championed by the people who develop javascript, and it is used by jQuery and D3 among others. It helps you to write good, reusable code.

So – please let me know if you’ve used this pattern before, and whether my comparison to Python’s classes stacks up.

A final thought – perhaps it is wrong to compare Javascript to Python after all. Perhaps it is better compared to LISP!

# 9 Lessons from PyConAU 2014

A summary of what I learned at PyCon AU in Brisbane this year. (Videos of the talks are here.)

## 1. PyCon’s code of conduct

Basically, “Be nice to people. Please.”

I once had a boss who told me he saw his role as maintaining the culture of the group.  At first I thought that seemed a strange goal for someone so senior in the company, but I eventually decided it was enlightened: a place’s culture is key to making it desirable, and making the work sustainable. So I like that PyCon takes the trouble to try to set the tone like this, when it would be so easy for a bunch of programmers to stay focused on the technical.

## 2. Django was made open-source to give back to the community

Ever wondered why a company like Lawrence Journal-World would want to give away its valuable IP as open source? In a “fireside chat” between Simon Willison (Django co-creator) and Andrew Godwin (South author), it was revealed that the owners knew that much of their CMS framework had been built on open source software, and they wanted to give back to the community. It just goes to show, no matter how conservative the organisation you work for, if you believe some of your work should be made open source, make the case for it.

## 3. There are still lots more packages and tools to try out

That lesson’s copied from my post last year on PyCon AU. Strangely this list doesn’t seem to be any shorter than last year – but it is at least a different list.

• Varnish – “if your server’s not fast enough, just add another”.  Apparently a scary scripting language is involved, but it can take your server from handling 50 users to 50,000. Fastly is a commercial service that can set this up for you.
• Solr and elasticsearch are ways to make searches faster; use them with django-haystack.
• Statsd & graphite for performance monitoring.
• Docker.io

Some other stuff -

• mpld3 – convert matplotlib to d3. Wow! I even saw this in action in an ipython notebook.
• you can use a directed graph (eg using networkx) to determine the order of processes in your code

Here are some wider tools for bioinformaticians (if that’s a word), largely from Clare Sloggett’s talk -

• rosalind.info – an educational tool for teaching bioinformatics algorithms in python.
• nectar research cloud – a national cloud for Australian researchers
• biodalliance – a fast, interactive, genome visualization tool that’s easy to embed in web pages and applications (and ipython notebooks!)
• ensembl API – an API for genomics – cool!

And some other sciency packages -

• Natural Language Toolkit NLTK
• Scikit Learn can count words in docs, and separate data into training and testing sets
• febrl – to connect user records together when their data may be incorrectly entered

One standout talk for me was Ryan Kelly’s pypy.js, implementing a compliant and fast python in the browser entirely in javascript. The only downside is it’s 15 Mb to download, but he’s working on it!

And finally, check out this alternative to python: Julia, “a high-level, high-performance dynamic programming language for technical computing”, and Scirra’s Construct 2, a game-making program for kids (Windows only).

## 4. Everyone loves IPython Notebook

I hadn’t thought to embed javascript in notebooks, but you can. You can even use them collaboratively through Google docs using Jupyter‘s colaboratory. You can get a table-of-contents extension too.

## 5. Browser caching doesn’t have to be hard

Remember, your server is not just generating html – it is generating an http response, and that includes some headers like “last modified”, “etag”, and “cache control”. Use them. Django has decorators to make it easy. See Mark Nottingham’s tutorial. (This from a talk by Tom Eastman.)

## 6. Making your own packages is a bit hard

I had not heard of wheels before, but they replace eggs as a “distributable unit of python code” – really just a zip file with some meta-data, possibly including operating-system-dependent binaries. Tools that you’ll want to use include tox (to run tests in lots of different environments); sphinx (to auto-generate your documentation) and then ReadTheDocs to host your docs; check-manifest to make sure your manifest.in file has everything it needs; and bumpversion so you don’t have to change your version number in five different places every time you update the code.

If you want users to install your package with “`pip install python-fire`“, and then import it in Python with “`import fire`“, then you should name your enclosing folder `python_fire`, and inside that you should have another folder named `fire`. Also, you can install this package while you are testing it by `cd`ing to the `python-fire` directory and typing `pip install -e .` (note the final full-stop; the `-e` flag makes it editable).

Once you have added a `LICENSE`, `README`, docs, tests, `MANIFEST.in``setup.py` and optionally a `setup.cfg` (to the `python-fire` directory in the above example) and you have `pip install`ed `setuptools``wheel` and `twine`, you run both

```python setup.py bdist_wheel [--universal]
python setup.py sdist```

The `bdist` version produces a binary distribution that is operating-system-specific, if required the universal flag says it will run on all operating systems in both Python 2 and Python 3). The `sdist` version is a source distribution.

To upload the result to pypi, run

`twine upload dist/*`

(This from a talk by Russell Keith-Magee.)  Incidentally, piprot is a handy tool to check how out-of-date your packages are. Also see the Hitchhiker’s Guide to Packaging.

## 7. Security is never far from our thoughts

This lesson is also copied from last year’s post. If you offer a free service (like Heroku), some people will try to abuse it. Heroku has ways of detecting potentially fraudulent users very quickly, and hopes to open source them soon. And be careful of your APIs which accept data – XML and YAML in particular have scary features which can let people run bad things on your server.

## 8. Database considerations

Some tidbits from Andrew Godwin’s talk (of South fame)…

• Virtual machines are slow at I/O, so don’t put your database on one – put your databases on SSDs. And try not to run other things next to the database.
• Setting default values on a new column takes a long time on a big database. (Postgres can add a NULL field for free, but not MySQL.)
• Schema-less (aka NoSQL) databases make a lot of sense for CMSes.
• If only one field in a table is frequently updated, separate it out into its own table.
• Try to separate read-heavy tables (and databases) from write-heavy ones.
• The more separate you can keep your tables from the start, the easier it will be to refactor (eg. shard) later to improve your database speed.

## 9. Go to the lightning talks

I am constantly amazed at the quality of the 5-minute (strictly enforced) lightning talks. Russell Keith-Magee’s toga provides a way to program native iOS, Mac OS, Windows and linux apps in python (with Android coming). (Keith has also implemented the constraint-based layout engine Cassowary in python, with tests, along the way.) Produce displays of lightning on your screen using the von mises distribution and amazingly quick typing. Run python2 inside python3 with sux (a play on six).  And much much more…

Finally, the two keynotes were very interesting too. One was by Katie Cunningham on making your websites accessible to all, including people with sight or hearing problems, or dyslexia, or colour-blindness, or who have trouble with using the keyboard or the mouse, or may just need more time to make sense of your site. Oddly enough, doing so tends to improve your site for everyone anyway (as Katie said, has anyone ever asked for more flashing effects on the margins of your page?). Examples include captioning videos, being careful with red and green (use vischeck), using aria, reading the standards, and, ideally, having a text-based description of any graphs on the site, like you might describe to a friend over the phone. Thinking of an automated way to do that last one sounds like an interesting challenge…

The other keynote was by James Curran from the University of Sydney on the way in which programming – or better, “computational thinking” – will be taught in schools. Perhaps massaging our egos at a programming conference, he claimed that computational thinking is “the most challenging thing that people do”, as it requires managing a high level of complexity and abstraction. Nonetheless, requiring kindergarteners to learn programming seemed a bit extreme to me – until he explained at that age kids would not be in front of a computer, but rather learning “to be exact”. For example, describing how to make a slice of buttered bread is essentially an algorithm, and it’s easy to miss all the steps required (like opening the cupboard door to get the bread). If you’re interested, some learning resources include MIT’s scratch, alice (using 3D animations), grok learning and the National Computer Science School (NCSS).

All in all, another excellent conference – congratulations to the organisers, and I look forward to next year in Brisbane again.

# Serve datatables with ajax from Django

Datatables is an amazing resource which lets you quickly display lots of data in tables, with sorting, searching and pagination all built in.

The simplest way to use it is to populate the table when you load the page.  Then the sorting, searching and pagination all just happen by themselves.

If you have a lot of data, you can improve page load times by just serving the data you need to, using ajax. On first sight, this is made easy too.  However, be warned: if the server is sending only the data needed, then the server needs to take care of sorting, searching and pagination. You will also need to control the table column sizes more carefully.

There’s quite a lot required to get this right, so I thought I’d share what I’ve learned from doing this in Django.

Start with the following html. This example demonstrates using the render function to insert a link into the table.

```</pre>
<div class="row">
<table class="table table-striped table-bordered" id="example" style="clear: both;">
<tr>
<th>Name</th>
<th>Value</th>
</tr>
</table>
</div>
<pre>
```

and javascript:

```\$(document).ready(function() {
exampleTable = \$('#example').dataTable( {
"aaSorting": [[ 2, "asc" ]],
"aoColumns": [
{ "mData":"name", "sWidth":"150px" },
{ "mData":"supplier", "sWidth":"150px",
"mRender": function (supplier, type, full)  {
return '<a href="'+supplier.slug+'">' + supplier.name + '</a>';
},
},
{ "sType": 'numeric', "sClass": "right", "mData":"price", "sWidth":"70px" },
],
"bServerSide": true,
"sAjaxSource": "{% url 'api' 'MyClass' %}",
"bStateSave" : true, // optional
fnStateSave :function(settings,data){
localStorage.setItem("exampleState", JSON.stringify(data));
},
return JSON.parse(localStorage.getItem("exampleState"));
},
fnInitComplete: function() { // use this if you don't hardcode column widths
}
});
\$('#example').click(function() { // only if you don't hardcode column widths
});
```

Next you need to write an API for the data. I’ve put my api in its own file, `apis.py`, and made it a generic class-based view, so I’ve added to `urls.py`:

```from django.conf.urls import patterns, url
from myapp import views, apis

urlpatterns = patterns('',
...
url(r'^api/v1/(?P<cls_name>[\w-]+)/\$',apis.MyAPI.as_view(),name='api'),
)
```

Then in `apis.py`, I put the following. You could use Django REST framework or TastyPie for a fuller solution, but this is often sufficient. I’ve written it in a way that can work across many classes; just pass the class name in the URL (with the right capitalization). One missing feature here is an ability to sort on multiple columns.

```import sys
import json

from django.http import HttpResponse
from django.views.generic import TemplateView
from django.core.serializers.json import DjangoJSONEncoder

import myapp.models

class JSONResponse(HttpResponse):
"""
Return a JSON serialized HTTP response
"""
def __init__(self, request, data, status=200):
# pass DjangoJSONEncoder to handle Decimal fields
json_data = json.dumps(data, cls=DjangoJSONEncoder)
super(JSONResponse, self).__init__(
content=json_data,
content_type='application/json',
status=status,
)

class JSONViewMixin(object):
"""
Return JSON data. Add to a class-based view.
"""
def json_response(self, data, status=200):
return JSONResponse(self.request, data, status=status)

# API

# define a map from json column name to model field name
# this would be better placed in the model
col_name_map = {'name': 'name',
'supplier': 'supplier__name', # can do foreign key look ups
'price': 'price',
}
class MyAPI(JSONViewMixin, View):
"Return the JSON representation of the objects"
def get(self, request, *args, **kwargs):
class_name = kwargs.get('cls_name')
params = request.GET
# make this api general enough to handle different classes
klass = getattr(sys.modules['myapp.models'], class_name)

# TODO: this only pays attention to the first sorting column
sort_col_num = params.get('iSortCol_0', 0)
# default to value column
sort_col_name = params.get('mDataProp_{0}'.format(sort_col_num), 'value')
search_text = params.get('sSearch', '').lower()
sort_dir = params.get('sSortDir_0', 'asc')
start_num = int(params.get('iDisplayStart', 0))
num = int(params.get('iDisplayLength', 25))
obj_list = klass.objects.all()
sort_dir_prefix = (sort_dir=='desc' and '-' or '')
if sort_col_name in col_name_map:
sort_col = col_name_map[sort_col_name]
obj_list = obj_list.order_by('{0}{1}'.format(sort_dir_prefix, sort_col))

filtered_obj_list = obj_list
if search_text:
filtered_obj_list = obj_list.filter_on_search(search_text)

d = {"iTotalRecords": obj_list.count(),                # num records before applying any filters
"iTotalDisplayRecords": filtered_obj_list.count(), # num records after applying filters
"sEcho":params.get('sEcho',1),                     # unaltered from query
"aaData": [obj.as_dict() for obj in filtered_obj_list[start_num:(start_num+num)]] # the data
}

return self.json_response(d)

```

This API depends on the model for two extra things:

• the object manager needs a `filter_on_search` method, and
• the model needs an `as_dict` method.

The `filter_on_search` method is tricky to get right. You need to search with OR on the different fields of the model, and AND on different words in the search text. Here is an example which subclasses the QuerySet and object Manager classes to allow chaining of methods (along the lines of this StackOverflow answer).

```from django.db import models
from django.db.models import Q
from django.db.models.query import QuerySet

class Supplier(models.Model):
name = models.CharField(max_length=60)
slug = models.SlugField(max_length=200)

class MyClass(models.Model):
name = models.CharField(max_length=60)
supplier = models.ForeignKey(Supplier)
price = models.DecimalField(max_digits=8, decimal_places=2)
objects = MyClassManager()

def as_dict(self):
"""
Create data for datatables ajax call.
"""
return {'name': self.name,
'supplier': {'name': self.supplier.name, 'slug': self.supplier.slug},
'price': self.price,
}

class MyClassMixin(object):
"""
This will be subclassed by both the Object Manager and the QuerySet.
By doing it this way, you can chain these functions, along with filter().
(A simpler approach would define these in MyClassManager(models.Manager),
but won't let you chain them, as the result of each is a QuerySet, not a Manager.)
"""
def q_for_search_word(self, word):
"""
Given a word from the search text, return the Q object which you can filter on,
to show only objects containing this word.
Extend this in subclasses to include class-specific fields, if needed.
"""
return Q(name__icontains=word) | Q(supplier__name__icontains=word)

def q_for_search(self, search):
"""
Given the text from the search box, search on each word in this text.
Return a Q object which you can filter on, to show only those objects with _all_ the words present.
Do not expect to override/extend this in subclasses.
"""
q = Q()
if search:
searches = search.split()
for word in searches:
q = q & self.q_for_search_word(word)
return q

def filter_on_search(self, search):
"""
Return the objects containing the search terms.
Do not expect to override/extend this in subclasses.
"""
return self.filter(self.q_for_search(search))

class MyClassQuerySet(QuerySet, MyClassMixin):
pass

class MyClassManager(models.Manager, MyClassMixin):
def get_query_set(self):
return MyClassQuerySet(self.model, using=self._db)

```

This is a stripped down version of my production code. I haven’t fully tested this stripped down version, so please let me know if you find any problems with it.

Hope it helps!

# Solve this puzzle – @memo and dynamic programming

Here’s the situation.  You have a puzzle with 15 round coloured plates arranged in an equilateral triangle, and 14 coloured pebbles on top of them.  One plate does not have a pebble – it is called the hole.  Your goal is to rearrange the pebbles so that they are on the matching coloured plates, in the minimum number of moves possible.  For each move, you can only move one pebble in a straight line to the hole, possibly leaping over other pebbles on the way.

The question is – can you design an algorithm to calculate, for any starting board, the minimum number of moves to solve it?

In fact this describes the game Panguru, recently produced by Second Nature Games and available as a board game and online.  In Panguru, there are two pebbles and plates of each colour, and one additional legal move down the centreline of the triangle.  If my quick description is too wordy, the online game will give you a much better feel for it; there are rules available too.

## A dynamic programming solution, with memoization

Here’s the approach I came up with, which I implemented in python. This was informed by the excellent book Python Algorithms, Mastering Basic Algorithms in the Python Language, by Magnus Lie Hetland, which I highly recommend.

Think of all the possible moves used to solve the puzzle as a tree. Number the positions 0 to 14.  The root node of the tree is the hole position, and child nodes give the position of the pebble that is being moved. So with the hole starting at the top of the triangle (position 0), one path through the tree might be 0-3-5.  (If you think about it, the last move always tells you where the hole finishes up.)

It is easy to measure how close we are to the solved board: we just count how many pebbles have the same colour as their plates.  When we hit 14, we are done!

In fact we can turn it around.  Let’s assume we have a solution (14 matching pebbles) after m moves. How did we get here?  Clearly, we must have had 13 matching pebbles the move before. (Because you only move one pebble at a time, each move can only increase or decrease the number of matching pebbles by one, or leave the number unchanged.)

And the move before that, we must have had either 12 or 13 matching pebbles. And so on.

This sounds like induction and lends itself to a recursive solution, like this. First define the tree structure via the `Node` class. The core of this is the `__init__` method which keeps track of the node’s parent. I’ve also added `original_pos` to help us later, and two extra methods to display nodes nicely on the command line (`__repr__`) and to make it easier to access the node’s ancestry (`__getitem__`).

```class Node():
def __init__(self, parent, move, num_matching):
self.parent = parent
self.move = move
self.num_matching = num_matching

def original_pos(self, pos):
"""
Return the original position of the pebble now at position pos.
"""
if not self.parent:
return pos
if pos==self.move:
prev_pos = self.parent.move
elif pos==self.parent.move:
prev_pos = self.move
else:
prev_pos = pos
return self.parent.original_pos(prev_pos)

def __getitem__(self, key):
"""
If m is a node which you think of as the last move in a sequence,
m[-3] is the third last move as a Node.
m[-3].move is the third last move as an integer (which position was moved)
"""
if key>=0:
raise IndexError, "You must index moves from the end, e.g. m[-1] for the last move."
if key==-1:
return self
else:
try:
return self.parent[key+1]
except TypeError:
raise IndexError, "Out of range."

def __repr__(self):
return "%s%d" % ((self.parent and ("%s-" % str(self.parent)) or ""),
self.move)
```

Then the puzzle-solving approach could be implemented like this:

```class Puzzle():
def __init__(self, plates, pebbles, allowed_moves):
"""
Set up a puzzle instance for solving.
Args:
plates is a string of 15 chars representing colours
e.g. "WCBRBOGYGPCORPY"
pebbles is the same, with "-" for the hole
e.g. "-PCBRYOGYGBCORP"
allowed_moves is a list s.t.
allowed_moves[i] = list of positions that a pebble at pos i can move to
= list of positions that can move to this position, if it is the hole
e.g. [[1, 2, 3, 4, 5, 6, 9, 10, 12, 14],
[0, 2, 3, 4, 6, 8, 10, 13],
[0, 1, 4, 5, 7, 9, 11, 14],
[0, 1, 4, 5, 6, 7, 10, 12],
[0, 1, 2, 3, 5, 7, 8, 11, 12, 13],
[0, 2, 3, 4, 8, 9, 12, 14],
[0, 1, 3, 7, 8, 9, 10, 11],
[2, 3, 4, 6, 8, 9, 11, 12],
[1, 4, 5, 6, 7, 9, 12, 13],
[0, 2, 5, 6, 7, 8, 13, 14],
[0, 1, 3, 6, 11, 12, 13, 14],
[2, 4, 6, 7, 10, 12, 13, 14],
[0, 3, 4, 5, 7, 8, 10, 11, 13, 14],
[1, 4, 8, 9, 10, 11, 12, 14],
[0, 2, 5, 9, 10, 11, 12, 13]]
"""
self.plates = plates
self.pebbles = pebbles
self.allowed_moves = allowed_moves
self.num_pebbles = len(filter(lambda x: x not in ["-"], self.pebbles))
self.num_matching = sum(plates[i] == pebbles[i] for i in range(len(pebbles)))
hole_pos = self.pebbles.find('-')
self.root = Node(None, hole_pos, self.num_matching)

def matching_nodes(self, turn, num_matching):
"""
Return all the series of moves (as BoardNodes with parents)
that have 'num_matching' matching spots after 'turn' turns.
"""
if turn==0:
if num_matching==self.num_matching:
return [self.root]
else:
return []
result = []
for change in (-1,0,1):
for prev_node in self.matching_nodes(turn-1, num_matching+change):
for move in self.allowed_moves[prev_node.move]:
pebble_colour = self.pebbles[prev_node.original_pos(move)]
# was the moved pebble on a matching plate already?
old_pos_match = (self.plates[move]==pebble_colour)
# does the prev board's hole plate match the moved pebble?
new_pos_match = (self.plates[prev_node.move]==pebble_colour)
# did the move change how many positions were matching,
# by exactly the number we're looking at?
if (old_pos_match-new_pos_match)==change:
result += [Node(prev_node, move, num_matching)]
return result

```

The interesting recursion here is going on in the `matching_nodes` method.  It just implements the idea that the solutions that have, say, 10 matching positions at turn 10, must have had either 9,10 or 11 matching positions at turn 9. It then works back to turn 0, at which point we know how many matching positions there were.

On top of this, we need a further method which finds the right question to ask. It could start by saying – give me all the solutions after 0 moves. If there are none, find all the solutions after 1 move. And keep trying until you find some solutions, e.g.:

```    def optimal_moves(self, stop_at=20):
"""
Return a tuple (fewest possible number of moves, [optimal moves for this board]).
"""
num = 0
result = []
while not result and num<=stop_at:
result = self.matching_nodes(num, self.num_pebbles)
num += 1
return (num, result)

```

The code above will work, but it will be slow and inefficient, because `matching_nodes` will often recurse through territory it has already covered. And that’s where Hetland’s memo decorator comes to the rescue. (This is only a few lines of code which you can find by searching at Google books.) This will cache the results of the decorated method, so that it does not need to be recalculated.  To use it, simply apply `@memo` to the `def matching_nodes` line, like so:

```@memo
def matching_nodes(self, turn, num_matching):```

And that’s it!

You can see this code in action when you ask for a hint in the online game. Each time you press the hint button, the javascript sends off an ajax query which triggers the server to run the above code on the player’s board, and return which moves would come next if the player plays optimally.

In fact, to get around annoying ever-expanding memory problems on the server, I’m running this asynchronously using Celery and Redis (as covered in an earlier blog post), and restarting the celery worker after every request.  But that’s another story…

I hope that has been of interest, and please let me know if you have any comments or improvements.

# Better than jQuery.ajax() – Django with Angular

I have built a website for my games company Second Nature Games using Django.  Django brings lots of benefits like a nice admin panel where my partner can upload new game boards and add new content to the site.

However, to write games on the web you can’t rely too much on a server-side framework like Django. You are going to have to write some javascript as well.  I was keen for the challenge, having used some jQuery before.

In my first attempt, the boards were rendered by Django on the server, and then the javascript on the client-side would manipulate the document object model (DOM) as the game was played. As I described in an earlier post, I used jQuery’s \$.ajax() command to alert the server when the game was finished, and then the javascript would update scores and stats as required.

But this is not easily scalable:  as I added a leaderboard and more stats, more and more things needed to be updated, and I got into a tangle.

The solution is to use a javascript MVC framework like Backbone, Angular or Ember (or Meteor even more so).  I chose to use Angular. There is a great website, To Do MVC, which shows the same To Do application written in many different frameworks, so you can make an informed choice if you’re facing the same decision.

Using one of these frameworks changes how you view the server: the server just sends data and the client renders it. This is known as “data on the wire”.  The effect is to turn the server into an API. (Unfortunately for Django fans, I suspect Django is really overkill for this. But having developed the rest of the site with Django, I stuck with it.)  Meanwhile, whenever the data updates, the javascript framework will update the DOM for you.

Here is a sample template for the Panguru leaderboard, using Angular:

```<div class="info">
<ul>
<span class="name">{{ entry.name }}</span>
<span class="score">{{ entry.score }}</span>
</li>
</ul>
</div>
</div>
```

You can see it’s pretty straightforward and elegant. The code to load it from the API is:

```\$scope.updateLeaderboard = function() {
var that = this;
})
});
}
```

As you may have noticed, Angular and Django both use the double-parentheses notation. You don’t want to have them both try to interpret the same file.

One approach is to use Django’s `{% verbatim %}` tag. I think it’s nicer to separate the Django-rendered templates from the Angular-rendered templates completely, into two separate directories. Django templates stay in their app’s `templates` directory. I put Angular’s html files into the app’s `static/partials` directory. If the server needs to provide data to the javascript, I let Django render it in its template, and pick it up in the static html file. One example is to pass the csrf token. Another example arises because I’m using the staticfiles app, so Django renames all the static files. Unfortunately this includes the locations of all the partial html files and images you need. E.g. in my Django `templates/game_page.html`:

```<html lang="en" ng-app="panguru">
<title>Panguru Online</title>
...
{% addtoblock "css" %}{# if you are using sekizai #}</pre>
<style type="text/css">
</style>
<script type="text/javascript">
csrf_token = '{{ csrf_token }}';
container_url = '{% static "partials/container.html" %}';
...
</script>

<body>
<div panguru-container></div>
</body>
</html>
```

This part I’m not especially proud of, and would love to hear if you have a better solution.

You can then either use a Django API framework like TastyPie to serve the `api/leaderboard/` URL, or you can write one yourself. I did this starting from a JSON response mixin like the one in the Django docs, or this one by Oz Katz, developed for use with Backbone. Then, in Django `views.py`, I put:

```class LeaderboardApi(JSONViewMixin, View):
"Return the JSON representation of the leaderboard"
def get(self, request, *args, **kwargs):
return self.json_response(Score.as_list(limit=20))
```

That just requires a method on your model which returns the object in an appropriate format, e.g.

```class Score(models.Model):
user = models.ForeignKey(User)
score = models.PositiveIntegerField(default=0)
class Meta:
ordering = ['-score']
@classmethod
def as_list(cls):
return [score.as_dict() for score in cls.objects.all()]
def as_dict(self):
return {'name': self.user.username, 'score': self.score }
```

There’s just one more thing, which is an inconsistency in how Django and Angular apply trailing slashes to URLs. Unfortunately if you use Angular’s `\$resource` approach to interacting with the API, it will strip off any final slash, and then Django will redirect the URL to one with a slash, losing any POST parameters along the way. There’s a fair bit of discussion about this online. My approach has been to just turn off Django’s default behavior in `settings.py`:

`APPEND_SLASH = False`

Get ready to thoroughly test your app when you make this change, especially if you are also using Django-CMS, as I am. I uncovered a few places where I had been a bit sloppy in hardwiring URLs, and was inconsistent in my use of slashes in them.

The full result is the logic puzzle game Panguru, which you can play online or buy as a physical boardgame.  Please check it out and let me know what you think!

# Get started with Python on a Mac

Here are the steps I recommend for getting started with python on a Mac (OS X), particularly if you are fairly new to the language.

1. The Mac comes with python installed, and you can run this directly from the terminal application with the command “python”.  However there is a very nice python interpreter called the “ipython notebook” which I strongly recommend using too.  Installation instructions are available here - specifically, I downloaded Enpackage Canopy and typed two commands into the terminal.  You should now be able to get an ipython notebook running in your internet browser with the terminal command:
```cd somewhere/sensible/
ipython notebook --pylab=inline```

The last part of this means that plots get drawn inside the notebook, which is handy. If you forget to add that, you can type into the notebook itself: `%pylab inline`.
Note you use shift-Enter to run all the python commands in a cell.

2. Install pip.  This lets you easily download python packages and will be very handy later.  This is easy – from the terminal type:
`sudo easy_install pip`
3. More advanced users may want to install virtualenv, so that different projects you work on can use different versions of packages. If you’re planning on putting any of your code in production, this is a must. But if you’re just getting started, ignore it for now.
4. Advanced users will also want to install git so you can incrementally save your own versions of code. If you’re just starting out, leave this for later too.

OK, now let’s dive in the deep end by loading some financial data from Quandl, then manipulating it with Pandas and plotting it with matplotlib. You’ll need Pandas and the Quandl package, which you can get by typing into the terminal:

```pip install pandas
pip install Quandl```

Now in your ipython notebook type (I recommend doing each group of statements in its own cell, so that you can run them separately; remember it’s shift-enter to run the statements):

```import numpy
import pandas
import Quandl

assets = ['OFDP.ALUMINIUM_21.3',
'OFDP.COPPER_6.3',
'OFDP.GOLD_2.3',
'OFDP.ZINC_26.3']

data = Quandl.get(assets)

data['2012':].plot()```

This link gives ways you can develop the plot further.

To give an example of how to manipulate the data, you could try:

```# show recent daily returns, normalised by asset volatility (ie. z-scores)
data = data.sort_index() # ensure oldest comes first
returns = (data/data.shift(1))
log_returns = numpy.log(returns)
vols = pandas.ewmstd(log_returns,span=180)  # daily vol, exp weighted moving avg
z = (returns-1)/vols # calc z-scores
z[z>50] = numpy.nan  # remove very bad data
z[z<-50] = numpy.nan
z.tail() # take a look at the most recent z-scores```

That’s a very quick introduction to installing and using python on the Mac. I hope it’s helpful!

Let me know if you have any comments or improvements.

# Asynchronous calls from Django

I have an optimisation I would like to run when the user presses a button on a Django page. For small cases, it is fine to run it synchronously.  However, when it takes more than a second or so, it is not great to have the web server held back by a process of unknown length.

The solution I have settled on is Celery, with Redis as the message broker.  I am using Redis over the alternatives, since it seems to have much lower memory requirements (I find it uses under 2 Mb, vs. 10-30 Mb per Celery process). And the equivalent commands if you want to use redis-queue (which uses about 10 Mb per worker) instead of Celery are given in this post.

There is a bit of a learning curve to get started with this, so I am making a guide for the next person by listing all the steps I have taken to get set up on both my development platform (running MacOS X) and a unix server (hosted by Webfaction).  Along the way I hope to answer questions about security and what the right settings are to put in the `redis.conf` file, the celery config file, and the usual Django `settings.py` file.

## Install Redis

Redis is the message broker. You will need to have this running at all times for Celery’s tasks to be executed.

Installing Redis on Mac OS X is described in this blog. Basically, just download the latest version from redis.io, and in the resulting untarred directory:

```make test
make
sudo mv src/redis-server /usr/bin
sudo mv src/redis-cli /usr/bin
mkdir ~/.redis
touch ~/.redis/redis.conf```

Installing Redis on your server is similar, though you may need to know how to download the code from the command line first (e.g. see this post):

```wget http://redis.googlecode.com/files/redis-2.6.14.tar.gz
tar xzf redis-2.6.14.tar.gz
cd redis-2.6.14
make test
make```

On the production server we don’t need to relocate the `redis-server` or `redis-cli` executables, as we’ll see in the next section.

## Run Redis

To run Redis on your Mac, just type one of:

```redis-server  # if no config required, or:
redis-server ~/Python/redis-2.6.14/redis.conf```

To run it on your Webfaction server, first add a custom app listening on a port, and note the port number you are assigned.

Now we need to daemonize it (see this post from the Webfaction community). In summary, in your redis directory, edit the `redis.conf` file like so (feel free to change the location of the `pid` file):

```daemonize yes
...
...
port xxxxx   # set to the port of the custom app you created```

To test this works, type the commands below. If all is well, the `pid` file will now contain a process id which you can check by providing it to the `ps` command.

```src/redis-server redis.conf
ps xxxxx # use the number in the pid file```

Note – when I did this without assigning the port number of the custom app, I got the following error:

```# Warning: no config file specified, using the default config. In order to specify a config file use src/redis-server /path/to/redis.conf
# Unable to set the max number of files limit to 10032 (Operation not permitted), setting the max clients configuration to 4064.

It turns out someone else was already using port `6379`, the default Redis port.

Now in practice you will want Redis to be managed with `cron`, so that it restarts if there is a problem. Webfaction has some docs on how to do this here; I used:

```crontab -e
# and add this line to the file, changing the path as necessary:
0,10,20,30,40,50 * * * * ~/webapps/redis/redis-2.6.14/src/redis-server ~/webapps/redis/redis-2.6.14/redis.conf```

FYI, for me the running Redis process uses 1.7 Mb (i.e. nothing compared to each celery process, as we’ll see).

## Install Celery

The Celery docs cover this.  Installation is simple, on both development and production machines (except that I install it in the web app’s environment with Webfaction, as explained here):

`pip install django-celery-with-redis`

I have added the following to `settings.py`, replacing the port number for production:

```BROKER_URL = 'redis://localhost:6379/0'
CELERY_RESULT_BACKEND = 'redis://localhost:6379/0'

import djcelery

INSTALLED_APPS = (
...
'djcelery',
...
)
```

And added the suggested lines to the top of `wsgi.py`:

```import djcelery

I found lots more detail here, but I haven’t yet established how much of this is required.

## Run a Celery worker

Now you need to start a Celery worker.

On your development server, you can enter your Django project directory and type:

`python manage.py celery worker --loglevel=info`

On your production server, I started by trying the same command above, to test out whether Celery could find the Redis process and run jobs – and it worked fine.  But in practice, the Celery docs say: “you will want to run the worker in the background as a daemon“.  (Note this link also talks about Celery beat, which “is a scheduler. It kicks off tasks at regular intervals, which are then executed by the worker nodes available in the cluster.” In my case, I do not need this.)

To do this, I copied the CentOS `celeryd` shell script file from the link at the end of the daemonization doc (since the server I am using runs CentOS), and placed it in a new `celerydaemon` directory in my Django project directory, along with the Django celeryd config file (I renamed the config file from `celeryd,` which was confusing as it is the same name as the shell script, to `celery.sysconfig`). I also created a new directory in my home directory called `celery` to hold the `pid` and `log` output files.

One more change is required, at least if you are using Webfaction to host your site: the call to `celery_multi` does not have a preceding python command by default.  While this works in an `ssh` shell, it does not work with `cron` - I believe because the `\$PATH` is not set up the same way in `cron`.  So I explicitly add the `python` command in the front, including the path to python.

The config file looks like this:

```# Names of nodes to start (space-separated)
CELERYD_NODES="myapp-node_1"

# Where to chdir at start. This could be the root of a virtualenv.

# How to call celeryd-multi (for Django)
# note python (incl path) added to front
CELERYD_MULTI="/home/user/bin/python \$CELERYD_CHDIR/manage.py celeryd_multi"

# Extra arguments
#CELERYD_OPTS="--app=my_application.path.to.worker --time-limit=300 --concurrency=8 --loglevel=DEBUG"
CELERYD_OPTS="--time-limit=180 --concurrency=2 --loglevel=DEBUG"
#  If you want to restart the worker after every 3 tasks, can use eg:
#  (I mention it here because I couldn't work out how to use

# Create log/pid dirs, if they don't already exist
CELERY_CREATE_DIRS=1

# %n will be replaced with the nodename

# Workers run as an unprivileged user
CELERYD_USER=celery
CELERYD_GROUP=celery

# Name of the projects settings module.
export DJANGO_SETTINGS_MODULE="myproject.settings"```

In the shell script, I changed the two references to `/var` (`DEFAULT_PID_FILE` and `DEFAULT_LOG_FILE`) and the reference to `/etc` (`CELERY_DEFAULTS`) in the shell script to directories I can write to, e.g.:

```DEFAULT_PID_FILE="/home/username/celery/%n.pid"
...

I found a problem in the CentOS script – it calls `/etc/init.d/functions`, which resets the `\$PATH` variable globally, so that the rest of the script cannot find python any more. I have raised this as an issue, where you can also see my workaround.

To test things out on the production server, you can type (use `sh` rather than `source` here because the script ends with an exit, and you don’t want to be logged out of your `ssh` session each time):

`sh celerydaemon/celeryd start`

and you should see a new `.pid` file in `~/celery` showing the process id of the new worker(s).

Type the following line to stop all the celery processes:

`sh celerydaemon/celeryd stop`

## Restart celery with cron if needed

As with Redis, you can ensure the celery workers are restarted by `cron` if they fail. Unlike with Redis, there are a lot of tricks here for the unwary (i.e. me).

1. Write a script to check if a celery process is running. Webfaction provides an example here, which I have changed the last line of to read:
`sh /home/username/webapps/webappname/projectname/celerydaemon/celeryd restart`
2. This is the script we will ask `cron` to run. Note that I use `restart` here, not `start`; I am doing this because I have found in a real case that if the server dies suddenly, celery continues to think it is still running even when it isn’t, and so `start` does nothing. So add to your crontab (assuming the above script is called `celery_check.sh`):
```crontab -e
1,11,21,31,41,51 * * * * ~/webapps/webappname/projectname/celerydaemon/celery_check.sh```
3. One last thing, pointed out to me in correspondence with Webfaction: the `celeryd` script file implements `restart` with:
``stop && start``

So if `stop` fails for any reason, the script will not restart celery.  For our purposes, we want `start` to occur regardless, so change this line to:

`stop; start;`

Your celery workers should now restart if there is a problem.

## Controlling the number of processes

If you’re like me you are now confused about the difference between a node, a worker, a process and a thread. When I run the `celeryd start` command, it kicks off three processes, one of which has the pid in the node’s `pid` file. This despite my request for one node, and “`--concurrency=2`” in the config file.

When I change the `concurrency` setting to 1, then I get two processes. When I also add another node, I get four processes.

So what I assume is happening is: workers are the same things as nodes, and each worker needs one process for overhead and “`concurrency`” additional processes.

For me, at first I found each celery process required about 30-35Mb (regardless of the number of nodes or concurrency). So three use about 100Mb.  When I looked again a week later, the processes were using only 10 Mb each, even when solving tasks.  I’m not sure what explains the discrepancy.

## Use it

With this much, you can adapt the Celery demo (adding two numbers) to your own site, and it should work.

On my site I use ajax and javascript to regularly poll whether the optimisation is finished. The following files hopefully give the basic idea.

`urls.py`

```# urls.py
from myapp.views import OptView, status_view
...
url(r'^opt/', OptView.as_view(), name="opt"),
url(r'^status/', status_view, name="status"), # for ajax
...
```

`views.py`

```# views.py
import json
from django.views.generic import TemplateView
from django.core.exceptions import SuspiciousOperation
from celery.result import AsyncResult

class OptView(TemplateView):
template_name = 'opt.html'

def get_context_data(self, **kwargs):
"""
Kick off the optimization.
"""
# save the task id so we can query its status via ajax
# if you need to cancel the task, use:
context = super(OptView, self).get_context_data(**kwargs)
return context

def status_view(request):
"""
Called by the opt page via ajax to check if the optimisation is finished.
If it is, return the results in JSON format.
"""
if not request.is_ajax():
raise SuspiciousOperation("No access.")
try:
except KeyError:
ret = {'error':'No optimisation (or you may have disabled cookies).'}
return HttpResponse(json.dumps(ret))
try:
# to do - check if it is really solved, or if it timed out or failed
ret = {'status':'solved'}
# replace this with the relevant part of the result
ret.update({'result':result})
else:
ret = {'status':'waiting'}
except AttributeError:
ret = {'error':'Cannot find an optimisation task.'}
return HttpResponse(json.dumps(ret))
return HttpResponse(json.dumps(ret))
```

`javascript`

```// include this javascript in your template (needs jQuery)
// also include the {% csrf_token %} tag, not nec. in a form
\$(function() {
function handle_error(xhr, textStatus, errorThrown) {
clearInterval(interval_id);
}

function show_status(data) {
var obj = JSON.parse(data);
if (obj.error) {
clearInterval(interval_id);
}
if (obj.status == "waiting"){
// do nothing
}
else if (obj.status == "solved"){
clearInterval(interval_id);
// show the solution
}
else {
clearInterval(interval_id);
}
}

function check_status() {
\$.ajax({
type: "POST",
url: "/status/",
data: {csrfmiddlewaretoken:
document.getElementsByName('csrfmiddlewaretoken')[0].value},
success: show_status,
error: handle_error
});
}

setTimeout(check_status, 0.05);
// check every second
var interval_id = setInterval(check_status, 1000);
});
```

As mentioned in the comments to the code above, if you need to cancel an optimisation, you can use:

`revoke(task_id, terminate=True)`

## Monitoring

You can monitor what’s happening in celery with celery flower, at least on dev:

```pip install flower
celery flower --broker=redis://localhost:PORTNUM/0```

And then go to `localhost:5555` in your web browser.

When you use `djcelery`, you will also find a `djcelery` app in the admin panel, where you can view workers and tasks.  There is a little bit of set up required to populate these tables.  More info about this is provided in the celery docs.

## Security

• http://redis.io/topics/security
• http://docs.celeryproject.org/en/latest/userguide/security.html

I hope that’s helpful – please let me know what you think.

# 9 Lessons from PyConAU 2013

A summary of what I learned at PyCon AU in Hobart in 2013. (Click here for 2014.)

## 1. In 2005, Django helped make it possible for a team of ONE to make a commercial web app

Building web apps with Django is not just possible, it’s fun. I hadn’t realised the key role that Django played, along with Ruby on Rails, in making this happen.

## 2. But in 2013 the goal posts are higher – can it still be done?

And that’s just the back-end. On the front-end you probably want to use javascript, ajax, jQuery, and probably another javascript library, e.g. I have been using kineticjs. But during the talks I learned I will need to consider meteor (heaps of cool stuff, but a starting point is that it drops a lot of the distinction between server and client, so that with very little code, a user can update the database and other users’ pages update to view it automatically), backbone.js (“models with key-value binding and custom events, collections with a rich API of enumerable functions,views with declarative event handling, and connects it all to your existing API over a RESTful JSON interface.”), angular.js (“lets you extend HTML vocabulary for your application”), D3.js (“data driven documents”), node.js, compass and SASS (to make css easier), ember.js (“a framework for creating ambitious web applications”), yeoman (“modern workflows for modern webapps” using Ruby and node.js)…

The keynote of DjangoCon AU by Alex Gaynor explained this in a historical context and sowed the idea in my mind that the time is ripe for a new framework (possibly an enhanced Django) that will make all these things easy as well (roughly speaking). Jacob Kaplan-Moss said to check out the Meteor screencast for what is possible.

## 3. Web security is never far from our thoughts

Jacob gave a great talk on web security.  As I mentioned above, Django takes care of the essential security features – CSRF tokens, SQL injections, password hashing and HTML cross-site scripting. Some immediately useful tips I picked up from Jacob are – always use https everywhere if you have user logins; django-secure makes this easy (“Helping you remember to do the stupid little things to improve your Django site’s security.”); use bcrypt for password hashing; use Django’s forms whenever there is user input, even if it’s not a form; turn off unused protocols (e.g. XML and yaml) in your API; and to emphasise how easy it is for others to intercept your unencrypted data, look up Firesheep.

## 4. Python packages for maths and science are making “big data” much more accessible to everyone

Lots of talks on this. Check out especially the scikit-learn documentation, which is incredibly thorough. But then there’s Pandas, scipy, and scikit-image, and for networks networkx.

For parallelization, the classic algorithm is mapreduce, and mrjob provides an python interface to this.  The easiest way to get started on parallelization is to use IPython.parallel. For an example, check out how to process a million songs in 20 minutes. For queuing jobs and running them in the background, redis-queue has a low barrier to entry. (One caveat – you may need to manually delete .pid files.)

An interesting quote – “Most of the world’s supercomputers are running Monte Carlo simulations.”

## 5. There are lots more packages and tools to try out

To improve my style, I want to check out django-model-utils (especially for “PassThroughManager”); and more generally, django-pipeline (for “CSS and JavaScript concatenation and compression, built-in JavaScript template support, and optional data-URI image and font embedding” – in preference to django-compressor), django-allauth (an “integrated set of Django applications addressing authentication, registration, account management as well as 3rd party (social) account authentication.”), django-taggit (to add tags to your project), Raven (the python client for Sentry, “notifies you when your users experience errors”), django-discover-runner (which will be part of Django 1.6 – it allows “you to specify which tests to run and organize your test code outside the reach of the Django test runner”), and django-sitetree (“introducing site tree, menu and breadcrumbs navigation”).

There’s more… Mock for testing (“allows you to replace parts of your system under test with mock objects and make assertions about how they have been used”), separate selenium tests into tests and page controllersGerrit (for online code reviews), Jenkins (“monitors executions of repeated jobs”), django-formrenderingtools (“customize layout of Django forms in templates, not in Python code.”). There’s a way to resize images in html5 before uploading them. And Fanstatic serves js and css files (e.g. specify you need jQuery through a python statement rather than in the template), though I’m not sure why I would need this yet.

If you need to kill off a process that’s taking too long you can use interrupting cow and django-timelimit.

There’s a way to compile clojure to javascript.  Since I don’t know clojure yet, this is a very speculative project for me, but I like the idea of avoiding javascript.

And if you’re writing tests in iOS, there’s a way to run selenium on the iOS simulator using appium.

## 6. I still have a lot to learn about Python

I won’t embarrass myself by listing all the things I learnt about Python here, though we were encouraged not to be afraid of the CPython source code, and even less so of the PyPy source code (which has the advantage that it is in python!).

I was convinced I should be trying to use Python 3.3 whenever possible, if only to save time later with unicode errors – Python 2.x doesn’t handle these well. Django 1.5 is actually written in Python 3.3, using a package called six to make it work with Python 2.x too.  Incidentally, it also seems the consensus is to use PostgreSQL over MySQL. Though admittedly that doesn’t really fit under this heading.

## 7. The Python community is friendly, humble and welcoming

Good news! This keeps it fun to program in Python as much as anything.

## 8. PyCon was a great conference

Of all the scientific and industry conferences I have been to, this one had the best-presented talks I have seen – and not just the scheduled presenters, but also the lightning (5 minute) talks. They were very engaging and intelligible.  Speakers used their slideshows in inventive ways (e.g. using memegenerator, prezi.com and the odd xkcd cartoon).  And the conference itself was well organised by Chris Neugebauer.

# Install lpsolve for Python

Today I wanted to try out using lpsolve with the python API on my Mac (OS X 10.7) and on my linux server.

Installing it on the Mac is tricky, but the essence is described in this blog post (here I clarify a few things that stumped me for a while, leave out some of the changes mentioned there that I didn’t need to do, and update a path or two):

• `lp_solve_5.5.0.15_source.tar.gz`
• `lp_solve_5.5.0.15_Python_source.tar.gz`

e.g. from sourceforge (I originally got a different version of one of these, and it did something quite different).

• The first will extract to a folder `lp_solve_5.5`. The second will extract to a folder with the same name, however it will only contain an `extra/Python` directory. Copy this extra directory into the first download’s `lp_solve_5.5` folder.
• `cd` into this `lp_solve_5.5` directory.
• `cd lpsolve55`
• `sh ccc.osx` . You will get a lot of warnings, but that’s ok. This will create a `bin/` directory. On my Mac it has a subdirectory `osx64/`, containing `liblpsolve55.a` and `liblpsolve55.dylib`.
• `sudo cp bin/osx64/liblpsolve55.a bin/osx64/liblpsolve55.dylib /usr/local/lib` (this step courtesy of this blog)
• `cd ../extra/Python`
• You now need to edit `setup.py`, as suggested in the blog post above (here I have updated the included directories to reflect current Xcode practice):
```...
LPSOLVE55 = '../../lpsolve55/bin/osx64' # not ux32
...
ext_modules = ...
...
include_dirs = ['../..', '/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.7.sdk/usr/include', '/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.7.sdk/usr/include/malloc'],
...
```

You may also want to update the version number (5.5.0.8) in `setup.py`, which does not match the originally downloaded files (5.5.0.15); I’m not sure which is correct.

• `python setup.py build`
• `python setup.py install` . This writes out …`/lib/python2.7/site-packages/lpsolve55-5.5.0.8-py2.7.egg-info` (in my case into my virtualenv)
• `>>> from lpsolve55 import *` should now work in python.

Once this was working, I also installed `lpsolve` on my linux (CentOS) server. I followed the same steps there, except I executed the straight `ccc` file rather than `ccc.osx` (I changed it to refer to `~/tmp` instead of `/tmp`, since I’m using shared hosting and cannot execute from `/tmp`). On linux there is no need to change the `setup.py` file. And I did not find it necessary to copy the `.a` or `.dylib` files anywhere.

If you’re using WebFaction, you will also want to change the final `install` command (as described here), to:

```python setup.py install --install-lib=\$HOME/webapps/web_app/lib/python2.7 \
--install-scripts=\$HOME/webapps/web_app/bin \
--install-data=\$HOME/webapps/web_app/lib/python2.7```

This also worked.

Hope that helps someone out there – let me know if you have any comments.