I am about to write some computationally-intensive Python code that'll almost certainly spend most of its time inside
's linear algebra functions.
The problem at hand is embarrassingly parallel
. Long story short, the easiest way for me to take advantage of that would be by using multiple threads. The main barrier is almost certainly going to be the Global Interpreter Lock (GIL).
To help design this, it would be useful to have a mental model for which
operations can be expected to release the GIL for their duration. To this end, I'd appreciate any rules of thumb, dos and don'ts, pointers etc.
In case it matters, I'm using 64-bit Python 2.7.1 on Linux, with
0.9.0rc2, built with Intel MKL 10.3.1.