--> Skip to main content



Python comparison - JIT, CUDA and DASK

Python has become the language of choice for data scientists and data analysts. It is easy to use with a lot analytical support libraries. Python programs aren't particularly fast. This has driven people to create a set of tools that help Python programs scale-up and scale-out. 

We can compare approaches with a simple program that I adapted to JIT, GPU and distributed computing.
Demonstration CaveatDistributed computing works best when there are problems with lots of I/O that can be spread across workers.  This program has no I/O and a high data transfer to computation ratio.
GPUs have a high cost for data load and unload data with GPU memory. This means they work best when there is a significantly higher computation to data transfer ratio than this program has. DemonstrationThe sample Python program creates 10,000,00 3 variable rows that it then uses as inputs for 10,000,000 iterations of log(x)*log(y)*log(Z).  The results are returned in a 10,000,000 long result array.  The times fo…

Latest Posts

Installing CUDA Python - Numba - Ubuntu 18.04 LTS

Installing CUDA 10.2 and RTX NVidia drivers on Ubuntu 18.04 LTS

Docker on a Chromebook on Crostini - Neverware CloudReady is ready

Sales Engineer Guide: Understanding the vendor elimination phase

Using authenticator applications with VIP 2FA protected sites

Protect messaging and streaming data in the cloud with "data key" encryption

Are you "on board or out" or "disagree and commit"

Azure Stack ASDK Hosts and Networks as 2020 February

Recognizing Violent Agreement

Azure Stack ASDK local installation workflow