Image by Author | Ideogram
Â
NumPy is a Python package data scientists use to perform many data operations. Many other Python packages are built on top of NumPy, so it’s good to know how to use them.
To improve our NumPy usage experience, we need to know where and why our code performs poorly or at least does not meet our expectations. This article will explore various methods for debugging and profiling the NumPy code to see the performance bottlenecks.
Â
NumPy Code Debugging
Â
Before anything else, we need to ensure that our code executes flawlessly. That is why we need to debug our code before we try to perform any code profiling.
Â
1. Using Assert
The easiest way to perform debugging is to use assert to ensure that our code output is as expected. For example, here is how we can perform the code.
import numpy as np
arr = np.array([1, 2, 3])
assert arr.shape == (3,)
Â
2. Using Python Debugger
We can use Python’s built-in debugger to review the code and assess the process.
import pdb
# Add the code when you need to pause in the middle of the execution.
pdb.set_trace()
Â
3. Using Try and Except Block
Using the Try and Except block is also an excellent way to debug and ensure we know what is going wrong.
try:
a = np.array([1, 2, 3])
print(a[5])
except IndexError as e:
print("Caught an Error:", e)
Â
Output:
Caught an Error: index 5 is out of bounds for axis 0 with size 3
Â
NumPy Code Profiling
Â
When we have finished the debugging process, we will profile the NumPy code execution to understand the performance bottleneck in our code.
Â
1. Profiling with Time
The simplest way to profile the performance is to understand the execution time manually. For example, we can use the following code to learn more about it.
import time
start_time = time.time()
np.dot(np.random.rand(1000, 1000), np.random.rand(1000, 1000))
end_time = time.time()
print(f"Execution Time: end_time - start_time seconds")
Â
Output:
Execution Time: 0.03861522674560547 seconds
Â
You can try various code combinations to see if the code execution becomes faster or slower.
Â
2. Profiling with cProfile
We can profile them in more detail with the cProfile package. Let’s see how it works.
import cProfile
def my_numpy_operation():
np.dot(np.random.rand(1000, 1000), np.random.rand(1000, 1000))
cProfile.run('my_numpy_operation()')
Â
Output:
7 function calls in 0.031 seconds
Ordered by: standard name
ncalls tottime percall cumtime percall filename:lineno(function)
1 0.016 0.016 0.031 0.031 :3(my_numpy_operation)
1 0.000 0.000 0.031 0.031 :1()
1 0.000 0.000 0.000 0.000 multiarray.py:741(dot)
1 0.000 0.000 0.031 0.031 built-in method builtins.exec
1 0.000 0.000 0.000 0.000 method 'disable' of '_lsprof.Profiler' objects
2 0.015 0.008 0.015 0.008 method 'rand' of 'numpy.random.mtrand.RandomState' objects
Â
As you can see, cProfile works by dissecting the whole process of our NumPy code and providing details. The information such as how many times the process is called, the time it took, and what method was called.
Â
3. Profiling with line_profiler
We can also use the package line_profiler to get the details for each line in our NumPy code. First, we need to install it.
pip install line_profiler
Â
After installing the package, we would use the following magic commands if you run it in the Jupyter Notebook.
Â
Let’s prepare the NumPy code that we want to profile.
import numpy as np
def matrix_multiplication(n):
a = np.random.rand(n, n)
b = np.random.rand(n, n)
result = np.dot(a, b)
return result
Â
Then, use the following magic command to see how the code works under the hood.
%lprun -f matrix_multiplication matrix_multiplication(500)
Â
Output:
Timer unit: 1e-09 s
Total time: 0.0069203 s
File:
Function: matrix_multiplication at line 3
Line # Hits Time Per Hit % Time Line Contents
==============================================================
3 def matrix_multiplication(n):
4 1 2165161.0 2e+06 31.3 a = np.random.rand(n, n)
5 1 1824265.0 2e+06 26.4 b = np.random.rand(n, n)
6 1 2930093.0 3e+06 42.3 result = np.dot(a, b)
7 1 780.0 780.0 0.0 return result
Â
The result will be similar to the above report. You can see the performance details and where the performance bottlenecks were.
Â
4. Profiling with memory_profiling
Lastly, we can also see the performance bottleneck from the memory standpoint. To do that, we can use the memory_profiling package.
pip install memory_profiler
Â
We will use the magic command below to initiate memory profiling in Jupyter Notebook.
%load_ext memory_profiler
Â
Then, let’s prepare the NumPy code we want to execute and use the magic command below to get the memory information.
def create_large_array():
a = [i for i in range(10**6)]
return sum(a)
%memit create_large_array()
Â
Output:
peak memory: 5793.52 MiB, increment: 0.01 MiB
Â
From the code above, we get the total memory information and the incrementation memory usage for the whole process. This information is vital, especially if your memory is limited.
Â
Conclusion
Â
Debugging and Profiling NumPy Code is important to understanding where our performance bottlenecks were. In this article, we explore various methods to pinpoint the problems. We can gain performance information From the simple assert Python command to a library such as memory_profiling.
I hope this has helped!
Â
Â
Cornellius Yudha Wijaya is a data science assistant manager and data writer. While working full-time at Allianz Indonesia, he loves to share Python and data tips via social media and writing media. Cornellius writes on a variety of AI and machine learning topics.