Dear Matplotlib users,
I need to plot a (time x distance) array of measurements associated with a color bar. Although time is fixed (i.e 0, 1, 2, etc..), to the number of rows, the distances are not evenly distributed (e.g are fixed to 1.22, 1.53, 1.84, 2,11.), although are always constant.
My question is simple. How can I modify the 'extent ' argument so that I represent the real distance values, and not simply evenly distributed events?
Any suggestion or comment would be greatly appreciated!
A minimal example plot is given bellow:
Thanks a lot in advance
import matplotlib.pyplot as plt
import numpy as np
# fake data
random = np.random.randint(0, 300, size=(55,127)
fig = plt.figure()
ax = fig.add_subplot(111)
myaspect = 100.0
# How can adjust this data to my REAL distances
# (e.g 1.22, 1.53, 1.84, 2,01) ???.
cax = ax.imshow(random, aspect = myaspect, extent=myextent, vmin=0, vmax=300)
colorbar = fig.colorbar(cax, ticks=[0,100,200,300,400])