# Revision history [back]

### How do I enter an imaginary number into a matrix

In a very simple example, I am created a 2 by 2 matrix like this

var('i')
ran40 = matrix(QQ,2,2,[[2*i,-2],[3,4]])
show(ran40)


I need to eventually call show(ran40.rref()) so that I can see the reduced row echelon form of the above matrix. But I get syntax errors in the creation of the matrix above.

How do I simply put the imaginary letter i into a matrix and then it actually be computed on. That is, the imaginart "part" is actually used in the RREF computation.