How do I enter an imaginary number into a matrix
In a very simple example, I am created a 2 by 2 matrix like this
var('i') ran40 = matrix(QQ,2,2,[[2*i,-2],[3,4]]) show(ran40)
I need to eventually call
show(ran40.rref()) so that I can see the reduced row echelon form of the above matrix. But I get syntax errors in the creation of the matrix above.
How do I simply put the imaginary letter i into a matrix and then it actually be computed on. That is, the imaginart "part" is actually used in the RREF computation.