Always try out examples, starting out with the simplest possible examples (it may take some thought as to which examples are the simplest). Does for instance the identity matrix have complex eigenvectors? This is pretty easy to answer, right?

Now for the general case: if $A$ is any real matrix with real eigenvalue $\lambda$, then we have a choice of looking for real eigenvectors or complex eigenvectors. The theorem here is that the $\mathbb{R}$-dimension of the space of real eigenvectors for $\lambda$ is equal to the $\mathbb{C}$-dimension of the space of complex eigenvectors for $\lambda$. It follows that (i) we will always have non-real eigenvectors (this is easy: if $v$ is a real eigenvector, then $iv$ is a non-real eigenvector) and (ii) there will always be a $\mathbb{C}$-basis for the space of complex eigenvectors consisting entirely of real eigenvectors.

As for the proof: the $\lambda$-eigenspace is the kernel of the (linear transformation given by the) matrix $\lambda I_n - A$. By the rank-nullity theorem, the dimension of this kernel is equal to $n$ minus the rank of the matrix. Since the rank of a real matrix doesn't change when we view it as a complex matrix (e.g. the reduced row echelon form is unique so must stay the same upon passage from $\mathbb{R}$ to $\mathbb{C}$), the dimension of the kernel doesn't change either. Moreover, if $v_1,\ldots,v_k$ are a set of real vectors which are linearly independent over $\mathbb{R}$, then they are also linearly independent over $\mathbb{C}$ (to see this, just write out a linear dependence relation over $\mathbb{C}$ and decompose it into real and imaginary parts), so any given $\mathbb{R}$-basis for the eigenspace over $\mathbb{R}$ is also a $\mathbb{C}$-basis for the eigenspace over $\mathbb{C}$.


If $A$ is a symmetric $n\times n$ matrix with real entries, then viewed as an element of $M_n(\mathbb{C})$, its eigenvectors always include vectors with non-real entries: if $v$ is any eigenvector then at least one of $v$ and $iv$ has a non-real entry.

On the other hand, if $v$ is any eigenvector then at least one of $\Re v$ and $\Im v$ (take the real or imaginary parts entrywise) is non-zero and will be an eigenvector of $A$ with the same eigenvalue. So you can always pass to eigenvectors with real entries.


If $x$ is an eigenvector correponding to $\lambda$, then for $\alpha\neq0$, $\alpha x$ is also an eigenvector corresponding to $\lambda$. If $\alpha$ is a complex number, then clearly you have a complex eigenvector. But if $A$ is a real, symmetric matrix ( $A=A^{t}$), then its eigenvalues are real and you can always pick the corresponding eigenvectors with real entries. Indeed, if $v=a+bi$ is an eigenvector with eigenvalue $\lambda$, then $Av=\lambda v$ and $v\neq 0$. So $A(a+ib)=\lambda(a+ib)\Rightarrow Aa=\lambda a$ and $Ab=\lambda b$. Thus, because $v\neq 0$ implies that either $a\neq 0$ or $b\neq 0$, you just have to choose.