Prove that every symmetric matrix is diagonalisable by an orthogonal matrix. Solution (i) Suppose \\( Sv=\\lambda v \\) Then \\( <Sx,v>=(Sx)^Tv=x^TS^Tv = x^TSv = <x,Sv> = \\lambda <x,v> \\) So if x is in W, we have <x,v>=0 and <Sx,v>=0 => Sx is in W. So the linear transformation x -> Sx takes vectors in W to W (ii) Let\'s define \\( T : W -> W \\) such that \\( T(w) = Sw \\) First we know that W has dimension (n-1) and let \\( B = \\{b1,...,b_{n-1}\\} \\) be an orthonormal basis for W. \\( Sb_j \\in W : \\exists (a_k)_{1\\leq k \\leq n-1} s.t. Sb_j = \\sum_{k=1}^{n-1} a_k b_k \\) Then since the basis B is orthonormal we have : \\( <b_i,Sb_j> = a_i \\) since we have \\( <b_i,b_j> = \\delta_{ij} \\) \\( T(b_i)=Sb_i=\\sum_{k=1}^{n-1} <b_k,Sb_j> b_i \\) Which proves the first statement as the B-matrix for T is composed of the coefficients of the \\( T(b_i) \\) that are exactly \\( <b_i,Sb_j> \\) at the ij-the entry. To conclude : \\( <bi,Sbj> = <Sbj,bi>=(Sbj)^Tbi=bj^TS^Tbi=bj^TSbi=<bj,Sbi> \\) (iii) Let\'s do induction on the size of S. Suppose n=1. Then \\( \\{b_1=v/<v,v>\\} \\) is an orthonormal basis of eigenvectors for S. Now suppose the statement is true for (n-1). We will prove it for n. Suppose S is of size n. We take \\( b_1 = v/<v,v> \\) one eigenvector of S. Then W as defined in (ii) has size (n-1) and we can apply our induction hypothesis to the symmetric matrix T. So we can find an orthonormal matrix of eigenvectors of T \\( \\{b_2, \\dots,b_{n}\\} \\) . Note that the eigenvectors of T are eigenvectors of S, since \\( T(w)=Sw \\) Then we have found an orthonormal basis of eigenvectors of S : \\( \\{ b_1, \\dots, b_n\\} \\). .