This is an old revision of the document!
I’ve been working on presenting Matrices in a way that really extends and puts in context some ideas that are rather big in the HSC. Here I am trying to show a few of those big ideas from quite a different perspective, and I hope this really helps your thinking about them — and your ability to apply them in problem solving!! This is growing by the hour, it has been for a few weeks, after a summer going seriously down the quantum rabbit hole and landing on matrices. They really really should be part of extension HSC, they are deeply relevant for a lot of modern physics and now even more seriously important in computing and making sense of a whole range of very current discussions.
Read these paragraphs slowly and think about these perspectives on a lot of the ideas you are studying in extension HSC. Think specially about how these areas are tied together. That is a key to being able to understand and solve problems that involve the methods you have memorised and practiced in class. Solving those problems is the goal of the course, a huge part of the exam and the maths as it lives beyond the classroom.
I want to add a few words to your mathematics vocabulary. Those headings and bold-face words here that are new, or that are more broadly defined here. They are about ideas you have tackled at school but have not named. Naming something certainly helps thinking about it. The topic Ext2 is about to hit in the first 5 weeks of year 12 — Mathematical Proof — is rather close to the Pure Maths perspective that underlies the whole of this discussion. The kind of insights I end up talking of in Isomorphism below are vital for the kinds of creative proofs that form the toughest questions in Ext2.
But that understanding, putting it all together as one big language is also required, perhaps less creatively and in more familiar contexts, in Ext1 problems and in the tough questions in Advanced.
These are like tables of numbers (or some other thing we can add and multiply in the usual fashion), but most commonly we talk of Real Numbers. When we move beyond real numbers the ideas can get very interesting, and very useful. They are similar to vectors, which we can write as a list or column, also often but not always Real Numbers. We can add matrices and multiply them by a scalar in the same way we can with vectors written this way. This multiplying by scalars and adding is called Linear Combination and is a very important idea in many contexts. We have other operations we can do with them — which makes a Matrix Algebra.
We can write them like this: \(\ \left[\matrix{1 & 4 & x \cr 0 & \frac12 & 2}\right],\ \) which is a \(\ 2{\times}3\ \) matrix.
Most data sets we look at through to statistics in year 12 can be thought of as a Matrix. A Vector can also be thought of as a single row or single column Matrix. Each Real Function can be thought of as an infinite-dimensional vector, but that is way beyond high school. Matrices are much more powerful than that! Square matrices can represent and perform the transformations we look at in Cartesian Space and in functions — Linear ones like the rotations, reflections, stretches and skews which preserve `0` and the more general Affine ones like translations. And the Complex Numbers (including the Reals) can be thought of as particular 2\(\times\)2 matrices — which match the linear transformations in 2D space in the very intriguing way I will work towards in this discussion. But there is much to fill in first.
This essentially means a value we can use to multiply something, or to measure a single aspect of something. The magnitude of a vector is a scalar, so are the elements of matrices. Multiplying by a scalar often relates to stretching or shrinking something — changing its scale. It certainly does when we are talking of measurements and Real Numbers, or Real Vectors. Sometimes we think of matrices as a list of vectors, so the rows of the table are vectors, or maybe the columns are vectors. Each vector in the list must be the same type – all the same dimension (have the same length list of scalars) and use the same kind of scalar. A scalar always has a Unit, multiplying by that value leaves the thing unchanged, it neither stretches nor shrinks. We usually use `1` as its symbol and call it the Identity.
We meet them in high school. At first they simply have a magnitude and a direction. We can use an arrow (its length is the magnitude) in a space to show one. The magnitude is a scalar. We can multiply vectors by these scalars. This stretches or shrinks them while preserving direction, reversing it if we multiply by a negative. We can also add vectors. The negative of a vector is just ‘in the the opposite direction’. Being negatives of each other means they add together to make zero and that adding one of them undoes the addition of the other. In this addition sense vectors are very like numbers. When the magnitude is `1` then we have some sense of a basic direction, we call that a Unit Vector. This becomes extremely important and powerful as we bring together geometry, number and algebra. We can move these arrows around in geometrical space, it does not matter — the magnitude and direction stay the same, it is the same vector everywhere. The arrow represents a value, familiar ideas like velocity have this kind of value and velocity is the same no matter where you are (at least until you get into modern physics and things get much more complicated!! Hmm — Matrices).
To add vectors we can put the base of the first arrow on a place we call `0` then put the next arrow on the tip of the first — and the sum is an arrow from `0` to the tip of the second. This is rather like jumping around on a number line, but we are not confined to a line. We can also think differently about our vectors. Put the arrow on `0` and put in an axis through that `0` for each dimension using our scalars as the scale — then look at where the tip of the arrow is in this Cartesian Space. We describe position with a list of numbers, one for each dimension. This description is an example of what we called a vector above, when we were talking about matrices.
A list of scalars or a magnitude and direction are both descriptions of a vector. We can do our vector operations in the list of scalars form, they are equivalent. They must be the same type, in the same space — that is have the same dimension with the same kind of scalar. We add each corresponding pair from their lists of scalars individually. We can multiply a vector by a scalar — multiply each scalar in the list individually. These are equivalent to the addition and scalar multiplication above.
There are some other vector operations. One that will be important in this discussion is the Dot Product, also called Scalar Product because the result is a scalar not a vector. This is between two vectors of the same type — the same dimension with the same scalars. Multiply each pair of corresponding elements individually then add these up. We often use bold face for vectors and the same letter with an index for its scalar components, so for `n`-dimensional vectors: \(\ {\bf v\bullet w}=\sum_{i=1}^n v_i w_i.\ \) Or if we are thinking magnitude and direction, with `|\ |` indicating ‘magnitude of’ and \(\ \theta\ \) the angle between them: \(\ {\bf v\bullet w}=|{\bf v}| |{\bf w}|\cos\theta\ \) is equivalent.
This makes for some useful calculations. In dynamics for example \(\texttt{force}\bullet\texttt{displacement}\) is work done, which is energy transfered. Considering movement along a curve with a varying force we can add up the infinitesimal contributions, \(\ \int{\bf F}\bullet{\rm d}{\bf s}\ \) but calculus with vectors or multiple variables is just beyond where we get at school so we often stay with constant \(\bf F\) and straight tracks. Dynamics questions are in our maths course because they provide some easy to visualise examples of mathematical models applied to the world. If the vectors are perpendicular dot product will be `0` since \(\ \cos\frac\pi2=0,\ \) if they are parallel it will be the product of their magnitudes. This is closely related to the idea of independent components. If everything is all on the same straight line our model is 1D, only the magnitudes count and direction is simply positive or negative. Work done (potential energy added) on a load lifting it in an elevator or with a crane is an example of this. The two quantities are parallel, we can treat them as scalars.
If one of the vectors in a Dot Product is a Unit Vector we get the scalar component of the other vector in that direction. But more usually we have a pair of vectors and we want to know the component of \(\bf v\) in the direction of \(\bf w\). In our maths course this often comes up in projectile motion. Here we have a force (hence acceleration) aways in the same direction so it only changes that component of the velocity, we consider the components of the velocity independently and put them back together in the end. This kind of analysis can be applied in many different situations. Divide a vector by its magnitude and it becomes a Unit Vector, so the component is \(\frac1{|{\bf w}|}{\bf w\bullet v}.\) With this magnitude the perpendicular component is the other leg of a right angle triangle and pythagoras rule gives that easily. Notice that Dot Product is Associative and Commutative along with scalar multiplication and scalar division is multiplication by reciprocal as usual. We can rearrange our vectors like we do variables in an algebraic Term. All this is especially useful if we have 3D (or higher!) vectors in cartesian form, Pythagoras gives us \(|{\bf w}|\) easily but the angle is a bit messy geometrically. Dot product finds angles between 3D vectors quickly, putting the magnitude-direction form and cartesian form together we get \[\sum_{i=1}^n v_i w_i=|{\bf v}| |{\bf w}|\cos\theta.\] Make that into a formula if you like but these should not be memorised formula for particular questions!
Be familiar with the two forms of vectors, the dot product in each form, its interpretation for component, angle or ‘are they perpendicular?’ and pythagoras for magnitudes. Be practiced noticing complementary or supplementary angles, right, isosceles and equilateral triangles, angle sums and all those geometrical features. Be familiar with noticing when these direct calculations or relationships could be part of the result you are seeking, the information you are given or some intermediate step. The problems in the exam are devised to test your imagination and understanding of the relationship between models and the calculations you can use in them.
Geometrically using right triangles and trig these are messy, with Dot Product easy to write down what needs to go in the calculator. It may be quicker just to use the coordinates, square roots and such in a big fraction but the school calculator models do have vector mode.
Unless you are doing several calculations with the same few vectors fussing around with those menus and the absence of a history in that mode probably outweighs any advantage you get with vector operations. Polar–Rectangular conversion is available in the standard mode, for 2D vectors.
We could certainly draw a diagram and find right angle triangles, use our trig and find the appropriate intermediate values then come up with the answer. There are many ways to tackle the problem. It is a very well known type of system and there are a lot of standard formula that can give a string of intermediate results and then a final outcome. But maybe we can use our vector knowledge and calculus to express the problem in a way that could generalise to much more complicated systems. In high school we look at highly simplified problems but we take away the usual computer maths support and expect them to be analysed using the tools we are teaching. Walk before you run! But really build the language from ground up. The HSC questions are very carefully structured to test understanding and flexibility when using the tools.
Can you frame an equation or two that answers it more directly. This is about understanding the systems we are describing and how the mathematical language we use in that description works. There are plenty of maths packages where you could type in the equation we might make to model the question and the answer would just be be returned. We describe systems in this way and being able to read the equation and ‘see’ the system is fluency with this language. Problems are often not so neat and more purely vector solutions can be needed. It is the component of the force along the track that is going to accelerate the train.
A deep understanding of the idea of dot product — its different meanings and methods, this equivalence and its relationship to spaces, functions and equations is needed for several extension maths problem types that always turn up. It has a wide range of meanings beyond what we see in school, including the Matrix ideas behind this discussion.
This idea we have also developed in high school. In geometry we think about lines, points, directions, angles, distances and surfaces. We think (mostly) about 2 or 3 dimensions and areas, volumes and such. Then we think about Cartesian Space, describing positions using axes as we did above. We get into describing shapes with functions and finding directions, slopes, areas and volumes with geometrical techniques and with calculus. We start to think about how vectors fit in with these ideas. We also think about transformations of shapes in this space: rotations, reflections, stretching in one or all directions and sliding in any direction. There is a great deal more to this idea of transformation than we have met so far!!
An important idea is Euclidean Space, which essentially means an ideal of flatness in the sense that parallel lines and our common idea of distance between two points make the usual geometrical sense. In Euclidean Space our familiar geometry applies, especially ideas like Pythagoras Rule about sides of right angle triangles and the angles of a triangle adding up to a straight angle. In reality our universe is not so simple, the space we exist in is not flat. Classical geometry is only an approximation. A very good one if size and temperature are in the range familiar to our senses. Time is not what it seems and the whole thing is springy so that we get ‘gravity waves’ travelling through the fabric of space itself. When things get big enough, small enough, energetic enough or cold enough we need to drastically rework our warm and human-scale intuition. Matrices are an important way to enable that.
We do not look much at non-euclidean spaces in school — the main one is looking at the surface of the earth, and in particular different maps of the world, different projections of the globe onto a flat piece of paper. Scales and straight lines get messed up, there needs to be stitching together, nearby places can be on the other side of cuts and a long way apart on the paper. Parallel lines do not make sense to a pilot of a jetliner. Pythagoras rule fails on triangles, the angle sum rule fails. We have a very different geometry. And then we see it embedded in 3-space, causing day and night, winter and summer, midnight sun in the arctic summer, timezones and the dateline.
Understanding spaces which are not flat we make a great deal of use of the idea of a Smooth space — where when we zoom in close enough it looks very much like Euclidean Space. There are no kinks, gaps, sudden jumps or hard corners. This is greatly helped by the ideas of Infinitesimal Calculus that we get a start on in high school. We can think of the typical functions we look at in calculus as describing a one dimensional space that mostly is not flat. Drawing a graph could be considered visualising embedding this space in the two dimensional Cartesian Plane. One way we developed calculus was zooming in closer and closer using the mathematical ideas of Limits and small changes — deltas like `Delta t` becoming infinitesimals like \({\rm d}t\). We found straight line or rectangle limits to describe the slopes or areas related to a curvy space described by a function and made new functions that gives a meaningful and useful answers to the questions “What is the slope of this space?”, “What is the curvature of this space?” and a way of going backwards on this that turns out to answer very different question about changes or cumulative changes in the position of this space as embedded in the larger space. Also questions like “How much area is there under a part of this space, in the higher dimensional euclidean space we embedded it in?”. Another way to say these are “How fast are things changing at a particular time?” (Derivative of a function if time is the function input) and “What is the total change during an interval of time?” (Definite Integral, of a function describing rate of change with time as its input). These ideas were first developed to describe movement and forces in Euclidean Space with an eternally ticking universal background time. They proved incredibly useful in many kinds of analysis about things that change. This completely new understanding of the world enabled the industrial revolution, and much of what we think of as a comfortable life.
We also talk about the set of all the vectors of a particular type and dimension along with their scalars. We call this collection a Vector Space. This set is in many ways like our various sets of numbers of a particular kind, its structure and behavior has much in common with the structure and behavior of these numbers. It also has much in common with the structure of Euclidian Space considered as a set of points.
The mathematical power here is in bringing the number properties addition, subtraction and difference together with the vector properties of scalar multiplication and dot product and considering them as working with the points in geometric space. That is quite a mouthful of words. That is why we take a few years and several big school topics working with different aspects of this one sentence — finally hitting Vectors and (for Ext2) Complex Numbers.
We define the structure Vector Space by making addition follow the rules of number addition. We call any set with such an operation a commutative Group, an Abelian Group. Then we define scalar multiplication to extend that by following a hybrid version of the rules of number multiplication, between one of each type of value. We call integers (or any other set with \(\ +\!\times\!=\ \) of its kind) a Ring. With a Vector Space we have two kinds of value (vector and scalar). The scalar must be more than just a Ring, it must support full division like Rational Numbers do. The vectors do not have this number kind of multiplication between them, Dot Product gives a scalar not a vector. Vectors gain a direction property which can also be interpreted as the set being `n`-dimensional. That means they lose the sense of order that many types of number have. With Real Numbers \(\ne\) implies either \(\lt\) or \(\gt\). This is not true for vectors. Order is essentially a one-dimensional quality.
Square matrices do have multiplication between them, and thus they have a Unit, an Identity and much else, but that multiplication is non-commutative, and only works between particular combinations of non-square matrices, More on that here.
By dimension above we essentially mean the number of independent directions our set of vectors has. We can start with a carefully chosen set of vectors which we call a Basis and generate all the vectors by the Linear Combinations of them. Put another way you can make every vector by a scalar multiplication of each basis vector then adding the results together. A basis must be one of the possible smallest sized sets of the vectors that can do this. In which case there is only one way to do it for each vector. When we were considering vectors and put them into a Cartesian Space above we effectively chose a basis. We chose the Unit Vectors along each axis. The dimension of a Vector Space is the number of vectors in its basis, the dimension of a Cartesian Space is the number of axes. The values of the components of a vector depend on the basis chosen, those operations using the scalar components are assuming we are using the same basis consistently. Suppose we have chosen a basis of Unit Vectors \(\ \{{\bf a, b, c}\}\ \) that means every vector in the space is uniquely represented by three scalars as \(\ [x\ y\ z]\ \) such that \(\ {\bf v}=x{\bf a}+y{\bf b}+z{\bf c}.\ \) This is a Linear Combination of the basis and \(\ x, y, z\ \) are the components of the vector, or the coordinates of its tip if it is placed on point \(\ (0,0,0)\ \) in the corresponding Cartesian Space (the axes use the same scalar and are in the same direction as the basis vectors). The idea of Linear Independence is important here, a vector is independent of a group of vectors if there is no Linear Combination of them that makes that vector.
When studying vectors seriously we prove that all those equivalences, uniquenesses, properties and the idea of a basis are valid and consistent. This does give us a much deeper insight into how it all fits together, but we only get started with this idea of Proof in high school, that kind of detailed analysis of mathematical structures must wait. But we must get a general intuition into the nature of the mathematical structures we use, and the relationships between them. This is the way we use them to solve problems, and the way we know how to tackle the HSC questions. Those questions are carefully devised to test if you have developed that intuitive understanding!
We get as far as them in high school, if you go to our top level of HSC maths. There are lots of simple algebraic equations that have no solutions in Real Numbers. Consider \(x^2+1=0.\) Nothing fancy or tricky, but no solution for `x`. But simply invent one more number, the ‘positive’ solution to this equation, give it the name \(\rm i\) and then include all the numbers you get by adding and multiplying it with the rest of the Real Numbers. Essentially create every number \(\ a+b\rm i\ \) with `a, b` real numbers and \(\ \rm i^2=-1.\ \) Now all our polynomials, and much else, always have solutions. And it is entirely consistent, these numbers obey all the usual number rules and we can easily use them in number algebra. They can multiply things like matrices or vectors easily enough if you allow them as the numbers in the table, list or magnitude. They can be the coordinates of a space. It is hard to imagine them as a size, there are very real problems here, and they do not have a place on our familiar number line. When they are scalars we need to do some serious rethinking.
Complex quantities, Complex Vectors, Complex Cartesian Space, Complex Fields and Complex Matrices have very natural applications to the physical world. It turns out nature is quite happy with them, there are quantities out there that make a lot of sense as Complex Numbers. The geometrical visualisations of these concepts is challenging, we are very accustomed to thinking within a Real Euclidian Space set in a flow of independent continuous time. The world is not so simple once we get big, small, energetic or cold enough. To understand it, work with it, predict it or make and use modern technologies we have to really shake up that habit. We need these mathematical structures as the language to do it. There are important operations with them that do result in Real Numbers, and these can give various senses of size but not really order because many share the same size. Again we have some serious new intuition building required if we are to imagine, predict and describe nature as we find it when we look closely, and that intuition will have to come from serious pure mathematical thinking. The science of the 20th century was a spectacular example of doing exactly that, and the 21st is looking wilder still.
Considered as numbers they have a two dimensional representation, we have a Complex Number Plane that is like the Real Number Line. We put \(\rm i\) directly above `0` on the number line. Same distance up as 1 is to the right. The rest all works, \(a+b\rm i\) is at the coordinates \( (a,b) \) in this plane. But keep going. That means they have a ‘distance’ from `0` and a direction, the angle measured anticlockwise from `1` to the point in this Complex Number Plane. Call these `r` and `theta`. Then we can see such a point is \(\ r(\cos\theta+{\rm i}\sin\theta)\ \) by considering the geometry of this number plane. Then considering Taylor Series or some other reasoning we can establish \(\ {\rm e^{i\theta} }=\cos\theta+{\rm i}\sin\theta.\ \) In this respect they are like 2D Vectors and very much more — you can multiply and divide them, take their log, and a great deal more that you cannot do with Vectors. They behave like both vectors and numbers, adding is the same as adding vectors and multiplying by a real number is the same as with Real Vectors. In high school we find that we can raise a number to a complex power quite sensibly and consistently. And with this there are some very powerful connections made between parts of maths that till now had seemed quite distinct. Algebra, Geometry and Trigonometry come together in Complex Numbers.
Some consider the equation that summarises this idea, \(\ \rm e^{i\pi}+1=0,\ \) the most remarkable in mathematics. It connects each of the five basic numbers and three basic number operations. But more on it soon, it has much to do with matrices and physics — and quantum computing.
Complex numbers are only in extension 2 high school here, it ties together a lot of the ideas you have been developing over the earlier topics. There is at least one fundamental matrix algebra idea that I very much should share with you that is part of this. The relationship between matrices and complex numbers is something I had not really explored before, and is very elegant as well as deeply core to the way the world works on the basic entity level, electrons, photons, atoms and such. But a representation of complex numbers (and hence also real numbers) within the 2\(\times\)2 square real matrices is intriguing. I’ll need to cover a bit about such matrices first though!
Real Vectors, with Real Number magnitudes and finite positive integer dimension usually 2 or 3, thought of with direction in geometric space are the kind we look at in high school — like displacements, forces, velocities or accelerations. But beyond that school topic we can have Complex Vectors, where the magnitude is a complex value. Thinking about arrows in geometric space breaks down here, but Nature does not care!
That is just the start. We can consider polynomials as vectors (the coefficients are the list of scalars). We can even consider functions that give a scalar for each element in some domain as vectors, we can add them and we can multiply them by a scalar and it works the same way as it does with vectors — thinking of the dimension of this one, it is big! Each such function an infinite dimensional vector. And functions do behave like vectors in this way. It can be a very useful for generalising and understanding functions as abstract entities.
In school, Extension 2 only, we meet the Cross Product. It is a 3D-only special case of the Exterior Product. The magnitude of the Cross Product is the area of the parallelogram formed by the two vectors, \(\ |{\bf v\times w}|=|{\bf v}| |{\bf w}|\sin\theta.\ \) It is complementary to Dot Product when considered geometrically, it uses sine rather than cosine. Dot Product is zero when perpendicular while Cross Product is zero when parallel. But Cross Product is a vector, its direction is perpendicular to each operand and up or down according to the Right Hand Rule. It is an area and a direction, an oriented area. The product of two directed lengths making a directed area fits with our geometrical ideas of area. The polar form comes directly from that. This Right Hand Rule business means it is not cleanly a well behaved vector. Sometimes we call it a Pseudovector, more generally we speak of a Bivector. Equivalently, working in cartesian form we find that \[\ \left[\matrix{v_1\cr v_2\cr v_3}\right]\times\left[\matrix{w_1\cr w_2\cr w_3}\right]= \left[\matrix{v_2 w_3 - v_3 w_2\cr v_1 w_3 - v_3 w_1\cr v_1 w_2 - v_2 w_1}\right]\] In higher dimensions we can have an `n`-vector, for example a 3-vector has a volume as its magnitude. But this is again going far beyond school. The two different representations of Cross Product are the power of this idea, we establish that the two forms are equivalent in general, as abstract mathematical models — then back and forth between magnitude-direction and cartesian forms will give insight into questions and help solve problems!
How does that rectangular form work for directed areas? It seems initially quite a long way from the geometry! Consider three perpendicular unit vectors, \(\{\bf i, j, k\}\) as the basis. Then consider their Cross Products geometrically using the parallelograms formed and the Right Hand Rule. \[\bf i\times i=j\times j=k\times k=0\] they are parallel to themselves so the parallelogram area is zero. There are two possible orientations for our basis, say we have \(\bf i\) east, \(\bf j\) north and \(\bf k\) upwards then \[{\bf i\times j}={\bf k}=-({\bf j\times i}),\quad{\bf j\times k}={\bf i}=-({\bf k\times j}),\quad{\bf k\times i}={\bf j}=-({\bf i\times k}).\] The parallelograms are all unit squares and the Right Hand Rule gives us the positive or negatives. Considering geometry and areas again we can see that Cross Product has Distributivity and Linearity as an operation. Scalar multiply one vector and you scalar multiply the area. You can add two vectors then find the area, or you can find two areas then add them, geometrically the same outcome. We’d need to get a bit more rigorous and ensure this is consistent, but it does work fine. Using these properties and some algebra we get: \begin{align*} {\bf v}&=v_1{\bf i}+v_2{\bf j}+v_3{\bf k} \\ {\bf w}&=w_1{\bf i}+w_2{\bf j}+w_3{\bf k} \\ {\bf v\times w}&=(v_1{\bf i}+v_2{\bf j}+v_3{\bf k})\times(w_1{\bf i}+w_2{\bf j}+w_3{\bf k})\\ &=v_1w_1({\bf i\times i})+v_1w_2({\bf i\times j})+v_1w_3({\bf i\times k})\\ &{\quad}+v_2w_1({\bf j\times i})+v_2w_2({\bf j\times j})+v_2w_3({\bf j\times k})\\ &{\quad}+v_3w_1({\bf k\times i})+v_3w_2({\bf k\times j})+v_3w_3({\bf k\times k})\\ &=v_1w_1{\bf 0}+v_1w_2{\bf k}-v_1w_3{\bf j}\\ &{\quad}-v_2w_1{\bf k}+v_2w_2{\bf 0}+v_2w_3{\bf i}\\ &{\quad}+v_3w_1{\bf j}-v_3w_2{\bf i}+v_3w_3{\bf 0}\\ &=(v_2w_3-v_3w_2){\bf i}+(v_1w_3-v_3w_1){\bf j}+(v_1w_2-v_2w_1){\bf k} \end{align*} which is the rectangular form I gave above.
We see that both the Polar and Rectangular forms of Cross Product describe the vector which has magnitude the area of the parallelogram and direction defined by the Right Hand Rule. They describe the same thing, using very different methods. They are equivalent methods. Considering the Polar forms of dot and cross products and a parallelepiped (a skew parallelogram prism) formed by three vectors we see that \(\ \left|\,(\bf a\times b)\bullet c\,\right|\ \) is its volume, \(\texttt{base}\cdot\texttt{height.}\) If the angle of \(\bf c\) from the perpendicular is `phi` then the height is \(\,|{\bf c}|\cos\phi\,\) (the angle of \(\bf c\) from the \(\texttt{base}\) is \(\frac\pi2-\phi\) which is why we mix Dot and Cross). Hence the triple product, \(\,|{\bf a\times b}||{\bf c}|\cos\phi,\,\) is used. It does not matter which pair are Cross, the geometry is symmetrical, the sign will change depending on the orientation of the vectors, that is the cyclic order of the operations. Flipping the order of the Cross will flip the sign, Dot is commutative. The absolute value will remain the same.
In Rectangular Form we find that the triple product is equivalent to the determinant of the \(3{\times}3\) Matrix formed by the three vectors. this is written \(|A|,\) so; \[({\bf a\,b\,c})={\bf a}\bullet({\bf b\times c})=\left|\matrix{a_1&a_2&a_3\cr b_1&b_2&b_3\cr c_1&c_2&c_3}\right|\]
This suggests extending to more dimensions — well beyond the scope of our school of course. The Determinant has several interpretations. The idea is millenia old, imported into europe along with algebra a bit over 500 years ago. The european name is well before Matrices, they were developed to support linear algebra. Determinant comes from linear algebra, in part studying the simultaneous equations we touch on early in our school algebra. It appears in the “The Nine Chapters on the Mathematical Art” the classical chinese mathematical text, a compilation of older knowledge in their typical classical style, published around 2300 years ago. It shows whether the equations have a unique solution. The determinant is also closely related to eigenvalues and eigenvectors and in turn deeply part of quantum mechanics where Complex Matrices play a large role. It is often more important as a theoretical rather than practical method — there are more efficient approaches to most actual calculations. It is about understanding a system in abstract, and matrices are fundamental to the description of the cosmos and the descriptions on atomic scale. And to modelling any big system, say climate or business logistics. And to computation, hence most of the technology around us.
We are starting to see here how Space, Scalar, Vectors and Matrices fit together. We have a vector, \(\bf v\), in `n`-space. We can describe it as an arrow, a magnitude with a direction. We can describe the same entity as a list of scalars, `v_i`. In another interpretation it is a single column or single row matrix `V` and can participate in matrix operations. Each of these representations can also describe a position in `n`-dimensional geometrical space.
Position in one dimensional space can be described by a scalar, often a Real Number. Direction here is sign, positive or negative, forward or back. That is our very familiar number line — we need to pick a ‘zero’ point and also a ‘one’ point which gives the scalar its ‘unit’. Then we can use numbers to describe position. Complex Numbers, another kind of scalar, do not fit on a line. They do not have that kind of ‘order’. But they do behave like numbers in other ways, \({+}{\times}{=}{1}{\div}{0}{-}\) all behave properly. \(\ b^a\sin a\,\log_b a\ \) and number algebra are better behaved than Real Numbers. They have this extra number, \(\ \rm i.\,\) We can place it distance `1` from ‘zero’ perpendicular to the number line. We pick that point \(\,\rm i\,\) and we see the Complex Numbers as all the points on a Number Plane. Now we have a magnitude and direction sense of these numbers. Direction here is angle rather than sign. It no longer defines an ‘order’, it is a different kind of positional relationship instead. Numbers `pi` and \(\rm e\) emerge as fundamental to this system, linking its geometrical and component representations. The trig and exponential functions are deeply involved here.
A scalar which we can also see as 2D, things are getting interesting. Vectors or Cartesian points can represent a plane, but \(\ {\times}{\div}a^b\sin\log\ \)do not make much sense with vectors or with geometrical points. They are perfectly sensible as Complex Number operations. We are pulling together algebra, geometry and trigonometry, creating a single framework. It turns out that nature is quite happy with this kind of quantity, several values we can measure make a great deal of sense as Complex Numbers. But that does challenge our geometrical intuition, and we need to develop our mathematical language to support this. And to support the fact that space and time are not simple, flat and independent in the way that geometrical intuition derived from our senses and human-scale experience might suggest. Here matrices become very important.
Consider a scalar, \(s,\) as the one-element matrix \(S=\left[s\right]\) and as a 1D vector \(\bf s\) with magnitude `|s|`, as above. Consider \(\bf v\) and corresponding single column matrix `V,` \(s{\bf v}\) is \(SV.\) Consider another vector `W`. `W^T` is then a row vector, \({\bf w\bullet v}\) is \(W^TV\) if the single entry matrix result is considered a scalar. Look here for more detail on these matrix operations — I have details a little outside this page since here I am interested in exploring and extending the ideas covered in the HSC.
If convenient we can just work with Matrices, scalars are a single entry matrix, vectors are a single column matrix, their transpose is a single row matrix, scalar multiplication and dot product are defined as above. There is much more that can be done to extend these with matrices.
Linear transformations with matrices is very important, but I’ve seperated it out since we do not touch then formally in school. They become deeply part of the language of physics. They are fundamental to computations in many areas. As mathematical language they have many applications.
A linear transformation which is an isomorphism from a space to itself is a square matrix. Matrix multiplication is the composition of the transformations.
We can go another step here and find matrices that behave like Complex Numbers. They can be added and multiplied, scalar multiplication by Reals works. Multiplying as a Matrix acts like multiplying as a Complex Number. These 2D-scale–rotate transformation matrices are a subset of all the linear transformations with matrices or operators on a 2D vector space, they are a subset of all the 2\({\times}\)2 matrices.
\(\left[\matrix{\cos\theta & -\sin\theta \cr \sin\theta & \cos\theta}\right],\quad\) Rotation by `theta` anticlockwise
These correspond to the Complex Numbers with unit magnitude, consider \(\,{\rm e}^{ {\rm i}\theta}=\cos\theta+{\rm i}\sin\theta.\) The rest of the Complex Numbers are found by scalar multiplication, that is rotate then scale.
\[z=a+b{\sf i}=r(\cos\theta+{\sf i}\sin\theta)\quad\text{is}\quad\left[\matrix{a&-b\cr b&a}\right]=Z\]
Multiplying in Complex Numbers can be done equivalently in either Polar form or algebraically in \(a+b\rm i\) form. In Polar form we multiply the magnitudes and add the angles. Considered in the Complex Number Plane this is a scale and rotate transformation. Matrix multiplication performs transformations, much more varied transformations than just these. In the 2D case \(Z\) above performs exactly the same stretch and rotation that \(z\) does, this is broadly consistent and we can represent Complex Numbers this way as 2\({\times}\)2 square real matrices.
Here we open the door to a whole world of mathematics that can describe the world, but we end up with entities being essentially functions, complex numbers as values and all kinds of things very different from the intuition we get from observations with our senses. operator (physics)#Operators_in_quantum_mechanics linear transformations with matrices
This idea we touch on in high school, the idea that every point in a space might be associated with a value. There is another very important mathematical meaning for the word Field, quite different, referring to the kind of numbers which allow division for any pair except dividing by `0` — which are a very important kind of scalar, well known examples are Rational, Real and Complex Numbers — the scalars of a Vector Space must be this kind but I avoided using the word because this new usage was coming up. Initially these values are scalars in a Scalar Field, generally our familiar real numbers. We do not use the word field at first, but graphing a function in a 2D space is really showing a Scalar Field on a one dimensional space by adding another axis to show the value at each point — as distance above the number line. It is a very powerful way of imagining and working with a function. A function is a description of a field if the input can be considered the coordinates in a space, since then it assigns a unique value to each point. A lot of the last three years at school deals with this!
But we are also familiar with a Vector Field, and certainly deal with them in high school — the wind map that the weather bureau puts out is a Vector Field on two dimensions, every point on the map has a wind speed and direction so every point has a vector associated with it and we draw that with little arrows for the vectors placed on their points in the map. This is a very important kind of analysis used in engineering, manufacturing, computation, planning, predicting complex behaviors and understanding the results of studies. A lot of the core university maths is the calculus that deals with that, looking at rates of change of those vectors and sums of those vectors in different ways. An extension of the way school calculus looks at rates of change and sums of velocity in projectile motion.
The reason I dove into this quantum computing rabbit hole and landed here on matrices is because these can be very hard computations which quantum computing can tackle in an entirely different way. And that wave is just breaking, especially in the last few months. Many calculations essentially impossible classically are starting to become quite reasonable within a Quantum Computer and they have become usefully operational in the last few months, beyond a few very specialist applications that have been around for a few years. They are rapidly scaling up towards doing what was very recently impossible. Neural Nets are another computational approach that has already shattered old limitations. Work 40 years ago that ultimately led to them as today’s very effective mathematical and computational tool was given the physics Nobel Prize 2024, and applications were given the Chemistry Nobel Prize. They are also known as AI. Both Quantum and Neural Net calculations are profoundly mathematical approaches that are only comprehensible as pure mathematical structures, and matrices are deeply part of them.
At quantum level reality is pure equations, matrices and complex numbers are very real and to use that for computation we look for how we can turn a problem into equations that we can set up across a collection of entangled electrons or photons or structures made with them and watch them resolve the question almost instantly – then read out the answer. That sounds complicated, but is actually a gross oversimplification. This stuff truly is hard and counter intuitive, and it is pure maths that gives the handles to help the imagination get there and shift our intuition into this very different scale and world.
Consider our familiar functions like \({\rm f}(x)\) giving a real number for each real number. We can view this collection as a Vector Space which includes all possible examples of such a function. It has infinite dimension, and not just the infinite of the integers. In that case each would be an endless list of values one after the other. But this is the much bigger infinite of the number of real numbers. These vectors have an independent component for every distinct input. Each different list of such components describes a different function. The hypothetical list defines each output given the input. It is a way to give the function mapping rule. A very long and dense list, thats for sure! But that is an infinitely long, infinitely dense list of real numbers otherwise nicely ordered (like Reals are). As a vector there is a distinct axis or direction for each potential distinct input, whatever that might mean.
Imagine a graph, the number line with graph above or below it as given by the output of one of the functions in that collection. The usual kind of graph we draw. Now consider that graph as a list of values along the number line. The value ‘listed’ for each point on the horizontal axis is the height above that axis of the line we drew. Some nice, detailed list of values indeed! With a value given for each point on the number line the dimension is the size of the set of all real numbers. Earlier we had thought of such a function as a Scalar Field, or thought of a Scalar Field as a function. We can alternatively consider this curved line as a curved one dimensional Space, though this is more interesting when the dimensions are higher. An example of a Scalar Field on 2D cartesian space is a contour map, it describes the surface of the land by giving the elevation at each place. This surface we walk around in is certainly a curved 2D space, not a euclidean one. We can certainly perceive its slope, see its curvature and consider volume under (or over) — which becomes rather real when the contours go below sea level and the ocean fills that negative space.
This way we can start to build intuition, each time we reimagine an abstract something in different ways it gives us a new perspective on those things. Pushing that to a 4D minkowski spacetime which is extended as elastic and with curvature everywhere is our main description of the cosmos. Add torsion, twist, everywhere and perhaps we are approaching the next cosmological description, but leave that to today and tomorrows leading physicists to think about, it is very much one approach to some of the open questions in current physics. Considering the underlying spacetime and the Big Bang one suggestion, certainly not majority view at present but backed by some very interesting calculations from recent detailed observations, is that it is closed in on itself, finite, in an essentially flat hyper-torus kind of structure, expanding and presently maybe only 3 or 4 times longer across than the longest observable distance, distance light could have travelled since the Big Bang. This combines very cleanly with the torsion ideas, and shifts our models on a fairly deep philosophical level and that kind of shift does not happen overnight! But I personally would certainly not bet against it.
In calculus we consider the function, its derivatives and integrals. An interpretation of derivatives and second derivatives is slope and curvature of that line, that is that space. And when we dig deep in the world, going beyond the scales we can touch and see, we must rely on very abstract descriptions, we need to build this intuition about how what we are describing behaves, what it is. If something in the world is a function we need to expand how we think about functions. At the atomic scale entities do not have a nice position and trajectory, this fails completely as a description. They are fuzzy, they interact, they merge into larger entities, they are best described as functions.
The space in which we exist does not fit a nice three dimensional euclidean space with time flowing independently. That description is a reasonably good approximation as long as we stay very much within the scales and energies we can live in and perceive with our senses. But it does not suffice for developing modern technologies, or describing the cosmos. We exist in a space much more accurately described as a four dimensional spacetime which is dynamically curved and fundamentally elastic. Mass, energy, gravity and this curvature are intimately linked. But that is well beyond school maths, after extension 2 you are in a place to get started, you have some of the language required to talk about all this. You will need a lot more.
This kind of classifying things can be very useful in understanding how they work, and generalising and abstracting them. And building our intuition about structures we cannot just pick up and look at with our human senses. This is getting way outside the scope of high school maths, so I will leave it there. To tackle extension maths means building towards that kind of intuition but quite informally. We do not do a great deal of systematic proving of the properties of the structures we are learning to work with. We do not look at the formal relationships and language that can describe much of what I have spoken of in this discussion much more rigorously.
But it is useful to remember that we can consistently add pure functions and multiply them by a scalar — like we can vectors, they have linearity in this sense, they can be entities in a space, a vector aka linear space. Reasoning this way is very useful understanding what is happening when we differentiate more complicated functions and we certainly do this in high school.
This is something we use all the time in school problem solving, without ever naming it. It is fundamental to mathematical methods and thinking. The word and idea is part of formal thinking and reasoning about mathematical structures, for example about the relationship between \(\ y={\rm f}(x)\ \) and a graph of that equation on the cartesian plane with `x` and `y` on the axes.
An Isomorphism is a mapping between structures that preserves significant aspects of their behavior. It can give us a great insight into either form because we have developed insight in each form quite differently. An early example of this is decimals, mixed numerals and improper fractions. We learn to convert between them and each gives us a different insight into the quantity or ratio represented. Each helps us solve different kinds of problems. The process of drawing a graph of an equation is an Isomorphism. Two structures can be Isomorphic. In this case the equation and the graph both show the same underlying relationship between the values of `x` and `y`. For some questions about that relationship an equation and the algebra we know about equations will give us what we need. For other questions about the relationship the shape of the graph and what we know about geometry will be a much more direct answer. More likely in the context of our school program is that the graph will give us essential insights into the problem which can then be followed up with techniques we have learnt about functions.
A graph is not much help understanding the relationship it describes if you do not have strong geometric insight. Developing that insight is the main goal of the Measurement and space thread that has been in the syllabus since kindy. Recently renamed — it had been called Measurement and Geometry. Geometry was demoted below the core threads in 2024, but remains in many content areas. Geometrical construction has been dropped, it seems, as explicit content. But geometrical reasoning most certainly has not been, and to teach and develop it surely requires understanding constructions. So it is still in there. Without understanding, problem solving, reasoning, communications and fluency in these remaining content areas there is little geometrical insight, half of ‘relationships’ is missing. They are vital elements of the syllabus and must be tackled and tested seriously. Equations and functions are part of the Number and algebra thread, also since kindy. Again the main goal here is understanding mathematical descriptions of quantity of the numerical kind. The power of mathematical reasoning and problem solving is using very different insights into a situation enabled by multiple different models of it.
The kit you brought to the beginning of high school, assuming you did year 7 before 2024, almost certainly included a box with compass, dividers, protractor, a ruler and two triangles. Unless you were very lucky and had great teachers you probably hardly touched it. That part of the syllabus was usually very badly neglected. Somehow proper geometric construction on paper never quite got into the classroom. It could be replaced by geometrical construction on a computer, but is not in practice. A ruler to measure distances occasionally (but mostly make your working look neat), a couple of lessons were you had to measure angles with the protractor (that is still in content) and maybe drawing a few circles with the compass (mainly to make your diagrams pretty). If pencil and paper is too old fashioned you must fully embrace digital constructions in the mathematics classroom and in homework, which in many ways is more sophisticated while demanding the same kind of practical understanding of spatial relationships. In the digital realm the connection between number and geometry is much more explicit. There are Computer Aided Design programs that everybody now uses rather than a pencil — but with the real ones you do constructions the same way, drawing points, lines, angles, circles, intersections and such. You must understand the constructions and have, or develop, strong geometrical intuitions using them. Maybe you have independently done some design or computer programming that involves computer graphics. Maybe you happened to get interested in some older books of mathematical puzzles, there are a great many of those still around. Maybe you get out saws and drills, or a 3D printer, from time to time and build something.
Otherwise you have a lot of geometrical intuition building to catch up on! You do need it. It is expected in the exam questions, it is part of the maths syllabus. It is vital in many real world applications and to developing spatial understanding.
This becomes a problem with the Bearings topic, in the middle of trig a student suddenly needs to understand constructing lines, angles, points and such. Then the reasoning in geometric proofs is very often about constructing shapes. These are the alternatives but usually the first exposure to this kind of mathematical thinking is a few proofs. This is jumping in the deep end to learn to swim — most sink, some quickly realise they can flap their arms and legs around and effectively keep their head above water. For this reason very few test questions actually require a proof — teachers and parents do not like lots of blank or just plain wrong answers. But these questions are in the HSC, and the geometrical insight they depend on is part of many of those questions. The reasoning (if no longer explicitly the construction required to do it) is part of the syllabus, in gradually more sophisticated ways, every year since early primary. Most classrooms sadly fail to develop it. Like number sense, geometric sense takes a long time to develop. Isomorphisms between graphical and numeric representations are not an escape from geometry, rather they bring geometric reasoning into solutions.
In this discussion we have considered Vectors as ‘magnitude with direction’ and as ‘a list of scalars’, the second is the Cartesian form. We sometimes call these the ‘Polar’ and ‘Rectangular’ forms. They are different ways of writing and saying the same underlying value. As with finite decimals and fractions. We learn to convert between them. This conversion is another isomorphism, and each form gives us different insights and different possibilities for calculations. One is algebraic, one is geometric. The power comes from the combination, not ‘oh goody I can get an answer using numbers instead!’.
A scatter graph is isomorphic with a 2-column matrix showing two distinct aspects measured for many examples. It shows that data quite clearly, and can help us ‘see’ that matrix and the data it represents. We can find or describe patterns in that data, patterns which may have been difficult to ‘see’ in the 2 column table of data. Kepler, Dirac and a few other very unique minds are able to see very complex and surprising relationships in raw data and break very new ground indeed. For the rest of us an equation and a graph helps a great deal — and calculus when we are trying to understand changes in something. The whole language of graphs, calculus and functions was invented (in that order) to explain and extend Kepler’s insights, they were not available to him at the time. They had studied formal geometry very closely. The algebraic formulations came next. Again equations and graphs, one algebraic one geometric.
The modern description of the universe, and just about everything in it, is pretty close to pure equations. Entities are blurry geometrically, they are equations that interact. Space, time and spacetime is not some kind of simple cartesian space that they, and we, exist in. It is warped and springy, and time is not what it seems. We understand space, time and entities by the various equations that describe them. Alternative equations give different insights. Geometric interpretations of the equations are critical. Algebraic and geometric intuition must both be pushed very hard here.
The philosophical and scientific journey from Galileo (Infinitesimals) to Euler (Functions and Complex Numbers) was the context for Kepler’s fabulous insights. It took about a century and is a very big part of the maths we teach in high school. The journey into Pure Maths as we understand it now came next, getting started maybe 300 years ago. That is towards a formal understanding of what Maths is, making the language and ideas rigorous, seeing the structures of structures, digging down to deeper foundations (they are endless, there is no bottom) and up to broader abstractions and generalisations. This is the language required for modern science, analysis, computation and planning. We do not go there in high school. We do look at applications of calculus and statistics in ways that developed alongside that Pure Maths journey. We hope to put a few students into a place they could be comfortable starting to follow that journey during university maths. Maybe then to participate in the scientific, analytical, engineering or computation worlds that have come from that. Dirac’s incredible insight was early in the shift from classical dynamics to the relativity and quantum mechanics understanding of the world, the shift to much more pure equations.
Going back to that 2 column data above we might be able to find an equation that fits the relationship between `x` and `y` quite closely and that would make a much more concise description of the data. It might help us understand the nature of the relationship much more deeply because we have become familiar with the nature certain equations, and those insights can be applied to any relationship described by those equations. We aim to become very familiar with linear, quadratic, exponential and trigonometric relationships especially, and if we get there we can nail lots of problems in the HSC exams. And they all turn up everywhere in nature, technology, economics and planning.
Throughout school we have learnt procedures where we can represent something in a different way, with different notation or in a quite distinct form. Then we use methods we have learnt in one representation to solve problems presented in another way. But we do not think too much about what we are actually doing. Though I certainly try to make it clear what this flipping between forms is, that these forms are equivalent but different representations of the same thing, I never use the term Isomorphic — it is a bit intimidating and following it up (something I hope students might do occasionally after our discussions!) you would get into all manner of very formal pure maths of a kind very different to what we look at in school. That is for uni, but only if you choose the Pure Maths rather than the stats or analysis paths.