In computing, row-major order and column-major order are methods for storing multidimensional arrays in linear storage such as random access memory.
The difference between the orders lies in which elements of an array are contiguous in memory. In row-major order, the consecutive elements of a row reside next to each other, whereas the same holds true for consecutive elements of a column in column-major order. While the terms allude to the rows and columns of a two-dimensional array, i.e. a matrix, the orders can be generalized to arrays of any dimension by noting that the terms row-major and column-major are equivalent to lexicographic and colexicographic orders, respectively. It is also worth noting that matrices, being commonly represented as collections of row or column vectors, using this approach are effectively stored as consecutive vectors or consecutive vector components. Such ways of storing data are referred to as AoS and SoA respectively.
Data layout is critical for correctly passing arrays between programs written in different programming languages. It is also important for performance when traversing an array because modern CPUs process sequential data more efficiently than nonsequential data. This is primarily due to CPU caching which exploits spatial locality of reference.[1] In addition, contiguous access makes it possible to use SIMD instructions that operate on vectors of data. In some media such as magnetic-tape data storage, accessing sequentially is orders of magnitude faster than nonsequential access.
The terms row-major and column-major stem from the terminology related to ordering objects. A general way to order objects with many attributes is to first group and order them by one attribute, and then, within each such group, group and order them by another attribute, etc. If more than one attribute participates in ordering, the first would be called major and the last minor. If two attributes participate in ordering, it is sufficient to name only the major attribute.
In the case of arrays, the attributes are the indices along each dimension. For matrices in mathematical notation, the first index indicates the row, and the second indicates the column, e.g., given a matrix
A
a1,2
Even though the row is indicated by the first index and the column by the second index, no grouping order between the dimensions is implied by this. The choice of how to group and order the indices, either by row-major or column-major methods, is thus a matter of convention. The same terminology can be applied to even higher dimensional arrays. Row-major grouping starts from the leftmost index and column-major from the rightmost index, leading to lexicographic and colexicographic (or colex) orders, respectively.
For example, the array
A=ay,x=\begin{bmatrix} \color{Blue}a11&\color{Blue}a12&\color{Blue}a13\\ \color{Orange}a21&\color{Orange}a22&\color{Orange}a23\end{bmatrix}
could be stored in two possible ways:
Address | Row-major order | Column-major order | |
---|---|---|---|
0 | \color{Blue}a11 | \color{Blue}a11 | |
1 | \color{Blue}a12 | \color{Orange}a21 | |
2 | \color{Blue}a13 | \color{Blue}a12 | |
3 | \color{Orange}a21 | \color{Orange}a22 | |
4 | \color{Orange}a22 | \color{Blue}a13 | |
5 | \color{Orange}a23 | \color{Orange}a23 |
Programming languages handle this in different ways. In C, multidimensional arrays are stored in row-major order, and the array indexes are written row-first (lexicographical access order):
Addressx + N_x*y | AccessA[y][x] | Value ay,x | |
---|---|---|---|
0 | A[0][0] | a11 | |
1 | A[0][1] | a12 | |
2 | A[0][2] | a13 | |
3 | A[1][0] | a21 | |
4 | A[1][1] | a22 | |
5 | A[1][2] | a23 |
On the other hand, in Fortran, arrays are stored in column-major order, while the array indexes are still written row-first (colexicographical access order):
Addressy + N_y*(x-1) | AccessA(y,x) | Value ay,x | |
---|---|---|---|
1 | A(1,1) | a11 | |
2 | A(2,1) | a21 | |
3 | A(1,2) | a12 | |
4 | A(2,2) | a22 | |
5 | A(1,3) | a13 | |
6 | A(2,3) | a23 |
Note how the use of A[i][j]
with multi-step indexing as in C, as opposed to a neutral notation like A(i,j)
as in Fortran, almost inevitably implies row-major order for syntactic reasons, so to speak, because it can be rewritten as (A[i])[j]
, and the A[i]
row part can even be assigned to an intermediate variable that is then indexed in a separate expression. (No other implications should be assumed, e.g., Fortran is not column-major simply because of its notation, and even the above implication could intentionally be circumvented in a new language.)
To use column-major order in a row-major environment, or vice versa, for whatever reason, one workaround is to assign non-conventional roles to the indexes (using the first index for the column and the second index for the row), and another is to bypass language syntax by explicitly computing positions in a one-dimensional array. Of course, deviating from convention probably incurs a cost that increases with the degree of necessary interaction with conventional language features and other code, not only in the form of increased vulnerability to mistakes (forgetting to also invert matrix multiplication order, reverting to convention during code maintenance, etc.), but also in the form of having to actively rearrange elements, all of which have to be weighed against any original purpose such as increasing performance. Running the loop row-wise is preferred in row-major languages like C and vice versa for column-major languages.
Programming languages or their standard libraries that support multi-dimensional arrays typically have a native row-major or column-major storage order for these arrays.
Row-major order is used in C/C++/Objective-C (for C-style arrays), PL/I,[4] Pascal,[5] Speakeasy, and SAS.[6]
Column-major order is used in Fortran, IDL,[7] MATLAB,[8] GNU Octave, Julia,[9] S, S-PLUS, R,[10] Scilab,[11] Yorick, and Rasdaman.[12]
A typical alternative for dense array storage is to use Iliffe vectors, which typically store pointers to elements in the same row contiguously (like row-major order), but not the rows themselves. They are used in (ordered by age): Java,[13] C#/CLI/.Net, Scala,[14] and Swift.
Even less dense is to use lists of lists, e.g., in Python,[15] and in the Wolfram Language of Wolfram Mathematica.[16]
An alternative approach uses tables of tables, e.g., in Lua.[17]
Support for multi-dimensional arrays may also be provided by external libraries, which may even support arbitrary orderings, where each dimension has a stride value, and row-major or column-major are just two possible resulting interpretations.
Row-major order is the default in NumPy[18] (for Python).
Column-major order is the default in Eigen[19] and Armadillo (both for C++).
A special case would be OpenGL (and OpenGL ES) for graphics processing. Since "recent mathematical treatments of linear algebra and related fields invariably treat vectors as columns," designer Mark Segal decided to substitute this for the convention in predecessor IRIS GL, which was to write vectors as rows; for compatibility, transformation matrices would still be stored in vector-major (=row-major) rather than coordinate-major (=column-major) order, and he then used the trick "[to] say that matrices in OpenGL are stored in column-major order".[20] This was really only relevant for presentation, because matrix multiplication was stack-based and could still be interpreted as post-multiplication, but, worse, reality leaked through the C-based API because individual elements would be accessed as M[vector][coordinate]
or, effectively, M[column][row]
, which unfortunately muddled the convention that the designer sought to adopt, and this was even preserved in the OpenGL Shading Language that was later added (although this also makes it possible to access coordinates by name instead, e.g., M[vector].y
). As a result, many developers will now simply declare that having the column as the first index is the definition of column-major, even though this is clearly not the case with a real column-major language like Fortran.
Torch (for Lua) changed from column-major[21] to row-major[22] default order.
As exchanging the indices of an array is the essence of array transposition, an array stored as row-major but read as column-major (or vice versa) will appear transposed. As actually performing this rearrangement in memory is typically an expensive operation, some systems provide options to specify individual matrices as being stored transposed. The programmer must then decide whether or not to rearrange the elements in memory, based on the actual usage (including the number of times that the array is reused in a computation).
For example, the Basic Linear Algebra Subprograms functions are passed flags indicating which arrays are transposed.[23]
The concept generalizes to arrays with more than two dimensions.
For a d-dimensional
N1 x N2 x … x Nd
(n1,n2,\ldots,nd)
nk\in[0,Nk-1]
In row-major order, the last dimension is contiguous, so that the memory-offset of this element is given by:
In column-major order, the first dimension is contiguous, so that the memory-offset of this element is given by:where the empty product is the multiplicative identity element, i.e., .
For a given order, the stride in dimension k is given by the multiplication value in parentheses before index nk in the right-hand side summations above.
More generally, there are d! possible orders for a given array, one for each permutation of dimensions (with row-major and column-order just 2 special cases), although the lists of stride values are not necessarily permutations of each other, e.g., in the 2-by-3 example above, the strides are (3,1) for row-major and (1,2) for column-major.