Linear Algebra: The sum of dot-products summarized in a matrix matrix product The 2019 Stack Overflow Developer Survey Results Are InRelationship between the rows and columns of a matrixWhat linear-algebra operation produces this vector?Linear Algebra - Getting Orthogonal Matrix?The “extended” dot productCombining Column- and Row-wise meanings of a matrixWhat is an intuitive way to understand the dot product in the context of matrix multiplication?Basic Linear Algebra VisualisationElement wise versus dot product rules for matrix multiplicationLinear combination of vectors vs Dot product in Matrix vector multiplicationMatrix-vector product: representing matrix as vector of vectors seemingly leads to paradox when transposing the matrix
Why not take a picture of a closer black hole?
Geography at the pixel level
Why isn't the circumferential light around the M87 black hole's event horizon symmetric?
What is the accessibility of a package's `Private` context variables?
How come people say “Would of”?
Did Scotland spend $250,000 for the slogan "Welcome to Scotland"?
Delete all lines which don't have n characters before delimiter
Loose spokes after only a few rides
Can one be advised by a professor who is very far away?
Is flight data recorder erased after every flight?
slides for 30min~1hr skype tenure track application interview
How to support a colleague who finds meetings extremely tiring?
Time travel alters history but people keep saying nothing's changed
Is there a symbol for a right arrow with a square in the middle?
How to answer pointed "are you quitting" questioning when I don't want them to suspect
Pokemon Turn Based battle (Python)
How to save as into a customized destination on macOS?
A poker game description that does not feel gimmicky
Does a dangling wire really electrocute me if I'm standing in water?
What to do when moving next to a bird sanctuary with a loosely-domesticated cat?
How to check whether the reindex working or not in Magento?
What do hard-Brexiteers want with respect to the Irish border?
Is bread bad for ducks?
Is "plugging out" electronic devices an American expression?
Linear Algebra: The sum of dot-products summarized in a matrix matrix product
The 2019 Stack Overflow Developer Survey Results Are InRelationship between the rows and columns of a matrixWhat linear-algebra operation produces this vector?Linear Algebra - Getting Orthogonal Matrix?The “extended” dot productCombining Column- and Row-wise meanings of a matrixWhat is an intuitive way to understand the dot product in the context of matrix multiplication?Basic Linear Algebra VisualisationElement wise versus dot product rules for matrix multiplicationLinear combination of vectors vs Dot product in Matrix vector multiplicationMatrix-vector product: representing matrix as vector of vectors seemingly leads to paradox when transposing the matrix
$begingroup$
While reading a section on orthogonal matrices, where $vecu_1, ..., vecu_n$ is an orthonormal set of vectors, I find the equation
$$(vecu_1vecu_1^t + ... + vecu_nvecu_n^t)vecx= beginbmatrix
| & ... & | \
vecu_1 & ... & vecu_n \
| & ... & |
endbmatrixbeginbmatrix
-vecu_1^t-\
...\
-vecu_n^t-endbmatrixvecx$$
Now I'm familiar with summarizing a linear combination of vectors by a matrix-vector product, but in doing so I use my concept of the dot product between a row vector in the matrix and the column vector written to the right. This is related to how, in the matrix product, you're basically doing a bunch of dot-products between rows in the left matrix and columns in the right.
Here is something different though, and I'm not sure how I ought to think about it. The columns are written in the left matrix and rows in the right. I don't see how this is consistent with the dot-product view of matrix products. I could chase indices by thinking of the coordinate-wise definition of matrix products, but I feel like the author intends for a better more natural way of understanding this equation.
So is there a good column-row understanding of matrix products, which makes the above equation sensible? Or is the only good way of understanding this by thinking in terms of the coordinate-wise definition of matrix products?
linear-algebra matrices matrix-equations
$endgroup$
add a comment |
$begingroup$
While reading a section on orthogonal matrices, where $vecu_1, ..., vecu_n$ is an orthonormal set of vectors, I find the equation
$$(vecu_1vecu_1^t + ... + vecu_nvecu_n^t)vecx= beginbmatrix
| & ... & | \
vecu_1 & ... & vecu_n \
| & ... & |
endbmatrixbeginbmatrix
-vecu_1^t-\
...\
-vecu_n^t-endbmatrixvecx$$
Now I'm familiar with summarizing a linear combination of vectors by a matrix-vector product, but in doing so I use my concept of the dot product between a row vector in the matrix and the column vector written to the right. This is related to how, in the matrix product, you're basically doing a bunch of dot-products between rows in the left matrix and columns in the right.
Here is something different though, and I'm not sure how I ought to think about it. The columns are written in the left matrix and rows in the right. I don't see how this is consistent with the dot-product view of matrix products. I could chase indices by thinking of the coordinate-wise definition of matrix products, but I feel like the author intends for a better more natural way of understanding this equation.
So is there a good column-row understanding of matrix products, which makes the above equation sensible? Or is the only good way of understanding this by thinking in terms of the coordinate-wise definition of matrix products?
linear-algebra matrices matrix-equations
$endgroup$
add a comment |
$begingroup$
While reading a section on orthogonal matrices, where $vecu_1, ..., vecu_n$ is an orthonormal set of vectors, I find the equation
$$(vecu_1vecu_1^t + ... + vecu_nvecu_n^t)vecx= beginbmatrix
| & ... & | \
vecu_1 & ... & vecu_n \
| & ... & |
endbmatrixbeginbmatrix
-vecu_1^t-\
...\
-vecu_n^t-endbmatrixvecx$$
Now I'm familiar with summarizing a linear combination of vectors by a matrix-vector product, but in doing so I use my concept of the dot product between a row vector in the matrix and the column vector written to the right. This is related to how, in the matrix product, you're basically doing a bunch of dot-products between rows in the left matrix and columns in the right.
Here is something different though, and I'm not sure how I ought to think about it. The columns are written in the left matrix and rows in the right. I don't see how this is consistent with the dot-product view of matrix products. I could chase indices by thinking of the coordinate-wise definition of matrix products, but I feel like the author intends for a better more natural way of understanding this equation.
So is there a good column-row understanding of matrix products, which makes the above equation sensible? Or is the only good way of understanding this by thinking in terms of the coordinate-wise definition of matrix products?
linear-algebra matrices matrix-equations
$endgroup$
While reading a section on orthogonal matrices, where $vecu_1, ..., vecu_n$ is an orthonormal set of vectors, I find the equation
$$(vecu_1vecu_1^t + ... + vecu_nvecu_n^t)vecx= beginbmatrix
| & ... & | \
vecu_1 & ... & vecu_n \
| & ... & |
endbmatrixbeginbmatrix
-vecu_1^t-\
...\
-vecu_n^t-endbmatrixvecx$$
Now I'm familiar with summarizing a linear combination of vectors by a matrix-vector product, but in doing so I use my concept of the dot product between a row vector in the matrix and the column vector written to the right. This is related to how, in the matrix product, you're basically doing a bunch of dot-products between rows in the left matrix and columns in the right.
Here is something different though, and I'm not sure how I ought to think about it. The columns are written in the left matrix and rows in the right. I don't see how this is consistent with the dot-product view of matrix products. I could chase indices by thinking of the coordinate-wise definition of matrix products, but I feel like the author intends for a better more natural way of understanding this equation.
So is there a good column-row understanding of matrix products, which makes the above equation sensible? Or is the only good way of understanding this by thinking in terms of the coordinate-wise definition of matrix products?
linear-algebra matrices matrix-equations
linear-algebra matrices matrix-equations
asked Mar 30 at 17:54
AddemAddem
1,7561429
1,7561429
add a comment |
add a comment |
1 Answer
1
active
oldest
votes
$begingroup$
This can be written as
$$
pmatrix vec u_1 & cdots & vec u_n
left[
pmatrix vec u_1^T \ vdots \ vec u_n^T
vec x
right],
$$
where the product in brackets are dot products of $x$ and vectors $u_i$. The resulting numbers are then used to get a linear combination of the vectors $u_i$.
$endgroup$
add a comment |
Your Answer
StackExchange.ifUsing("editor", function ()
return StackExchange.using("mathjaxEditing", function ()
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix)
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
);
);
, "mathjax-editing");
StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "69"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);
else
createEditor();
);
function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader:
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
,
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);
);
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3168578%2flinear-algebra-the-sum-of-dot-products-summarized-in-a-matrix-matrix-product%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
1 Answer
1
active
oldest
votes
1 Answer
1
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
This can be written as
$$
pmatrix vec u_1 & cdots & vec u_n
left[
pmatrix vec u_1^T \ vdots \ vec u_n^T
vec x
right],
$$
where the product in brackets are dot products of $x$ and vectors $u_i$. The resulting numbers are then used to get a linear combination of the vectors $u_i$.
$endgroup$
add a comment |
$begingroup$
This can be written as
$$
pmatrix vec u_1 & cdots & vec u_n
left[
pmatrix vec u_1^T \ vdots \ vec u_n^T
vec x
right],
$$
where the product in brackets are dot products of $x$ and vectors $u_i$. The resulting numbers are then used to get a linear combination of the vectors $u_i$.
$endgroup$
add a comment |
$begingroup$
This can be written as
$$
pmatrix vec u_1 & cdots & vec u_n
left[
pmatrix vec u_1^T \ vdots \ vec u_n^T
vec x
right],
$$
where the product in brackets are dot products of $x$ and vectors $u_i$. The resulting numbers are then used to get a linear combination of the vectors $u_i$.
$endgroup$
This can be written as
$$
pmatrix vec u_1 & cdots & vec u_n
left[
pmatrix vec u_1^T \ vdots \ vec u_n^T
vec x
right],
$$
where the product in brackets are dot products of $x$ and vectors $u_i$. The resulting numbers are then used to get a linear combination of the vectors $u_i$.
answered Apr 1 at 11:34
dawdaw
25.1k1745
25.1k1745
add a comment |
add a comment |
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3168578%2flinear-algebra-the-sum-of-dot-products-summarized-in-a-matrix-matrix-product%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown