Is there an explicit relationship between the eigenvalues of a matrix and its derivative? Announcing the arrival of Valued Associate #679: Cesar Manara Planned maintenance scheduled April 17/18, 2019 at 00:00UTC (8:00pm US/Eastern)Recovering a Matrix knowing its eigenvectors and eigenvalueseigenvalue of block matrix in terms of original matrixThe relationship between diagonal entries and eigenvalues of a diagonalizable matrixAntidiagonal block matrix (eigenvalues and eigenvectors)Eigenvalues of block matrix where blocks are relatedGradient of a function involving maximaAre the eigenvalues of the matrix AB equal to the eigenvalues of the matrix BAFinding approximate common eigenvector of two matricesMatrix transformations, Eigenvectors and EigenvaluesAre the following equations concerning matrices and their eigenvectors equivalent?
Can inflation occur in a positive-sum game currency system such as the Stack Exchange reputation system?
How do I determine if the rules for a long jump or high jump are applicable for Monks?
Why are there no cargo aircraft with "flying wing" design?
How do I stop a creek from eroding my steep embankment?
Java 8 stream max() function argument type Comparator vs Comparable
Why don't the Weasley twins use magic outside of school if the Trace can only find the location of spells cast?
Did Xerox really develop the first LAN?
Should I discuss the type of campaign with my players?
"Seemed to had" is it correct?
If Jon Snow became King of the Seven Kingdoms what would his regnal number be?
How to recreate this effect in Photoshop?
What makes black pepper strong or mild?
Should gear shift center itself while in neutral?
Gastric acid as a weapon
How widely used is the term Treppenwitz? Is it something that most Germans know?
What is the correct way to use the pinch test for dehydration?
Why did the IBM 650 use bi-quinary?
Antler Helmet: Can it work?
Why was the term "discrete" used in discrete logarithm?
Examples of mediopassive verb constructions
Is there a "higher Segal conjecture"?
How to draw this diagram using TikZ package?
Did Kevin spill real chili?
How much radiation do nuclear physics experiments expose researchers to nowadays?
Is there an explicit relationship between the eigenvalues of a matrix and its derivative?
Announcing the arrival of Valued Associate #679: Cesar Manara
Planned maintenance scheduled April 17/18, 2019 at 00:00UTC (8:00pm US/Eastern)Recovering a Matrix knowing its eigenvectors and eigenvalueseigenvalue of block matrix in terms of original matrixThe relationship between diagonal entries and eigenvalues of a diagonalizable matrixAntidiagonal block matrix (eigenvalues and eigenvectors)Eigenvalues of block matrix where blocks are relatedGradient of a function involving maximaAre the eigenvalues of the matrix AB equal to the eigenvalues of the matrix BAFinding approximate common eigenvector of two matricesMatrix transformations, Eigenvectors and EigenvaluesAre the following equations concerning matrices and their eigenvectors equivalent?
$begingroup$
If we consider a matrix $A$ dependent of a variable $x$, the eigenvalues and eigenvectors satisfying the equation
$$
A vecv=lambda vecv
$$
will also depend on $x$. If we consider the matrix $B$ such that
$$B_ij=frac mathrmd mathrmd x A_ij$$
Then, could we express the eigenvalues of $B$ in terms of the eigenvalues of $A$?
I found the question very interesting and was not able to find a satisfying answer myself.
For example in the case for $2times2$ matrices of the form
$$
A=left (
beginmatrix
a(x) & b(x) \
0 & c(x)
endmatrix
right ),implies B=left (
beginmatrix
a'(x) & b'(x) \
0 & c'(x)
endmatrix
right )
$$
I noticed that $lambda_B(x)= lambda_A'(x)$. But I cannot generalise it to general $2times 2$ matrices. Not even thinking about $ntimes n$ matrices...
Thank you for your help and any idea!
linear-algebra matrices derivatives eigenvalues-eigenvectors
$endgroup$
add a comment |
$begingroup$
If we consider a matrix $A$ dependent of a variable $x$, the eigenvalues and eigenvectors satisfying the equation
$$
A vecv=lambda vecv
$$
will also depend on $x$. If we consider the matrix $B$ such that
$$B_ij=frac mathrmd mathrmd x A_ij$$
Then, could we express the eigenvalues of $B$ in terms of the eigenvalues of $A$?
I found the question very interesting and was not able to find a satisfying answer myself.
For example in the case for $2times2$ matrices of the form
$$
A=left (
beginmatrix
a(x) & b(x) \
0 & c(x)
endmatrix
right ),implies B=left (
beginmatrix
a'(x) & b'(x) \
0 & c'(x)
endmatrix
right )
$$
I noticed that $lambda_B(x)= lambda_A'(x)$. But I cannot generalise it to general $2times 2$ matrices. Not even thinking about $ntimes n$ matrices...
Thank you for your help and any idea!
linear-algebra matrices derivatives eigenvalues-eigenvectors
$endgroup$
1
$begingroup$
One nice property of an upper (or lower) triangular matrix is that its eigenvalues are the same as its diagonal elements; which explains the $2times 2$ examples that you discovered. But this is not a general property of all matrices.
$endgroup$
– greg
Jan 27 at 20:51
add a comment |
$begingroup$
If we consider a matrix $A$ dependent of a variable $x$, the eigenvalues and eigenvectors satisfying the equation
$$
A vecv=lambda vecv
$$
will also depend on $x$. If we consider the matrix $B$ such that
$$B_ij=frac mathrmd mathrmd x A_ij$$
Then, could we express the eigenvalues of $B$ in terms of the eigenvalues of $A$?
I found the question very interesting and was not able to find a satisfying answer myself.
For example in the case for $2times2$ matrices of the form
$$
A=left (
beginmatrix
a(x) & b(x) \
0 & c(x)
endmatrix
right ),implies B=left (
beginmatrix
a'(x) & b'(x) \
0 & c'(x)
endmatrix
right )
$$
I noticed that $lambda_B(x)= lambda_A'(x)$. But I cannot generalise it to general $2times 2$ matrices. Not even thinking about $ntimes n$ matrices...
Thank you for your help and any idea!
linear-algebra matrices derivatives eigenvalues-eigenvectors
$endgroup$
If we consider a matrix $A$ dependent of a variable $x$, the eigenvalues and eigenvectors satisfying the equation
$$
A vecv=lambda vecv
$$
will also depend on $x$. If we consider the matrix $B$ such that
$$B_ij=frac mathrmd mathrmd x A_ij$$
Then, could we express the eigenvalues of $B$ in terms of the eigenvalues of $A$?
I found the question very interesting and was not able to find a satisfying answer myself.
For example in the case for $2times2$ matrices of the form
$$
A=left (
beginmatrix
a(x) & b(x) \
0 & c(x)
endmatrix
right ),implies B=left (
beginmatrix
a'(x) & b'(x) \
0 & c'(x)
endmatrix
right )
$$
I noticed that $lambda_B(x)= lambda_A'(x)$. But I cannot generalise it to general $2times 2$ matrices. Not even thinking about $ntimes n$ matrices...
Thank you for your help and any idea!
linear-algebra matrices derivatives eigenvalues-eigenvectors
linear-algebra matrices derivatives eigenvalues-eigenvectors
asked Jan 25 at 19:56
MattMatt
513
513
1
$begingroup$
One nice property of an upper (or lower) triangular matrix is that its eigenvalues are the same as its diagonal elements; which explains the $2times 2$ examples that you discovered. But this is not a general property of all matrices.
$endgroup$
– greg
Jan 27 at 20:51
add a comment |
1
$begingroup$
One nice property of an upper (or lower) triangular matrix is that its eigenvalues are the same as its diagonal elements; which explains the $2times 2$ examples that you discovered. But this is not a general property of all matrices.
$endgroup$
– greg
Jan 27 at 20:51
1
1
$begingroup$
One nice property of an upper (or lower) triangular matrix is that its eigenvalues are the same as its diagonal elements; which explains the $2times 2$ examples that you discovered. But this is not a general property of all matrices.
$endgroup$
– greg
Jan 27 at 20:51
$begingroup$
One nice property of an upper (or lower) triangular matrix is that its eigenvalues are the same as its diagonal elements; which explains the $2times 2$ examples that you discovered. But this is not a general property of all matrices.
$endgroup$
– greg
Jan 27 at 20:51
add a comment |
4 Answers
4
active
oldest
votes
$begingroup$
It is not true in general that the eigenvalues of $B(x)$ are the derivatives of those of $A(x)$. And this even for some square matrices of dimension $2 times 2$.
Consider the matrix
$$A(x) =
beginpmatrix
1& -x^2\
-x &1
endpmatrix$$
It’s characteristic polynomial is $chi_A(x)(t)=t^2-2t+1-x^3$, which has for roots $1pm x ^3/2$ for $x>0$. Those are the eigenvalues of $A(x)$.
The derivative of $A(x)$ is
$$B(x) =
beginpmatrix
0& -2x\
-1 &0
endpmatrix$$ and it’s characteristic polynomial is $chi_B(x)(t)=t^2-2x$, whose roots are $pm sqrt2 x^1/2$for $x>0$.
We get a counterexample as the derivative of $1+x^3/2$ is not $sqrt2x^1/2$.
However in the special case of upper triangular matrices (that you consider in your original question) the eigenvalues of the matrix derivative are indeed the derivatives of the eigenvalues.
$endgroup$
add a comment |
$begingroup$
It can be shown that the eigenvalues of the derivative of the matrix cannot be derived from the eigenvalues of the original matrix. Example:
$$
A_1 = beginpmatrix -1 & 0 \ 0 & 1 endpmatrix
;;; , ;;;
A_2 = beginpmatrix 0 & e^x \ e^-x & 0 endpmatrix
$$
Both of the matrices above have the eigenvalues $-1$ and $1.$ However, the derivative of the first matrix has the eigenvalue $0$ (with multiplicity 2), while the derivative of the second matrix has the eigenvalues $i$ and $-i.$
Just given the eigenvalues $-1$ and $1$, there is no way of telling which matrix they originate from, hence no way of getting the eigenvalues of the derivative.
$endgroup$
add a comment |
$begingroup$
Let $alpha_k,beta_k$ be the eigenvalues of $(A,B)$
where $B(x) = A'(x).$
A class of matrices for which $beta_k=alpha_k',$ can be constructed as follows.
Choose an orthogonal matrix $Q$ and an upper triangular matrix $U(x).$
$$eqalign
A &= Q,U,Q^-1 cr
A' &= Q,U'Q^-1 cr
$$
Since $Q$ is orthogonal, the eigenvalues of $(U,U')$ equal the eigenvalues of $(A,A')$, respectively.
Since the eigenvalues of a triangular matrix are its diagonal elements,
the EVs of $U$ are $U_kk=alpha_k,$ and
the EVs of $U'$ are $U'_kk=alpha_k'=beta_k.$
NB: $,Q$ must be independent of $x$ for this construction to apply.
$endgroup$
add a comment |
$begingroup$
If $A(x)v= lambda(x)v$, with v independent of x, then, differentiating on both sides of the equation by x, $A'(x)v= lambda'(x)v$. That is, the eigenvalues of A' are the derivatives of the eigenvalues of A, with same associated eigenvectors.
$endgroup$
3
$begingroup$
The bias with what you say is that in general, the eigenvector depends on $x$.
$endgroup$
– mathcounterexamples.net
Jan 25 at 20:49
add a comment |
Your Answer
StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "69"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);
else
createEditor();
);
function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader:
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
,
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);
);
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3087557%2fis-there-an-explicit-relationship-between-the-eigenvalues-of-a-matrix-and-its-de%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
4 Answers
4
active
oldest
votes
4 Answers
4
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
It is not true in general that the eigenvalues of $B(x)$ are the derivatives of those of $A(x)$. And this even for some square matrices of dimension $2 times 2$.
Consider the matrix
$$A(x) =
beginpmatrix
1& -x^2\
-x &1
endpmatrix$$
It’s characteristic polynomial is $chi_A(x)(t)=t^2-2t+1-x^3$, which has for roots $1pm x ^3/2$ for $x>0$. Those are the eigenvalues of $A(x)$.
The derivative of $A(x)$ is
$$B(x) =
beginpmatrix
0& -2x\
-1 &0
endpmatrix$$ and it’s characteristic polynomial is $chi_B(x)(t)=t^2-2x$, whose roots are $pm sqrt2 x^1/2$for $x>0$.
We get a counterexample as the derivative of $1+x^3/2$ is not $sqrt2x^1/2$.
However in the special case of upper triangular matrices (that you consider in your original question) the eigenvalues of the matrix derivative are indeed the derivatives of the eigenvalues.
$endgroup$
add a comment |
$begingroup$
It is not true in general that the eigenvalues of $B(x)$ are the derivatives of those of $A(x)$. And this even for some square matrices of dimension $2 times 2$.
Consider the matrix
$$A(x) =
beginpmatrix
1& -x^2\
-x &1
endpmatrix$$
It’s characteristic polynomial is $chi_A(x)(t)=t^2-2t+1-x^3$, which has for roots $1pm x ^3/2$ for $x>0$. Those are the eigenvalues of $A(x)$.
The derivative of $A(x)$ is
$$B(x) =
beginpmatrix
0& -2x\
-1 &0
endpmatrix$$ and it’s characteristic polynomial is $chi_B(x)(t)=t^2-2x$, whose roots are $pm sqrt2 x^1/2$for $x>0$.
We get a counterexample as the derivative of $1+x^3/2$ is not $sqrt2x^1/2$.
However in the special case of upper triangular matrices (that you consider in your original question) the eigenvalues of the matrix derivative are indeed the derivatives of the eigenvalues.
$endgroup$
add a comment |
$begingroup$
It is not true in general that the eigenvalues of $B(x)$ are the derivatives of those of $A(x)$. And this even for some square matrices of dimension $2 times 2$.
Consider the matrix
$$A(x) =
beginpmatrix
1& -x^2\
-x &1
endpmatrix$$
It’s characteristic polynomial is $chi_A(x)(t)=t^2-2t+1-x^3$, which has for roots $1pm x ^3/2$ for $x>0$. Those are the eigenvalues of $A(x)$.
The derivative of $A(x)$ is
$$B(x) =
beginpmatrix
0& -2x\
-1 &0
endpmatrix$$ and it’s characteristic polynomial is $chi_B(x)(t)=t^2-2x$, whose roots are $pm sqrt2 x^1/2$for $x>0$.
We get a counterexample as the derivative of $1+x^3/2$ is not $sqrt2x^1/2$.
However in the special case of upper triangular matrices (that you consider in your original question) the eigenvalues of the matrix derivative are indeed the derivatives of the eigenvalues.
$endgroup$
It is not true in general that the eigenvalues of $B(x)$ are the derivatives of those of $A(x)$. And this even for some square matrices of dimension $2 times 2$.
Consider the matrix
$$A(x) =
beginpmatrix
1& -x^2\
-x &1
endpmatrix$$
It’s characteristic polynomial is $chi_A(x)(t)=t^2-2t+1-x^3$, which has for roots $1pm x ^3/2$ for $x>0$. Those are the eigenvalues of $A(x)$.
The derivative of $A(x)$ is
$$B(x) =
beginpmatrix
0& -2x\
-1 &0
endpmatrix$$ and it’s characteristic polynomial is $chi_B(x)(t)=t^2-2x$, whose roots are $pm sqrt2 x^1/2$for $x>0$.
We get a counterexample as the derivative of $1+x^3/2$ is not $sqrt2x^1/2$.
However in the special case of upper triangular matrices (that you consider in your original question) the eigenvalues of the matrix derivative are indeed the derivatives of the eigenvalues.
edited Jan 25 at 21:29
answered Jan 25 at 20:48
mathcounterexamples.netmathcounterexamples.net
26.9k22158
26.9k22158
add a comment |
add a comment |
$begingroup$
It can be shown that the eigenvalues of the derivative of the matrix cannot be derived from the eigenvalues of the original matrix. Example:
$$
A_1 = beginpmatrix -1 & 0 \ 0 & 1 endpmatrix
;;; , ;;;
A_2 = beginpmatrix 0 & e^x \ e^-x & 0 endpmatrix
$$
Both of the matrices above have the eigenvalues $-1$ and $1.$ However, the derivative of the first matrix has the eigenvalue $0$ (with multiplicity 2), while the derivative of the second matrix has the eigenvalues $i$ and $-i.$
Just given the eigenvalues $-1$ and $1$, there is no way of telling which matrix they originate from, hence no way of getting the eigenvalues of the derivative.
$endgroup$
add a comment |
$begingroup$
It can be shown that the eigenvalues of the derivative of the matrix cannot be derived from the eigenvalues of the original matrix. Example:
$$
A_1 = beginpmatrix -1 & 0 \ 0 & 1 endpmatrix
;;; , ;;;
A_2 = beginpmatrix 0 & e^x \ e^-x & 0 endpmatrix
$$
Both of the matrices above have the eigenvalues $-1$ and $1.$ However, the derivative of the first matrix has the eigenvalue $0$ (with multiplicity 2), while the derivative of the second matrix has the eigenvalues $i$ and $-i.$
Just given the eigenvalues $-1$ and $1$, there is no way of telling which matrix they originate from, hence no way of getting the eigenvalues of the derivative.
$endgroup$
add a comment |
$begingroup$
It can be shown that the eigenvalues of the derivative of the matrix cannot be derived from the eigenvalues of the original matrix. Example:
$$
A_1 = beginpmatrix -1 & 0 \ 0 & 1 endpmatrix
;;; , ;;;
A_2 = beginpmatrix 0 & e^x \ e^-x & 0 endpmatrix
$$
Both of the matrices above have the eigenvalues $-1$ and $1.$ However, the derivative of the first matrix has the eigenvalue $0$ (with multiplicity 2), while the derivative of the second matrix has the eigenvalues $i$ and $-i.$
Just given the eigenvalues $-1$ and $1$, there is no way of telling which matrix they originate from, hence no way of getting the eigenvalues of the derivative.
$endgroup$
It can be shown that the eigenvalues of the derivative of the matrix cannot be derived from the eigenvalues of the original matrix. Example:
$$
A_1 = beginpmatrix -1 & 0 \ 0 & 1 endpmatrix
;;; , ;;;
A_2 = beginpmatrix 0 & e^x \ e^-x & 0 endpmatrix
$$
Both of the matrices above have the eigenvalues $-1$ and $1.$ However, the derivative of the first matrix has the eigenvalue $0$ (with multiplicity 2), while the derivative of the second matrix has the eigenvalues $i$ and $-i.$
Just given the eigenvalues $-1$ and $1$, there is no way of telling which matrix they originate from, hence no way of getting the eigenvalues of the derivative.
answered Jan 26 at 14:00
Reinhard MeierReinhard Meier
3,022310
3,022310
add a comment |
add a comment |
$begingroup$
Let $alpha_k,beta_k$ be the eigenvalues of $(A,B)$
where $B(x) = A'(x).$
A class of matrices for which $beta_k=alpha_k',$ can be constructed as follows.
Choose an orthogonal matrix $Q$ and an upper triangular matrix $U(x).$
$$eqalign
A &= Q,U,Q^-1 cr
A' &= Q,U'Q^-1 cr
$$
Since $Q$ is orthogonal, the eigenvalues of $(U,U')$ equal the eigenvalues of $(A,A')$, respectively.
Since the eigenvalues of a triangular matrix are its diagonal elements,
the EVs of $U$ are $U_kk=alpha_k,$ and
the EVs of $U'$ are $U'_kk=alpha_k'=beta_k.$
NB: $,Q$ must be independent of $x$ for this construction to apply.
$endgroup$
add a comment |
$begingroup$
Let $alpha_k,beta_k$ be the eigenvalues of $(A,B)$
where $B(x) = A'(x).$
A class of matrices for which $beta_k=alpha_k',$ can be constructed as follows.
Choose an orthogonal matrix $Q$ and an upper triangular matrix $U(x).$
$$eqalign
A &= Q,U,Q^-1 cr
A' &= Q,U'Q^-1 cr
$$
Since $Q$ is orthogonal, the eigenvalues of $(U,U')$ equal the eigenvalues of $(A,A')$, respectively.
Since the eigenvalues of a triangular matrix are its diagonal elements,
the EVs of $U$ are $U_kk=alpha_k,$ and
the EVs of $U'$ are $U'_kk=alpha_k'=beta_k.$
NB: $,Q$ must be independent of $x$ for this construction to apply.
$endgroup$
add a comment |
$begingroup$
Let $alpha_k,beta_k$ be the eigenvalues of $(A,B)$
where $B(x) = A'(x).$
A class of matrices for which $beta_k=alpha_k',$ can be constructed as follows.
Choose an orthogonal matrix $Q$ and an upper triangular matrix $U(x).$
$$eqalign
A &= Q,U,Q^-1 cr
A' &= Q,U'Q^-1 cr
$$
Since $Q$ is orthogonal, the eigenvalues of $(U,U')$ equal the eigenvalues of $(A,A')$, respectively.
Since the eigenvalues of a triangular matrix are its diagonal elements,
the EVs of $U$ are $U_kk=alpha_k,$ and
the EVs of $U'$ are $U'_kk=alpha_k'=beta_k.$
NB: $,Q$ must be independent of $x$ for this construction to apply.
$endgroup$
Let $alpha_k,beta_k$ be the eigenvalues of $(A,B)$
where $B(x) = A'(x).$
A class of matrices for which $beta_k=alpha_k',$ can be constructed as follows.
Choose an orthogonal matrix $Q$ and an upper triangular matrix $U(x).$
$$eqalign
A &= Q,U,Q^-1 cr
A' &= Q,U'Q^-1 cr
$$
Since $Q$ is orthogonal, the eigenvalues of $(U,U')$ equal the eigenvalues of $(A,A')$, respectively.
Since the eigenvalues of a triangular matrix are its diagonal elements,
the EVs of $U$ are $U_kk=alpha_k,$ and
the EVs of $U'$ are $U'_kk=alpha_k'=beta_k.$
NB: $,Q$ must be independent of $x$ for this construction to apply.
edited Mar 31 at 20:47
answered Jan 30 at 18:56
greggreg
9,3361825
9,3361825
add a comment |
add a comment |
$begingroup$
If $A(x)v= lambda(x)v$, with v independent of x, then, differentiating on both sides of the equation by x, $A'(x)v= lambda'(x)v$. That is, the eigenvalues of A' are the derivatives of the eigenvalues of A, with same associated eigenvectors.
$endgroup$
3
$begingroup$
The bias with what you say is that in general, the eigenvector depends on $x$.
$endgroup$
– mathcounterexamples.net
Jan 25 at 20:49
add a comment |
$begingroup$
If $A(x)v= lambda(x)v$, with v independent of x, then, differentiating on both sides of the equation by x, $A'(x)v= lambda'(x)v$. That is, the eigenvalues of A' are the derivatives of the eigenvalues of A, with same associated eigenvectors.
$endgroup$
3
$begingroup$
The bias with what you say is that in general, the eigenvector depends on $x$.
$endgroup$
– mathcounterexamples.net
Jan 25 at 20:49
add a comment |
$begingroup$
If $A(x)v= lambda(x)v$, with v independent of x, then, differentiating on both sides of the equation by x, $A'(x)v= lambda'(x)v$. That is, the eigenvalues of A' are the derivatives of the eigenvalues of A, with same associated eigenvectors.
$endgroup$
If $A(x)v= lambda(x)v$, with v independent of x, then, differentiating on both sides of the equation by x, $A'(x)v= lambda'(x)v$. That is, the eigenvalues of A' are the derivatives of the eigenvalues of A, with same associated eigenvectors.
answered Jan 25 at 20:03
user247327user247327
11.6k1516
11.6k1516
3
$begingroup$
The bias with what you say is that in general, the eigenvector depends on $x$.
$endgroup$
– mathcounterexamples.net
Jan 25 at 20:49
add a comment |
3
$begingroup$
The bias with what you say is that in general, the eigenvector depends on $x$.
$endgroup$
– mathcounterexamples.net
Jan 25 at 20:49
3
3
$begingroup$
The bias with what you say is that in general, the eigenvector depends on $x$.
$endgroup$
– mathcounterexamples.net
Jan 25 at 20:49
$begingroup$
The bias with what you say is that in general, the eigenvector depends on $x$.
$endgroup$
– mathcounterexamples.net
Jan 25 at 20:49
add a comment |
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3087557%2fis-there-an-explicit-relationship-between-the-eigenvalues-of-a-matrix-and-its-de%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
1
$begingroup$
One nice property of an upper (or lower) triangular matrix is that its eigenvalues are the same as its diagonal elements; which explains the $2times 2$ examples that you discovered. But this is not a general property of all matrices.
$endgroup$
– greg
Jan 27 at 20:51