How to show for any orthogonal vector set and any matrix $sum_u |Au|_2^2 leq |A|_F^2$?












3












$begingroup$


The following paper on page 16, in line 17
Online Principal Component Analysis
says for any orthogonal vector set and any matrix $sum_u |Au|_2^2 leq |A|_F^2$ is true. However, if we let $u_i$'s in $mathbb{R}^m$ be a set of orthogonal vectors and $A in mathbb{R}^{n times m}$ we have



$$sum_u |Au|_2^2=|AU|_F^2$$
where $U$ is the matrix that has $u_i$ as its columns and $|cdot|_F$ is Frobenius norm. Using Cauchy-Schwarz inequality



$$|AU|_F^2 leq |A|_F^2|U|_F^2$$



My question is how we can ignore $|U|_F^2$?










share|cite|improve this question









$endgroup$








  • 1




    $begingroup$
    $|AU|_F=|A|_F$ which follows from the trace property and $UU^*=I$.
    $endgroup$
    – A.Γ.
    Jan 3 at 15:43








  • 1




    $begingroup$
    But they are not orthanormal, it says they are orthogonal.
    $endgroup$
    – Saeed
    Jan 3 at 15:46










  • $begingroup$
    Some authors are sloppy about the distinction between "orthonormal" and "orthogonal". It's pretty straight-forward to normalize orthogonal vectors (orthogonalizing them using, say, Gram-Schmidt, is what takes all the effort!), so some authors don't trouble about it. Technically, you're correct, of course.
    $endgroup$
    – Adrian Keister
    Jan 3 at 15:56






  • 2




    $begingroup$
    It is sloppy written, but $U$ matrix they work with in this proof has orthonormal columns (Observation 14, page 14). Otherwise it is not true (take one vector and multiply it by a huge number).
    $endgroup$
    – A.Γ.
    Jan 3 at 16:46


















3












$begingroup$


The following paper on page 16, in line 17
Online Principal Component Analysis
says for any orthogonal vector set and any matrix $sum_u |Au|_2^2 leq |A|_F^2$ is true. However, if we let $u_i$'s in $mathbb{R}^m$ be a set of orthogonal vectors and $A in mathbb{R}^{n times m}$ we have



$$sum_u |Au|_2^2=|AU|_F^2$$
where $U$ is the matrix that has $u_i$ as its columns and $|cdot|_F$ is Frobenius norm. Using Cauchy-Schwarz inequality



$$|AU|_F^2 leq |A|_F^2|U|_F^2$$



My question is how we can ignore $|U|_F^2$?










share|cite|improve this question









$endgroup$








  • 1




    $begingroup$
    $|AU|_F=|A|_F$ which follows from the trace property and $UU^*=I$.
    $endgroup$
    – A.Γ.
    Jan 3 at 15:43








  • 1




    $begingroup$
    But they are not orthanormal, it says they are orthogonal.
    $endgroup$
    – Saeed
    Jan 3 at 15:46










  • $begingroup$
    Some authors are sloppy about the distinction between "orthonormal" and "orthogonal". It's pretty straight-forward to normalize orthogonal vectors (orthogonalizing them using, say, Gram-Schmidt, is what takes all the effort!), so some authors don't trouble about it. Technically, you're correct, of course.
    $endgroup$
    – Adrian Keister
    Jan 3 at 15:56






  • 2




    $begingroup$
    It is sloppy written, but $U$ matrix they work with in this proof has orthonormal columns (Observation 14, page 14). Otherwise it is not true (take one vector and multiply it by a huge number).
    $endgroup$
    – A.Γ.
    Jan 3 at 16:46
















3












3








3





$begingroup$


The following paper on page 16, in line 17
Online Principal Component Analysis
says for any orthogonal vector set and any matrix $sum_u |Au|_2^2 leq |A|_F^2$ is true. However, if we let $u_i$'s in $mathbb{R}^m$ be a set of orthogonal vectors and $A in mathbb{R}^{n times m}$ we have



$$sum_u |Au|_2^2=|AU|_F^2$$
where $U$ is the matrix that has $u_i$ as its columns and $|cdot|_F$ is Frobenius norm. Using Cauchy-Schwarz inequality



$$|AU|_F^2 leq |A|_F^2|U|_F^2$$



My question is how we can ignore $|U|_F^2$?










share|cite|improve this question









$endgroup$




The following paper on page 16, in line 17
Online Principal Component Analysis
says for any orthogonal vector set and any matrix $sum_u |Au|_2^2 leq |A|_F^2$ is true. However, if we let $u_i$'s in $mathbb{R}^m$ be a set of orthogonal vectors and $A in mathbb{R}^{n times m}$ we have



$$sum_u |Au|_2^2=|AU|_F^2$$
where $U$ is the matrix that has $u_i$ as its columns and $|cdot|_F$ is Frobenius norm. Using Cauchy-Schwarz inequality



$$|AU|_F^2 leq |A|_F^2|U|_F^2$$



My question is how we can ignore $|U|_F^2$?







linear-algebra cauchy-schwarz-inequality






share|cite|improve this question













share|cite|improve this question











share|cite|improve this question




share|cite|improve this question










asked Jan 3 at 15:39









SaeedSaeed

1,073310




1,073310








  • 1




    $begingroup$
    $|AU|_F=|A|_F$ which follows from the trace property and $UU^*=I$.
    $endgroup$
    – A.Γ.
    Jan 3 at 15:43








  • 1




    $begingroup$
    But they are not orthanormal, it says they are orthogonal.
    $endgroup$
    – Saeed
    Jan 3 at 15:46










  • $begingroup$
    Some authors are sloppy about the distinction between "orthonormal" and "orthogonal". It's pretty straight-forward to normalize orthogonal vectors (orthogonalizing them using, say, Gram-Schmidt, is what takes all the effort!), so some authors don't trouble about it. Technically, you're correct, of course.
    $endgroup$
    – Adrian Keister
    Jan 3 at 15:56






  • 2




    $begingroup$
    It is sloppy written, but $U$ matrix they work with in this proof has orthonormal columns (Observation 14, page 14). Otherwise it is not true (take one vector and multiply it by a huge number).
    $endgroup$
    – A.Γ.
    Jan 3 at 16:46
















  • 1




    $begingroup$
    $|AU|_F=|A|_F$ which follows from the trace property and $UU^*=I$.
    $endgroup$
    – A.Γ.
    Jan 3 at 15:43








  • 1




    $begingroup$
    But they are not orthanormal, it says they are orthogonal.
    $endgroup$
    – Saeed
    Jan 3 at 15:46










  • $begingroup$
    Some authors are sloppy about the distinction between "orthonormal" and "orthogonal". It's pretty straight-forward to normalize orthogonal vectors (orthogonalizing them using, say, Gram-Schmidt, is what takes all the effort!), so some authors don't trouble about it. Technically, you're correct, of course.
    $endgroup$
    – Adrian Keister
    Jan 3 at 15:56






  • 2




    $begingroup$
    It is sloppy written, but $U$ matrix they work with in this proof has orthonormal columns (Observation 14, page 14). Otherwise it is not true (take one vector and multiply it by a huge number).
    $endgroup$
    – A.Γ.
    Jan 3 at 16:46










1




1




$begingroup$
$|AU|_F=|A|_F$ which follows from the trace property and $UU^*=I$.
$endgroup$
– A.Γ.
Jan 3 at 15:43






$begingroup$
$|AU|_F=|A|_F$ which follows from the trace property and $UU^*=I$.
$endgroup$
– A.Γ.
Jan 3 at 15:43






1




1




$begingroup$
But they are not orthanormal, it says they are orthogonal.
$endgroup$
– Saeed
Jan 3 at 15:46




$begingroup$
But they are not orthanormal, it says they are orthogonal.
$endgroup$
– Saeed
Jan 3 at 15:46












$begingroup$
Some authors are sloppy about the distinction between "orthonormal" and "orthogonal". It's pretty straight-forward to normalize orthogonal vectors (orthogonalizing them using, say, Gram-Schmidt, is what takes all the effort!), so some authors don't trouble about it. Technically, you're correct, of course.
$endgroup$
– Adrian Keister
Jan 3 at 15:56




$begingroup$
Some authors are sloppy about the distinction between "orthonormal" and "orthogonal". It's pretty straight-forward to normalize orthogonal vectors (orthogonalizing them using, say, Gram-Schmidt, is what takes all the effort!), so some authors don't trouble about it. Technically, you're correct, of course.
$endgroup$
– Adrian Keister
Jan 3 at 15:56




2




2




$begingroup$
It is sloppy written, but $U$ matrix they work with in this proof has orthonormal columns (Observation 14, page 14). Otherwise it is not true (take one vector and multiply it by a huge number).
$endgroup$
– A.Γ.
Jan 3 at 16:46






$begingroup$
It is sloppy written, but $U$ matrix they work with in this proof has orthonormal columns (Observation 14, page 14). Otherwise it is not true (take one vector and multiply it by a huge number).
$endgroup$
– A.Γ.
Jan 3 at 16:46












1 Answer
1






active

oldest

votes


















1












$begingroup$

Note that $A^*A$ is a square matrix and $A^*A ge 0$ so there exists an orthonormal basis ${u_1, ldots, u_m}$ for $mathbb{R}^m$ such that $A^*A u_i = lambda_i u_i$ for some $lambda ge 0$.



We have
$$sum_{i=1}^m |Au_i|_2^2 = sum_{i=1}^m langle Au_i, Au_irangle = sum_{i=1}^m langle A^*Au_i, u_irangle = sum_{i=1}^m lambda_i =operatorname{Tr}(A^*A) = |A|_F^2$$



The interesting part is that the sum $sum_{i=1}^m |Au_i|_2^2$ is actually independent of the choice of the orthonormal basis ${u_1, ldots, u_m}$. Indeed, if ${v_1, ldots, v_m}$ is some other orthonormal basis for $mathbb{R}^m$, we have
begin{align}
sum_{i=1}^m |Au_i|_2^2 &= sum_{i=1}^m langle A^*Au_i, u_irangle\
&= sum_{i=1}^m leftlangle sum_{j=1}^mlangle u_i,v_jrangle A^*A v_j , sum_{k=1}^mlangle u_i,v_krangle v_krightrangle\
&= sum_{j=1}^m sum_{k=1}^m left(sum_{i=1}^mlangle u_i,v_jrangle langle v_k,u_irangleright)langle A^*A v_j,v_krangle\
&= sum_{j=1}^m sum_{k=1}^m langle v_j,v_kranglelangle A^*A v_j,v_krangle\
&= sum_{j=1}^m langle A^*A v_j,v_jrangle\
&= sum_{j=1}^m |Av_j|_2^2
end{align}



Hence we can extend any orthonormal set ${v_1, ldots, v_k} subseteq mathbb{R}^m$ to an orthonormal basis ${v_1, ldots, v_m}$ for $mathbb{R}^m$ to obtain
$$sum_{i=1}^k |Av_i|_2^2 lesum_{i=1}^m |Av_i|_2^2 = sum_{i=1}^m |Au_i|_2^2 = |A|_F^2 $$






share|cite|improve this answer









$endgroup$













    Your Answer





    StackExchange.ifUsing("editor", function () {
    return StackExchange.using("mathjaxEditing", function () {
    StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
    StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
    });
    });
    }, "mathjax-editing");

    StackExchange.ready(function() {
    var channelOptions = {
    tags: "".split(" "),
    id: "69"
    };
    initTagRenderer("".split(" "), "".split(" "), channelOptions);

    StackExchange.using("externalEditor", function() {
    // Have to fire editor after snippets, if snippets enabled
    if (StackExchange.settings.snippets.snippetsEnabled) {
    StackExchange.using("snippets", function() {
    createEditor();
    });
    }
    else {
    createEditor();
    }
    });

    function createEditor() {
    StackExchange.prepareEditor({
    heartbeatType: 'answer',
    autoActivateHeartbeat: false,
    convertImagesToLinks: true,
    noModals: true,
    showLowRepImageUploadWarning: true,
    reputationToPostImages: 10,
    bindNavPrevention: true,
    postfix: "",
    imageUploader: {
    brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
    contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
    allowUrls: true
    },
    noCode: true, onDemand: true,
    discardSelector: ".discard-answer"
    ,immediatelyShowMarkdownHelp:true
    });


    }
    });














    draft saved

    draft discarded


















    StackExchange.ready(
    function () {
    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3060689%2fhow-to-show-for-any-orthogonal-vector-set-and-any-matrix-sum-u-au-22-leq%23new-answer', 'question_page');
    }
    );

    Post as a guest















    Required, but never shown

























    1 Answer
    1






    active

    oldest

    votes








    1 Answer
    1






    active

    oldest

    votes









    active

    oldest

    votes






    active

    oldest

    votes









    1












    $begingroup$

    Note that $A^*A$ is a square matrix and $A^*A ge 0$ so there exists an orthonormal basis ${u_1, ldots, u_m}$ for $mathbb{R}^m$ such that $A^*A u_i = lambda_i u_i$ for some $lambda ge 0$.



    We have
    $$sum_{i=1}^m |Au_i|_2^2 = sum_{i=1}^m langle Au_i, Au_irangle = sum_{i=1}^m langle A^*Au_i, u_irangle = sum_{i=1}^m lambda_i =operatorname{Tr}(A^*A) = |A|_F^2$$



    The interesting part is that the sum $sum_{i=1}^m |Au_i|_2^2$ is actually independent of the choice of the orthonormal basis ${u_1, ldots, u_m}$. Indeed, if ${v_1, ldots, v_m}$ is some other orthonormal basis for $mathbb{R}^m$, we have
    begin{align}
    sum_{i=1}^m |Au_i|_2^2 &= sum_{i=1}^m langle A^*Au_i, u_irangle\
    &= sum_{i=1}^m leftlangle sum_{j=1}^mlangle u_i,v_jrangle A^*A v_j , sum_{k=1}^mlangle u_i,v_krangle v_krightrangle\
    &= sum_{j=1}^m sum_{k=1}^m left(sum_{i=1}^mlangle u_i,v_jrangle langle v_k,u_irangleright)langle A^*A v_j,v_krangle\
    &= sum_{j=1}^m sum_{k=1}^m langle v_j,v_kranglelangle A^*A v_j,v_krangle\
    &= sum_{j=1}^m langle A^*A v_j,v_jrangle\
    &= sum_{j=1}^m |Av_j|_2^2
    end{align}



    Hence we can extend any orthonormal set ${v_1, ldots, v_k} subseteq mathbb{R}^m$ to an orthonormal basis ${v_1, ldots, v_m}$ for $mathbb{R}^m$ to obtain
    $$sum_{i=1}^k |Av_i|_2^2 lesum_{i=1}^m |Av_i|_2^2 = sum_{i=1}^m |Au_i|_2^2 = |A|_F^2 $$






    share|cite|improve this answer









    $endgroup$


















      1












      $begingroup$

      Note that $A^*A$ is a square matrix and $A^*A ge 0$ so there exists an orthonormal basis ${u_1, ldots, u_m}$ for $mathbb{R}^m$ such that $A^*A u_i = lambda_i u_i$ for some $lambda ge 0$.



      We have
      $$sum_{i=1}^m |Au_i|_2^2 = sum_{i=1}^m langle Au_i, Au_irangle = sum_{i=1}^m langle A^*Au_i, u_irangle = sum_{i=1}^m lambda_i =operatorname{Tr}(A^*A) = |A|_F^2$$



      The interesting part is that the sum $sum_{i=1}^m |Au_i|_2^2$ is actually independent of the choice of the orthonormal basis ${u_1, ldots, u_m}$. Indeed, if ${v_1, ldots, v_m}$ is some other orthonormal basis for $mathbb{R}^m$, we have
      begin{align}
      sum_{i=1}^m |Au_i|_2^2 &= sum_{i=1}^m langle A^*Au_i, u_irangle\
      &= sum_{i=1}^m leftlangle sum_{j=1}^mlangle u_i,v_jrangle A^*A v_j , sum_{k=1}^mlangle u_i,v_krangle v_krightrangle\
      &= sum_{j=1}^m sum_{k=1}^m left(sum_{i=1}^mlangle u_i,v_jrangle langle v_k,u_irangleright)langle A^*A v_j,v_krangle\
      &= sum_{j=1}^m sum_{k=1}^m langle v_j,v_kranglelangle A^*A v_j,v_krangle\
      &= sum_{j=1}^m langle A^*A v_j,v_jrangle\
      &= sum_{j=1}^m |Av_j|_2^2
      end{align}



      Hence we can extend any orthonormal set ${v_1, ldots, v_k} subseteq mathbb{R}^m$ to an orthonormal basis ${v_1, ldots, v_m}$ for $mathbb{R}^m$ to obtain
      $$sum_{i=1}^k |Av_i|_2^2 lesum_{i=1}^m |Av_i|_2^2 = sum_{i=1}^m |Au_i|_2^2 = |A|_F^2 $$






      share|cite|improve this answer









      $endgroup$
















        1












        1








        1





        $begingroup$

        Note that $A^*A$ is a square matrix and $A^*A ge 0$ so there exists an orthonormal basis ${u_1, ldots, u_m}$ for $mathbb{R}^m$ such that $A^*A u_i = lambda_i u_i$ for some $lambda ge 0$.



        We have
        $$sum_{i=1}^m |Au_i|_2^2 = sum_{i=1}^m langle Au_i, Au_irangle = sum_{i=1}^m langle A^*Au_i, u_irangle = sum_{i=1}^m lambda_i =operatorname{Tr}(A^*A) = |A|_F^2$$



        The interesting part is that the sum $sum_{i=1}^m |Au_i|_2^2$ is actually independent of the choice of the orthonormal basis ${u_1, ldots, u_m}$. Indeed, if ${v_1, ldots, v_m}$ is some other orthonormal basis for $mathbb{R}^m$, we have
        begin{align}
        sum_{i=1}^m |Au_i|_2^2 &= sum_{i=1}^m langle A^*Au_i, u_irangle\
        &= sum_{i=1}^m leftlangle sum_{j=1}^mlangle u_i,v_jrangle A^*A v_j , sum_{k=1}^mlangle u_i,v_krangle v_krightrangle\
        &= sum_{j=1}^m sum_{k=1}^m left(sum_{i=1}^mlangle u_i,v_jrangle langle v_k,u_irangleright)langle A^*A v_j,v_krangle\
        &= sum_{j=1}^m sum_{k=1}^m langle v_j,v_kranglelangle A^*A v_j,v_krangle\
        &= sum_{j=1}^m langle A^*A v_j,v_jrangle\
        &= sum_{j=1}^m |Av_j|_2^2
        end{align}



        Hence we can extend any orthonormal set ${v_1, ldots, v_k} subseteq mathbb{R}^m$ to an orthonormal basis ${v_1, ldots, v_m}$ for $mathbb{R}^m$ to obtain
        $$sum_{i=1}^k |Av_i|_2^2 lesum_{i=1}^m |Av_i|_2^2 = sum_{i=1}^m |Au_i|_2^2 = |A|_F^2 $$






        share|cite|improve this answer









        $endgroup$



        Note that $A^*A$ is a square matrix and $A^*A ge 0$ so there exists an orthonormal basis ${u_1, ldots, u_m}$ for $mathbb{R}^m$ such that $A^*A u_i = lambda_i u_i$ for some $lambda ge 0$.



        We have
        $$sum_{i=1}^m |Au_i|_2^2 = sum_{i=1}^m langle Au_i, Au_irangle = sum_{i=1}^m langle A^*Au_i, u_irangle = sum_{i=1}^m lambda_i =operatorname{Tr}(A^*A) = |A|_F^2$$



        The interesting part is that the sum $sum_{i=1}^m |Au_i|_2^2$ is actually independent of the choice of the orthonormal basis ${u_1, ldots, u_m}$. Indeed, if ${v_1, ldots, v_m}$ is some other orthonormal basis for $mathbb{R}^m$, we have
        begin{align}
        sum_{i=1}^m |Au_i|_2^2 &= sum_{i=1}^m langle A^*Au_i, u_irangle\
        &= sum_{i=1}^m leftlangle sum_{j=1}^mlangle u_i,v_jrangle A^*A v_j , sum_{k=1}^mlangle u_i,v_krangle v_krightrangle\
        &= sum_{j=1}^m sum_{k=1}^m left(sum_{i=1}^mlangle u_i,v_jrangle langle v_k,u_irangleright)langle A^*A v_j,v_krangle\
        &= sum_{j=1}^m sum_{k=1}^m langle v_j,v_kranglelangle A^*A v_j,v_krangle\
        &= sum_{j=1}^m langle A^*A v_j,v_jrangle\
        &= sum_{j=1}^m |Av_j|_2^2
        end{align}



        Hence we can extend any orthonormal set ${v_1, ldots, v_k} subseteq mathbb{R}^m$ to an orthonormal basis ${v_1, ldots, v_m}$ for $mathbb{R}^m$ to obtain
        $$sum_{i=1}^k |Av_i|_2^2 lesum_{i=1}^m |Av_i|_2^2 = sum_{i=1}^m |Au_i|_2^2 = |A|_F^2 $$







        share|cite|improve this answer












        share|cite|improve this answer



        share|cite|improve this answer










        answered Jan 3 at 21:02









        mechanodroidmechanodroid

        28k62447




        28k62447






























            draft saved

            draft discarded




















































            Thanks for contributing an answer to Mathematics Stack Exchange!


            • Please be sure to answer the question. Provide details and share your research!

            But avoid



            • Asking for help, clarification, or responding to other answers.

            • Making statements based on opinion; back them up with references or personal experience.


            Use MathJax to format equations. MathJax reference.


            To learn more, see our tips on writing great answers.




            draft saved


            draft discarded














            StackExchange.ready(
            function () {
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3060689%2fhow-to-show-for-any-orthogonal-vector-set-and-any-matrix-sum-u-au-22-leq%23new-answer', 'question_page');
            }
            );

            Post as a guest















            Required, but never shown





















































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown

































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown







            Popular posts from this blog

            Quarter-circle Tiles

            build a pushdown automaton that recognizes the reverse language of a given pushdown automaton?

            Mont Emei