Linearly constrained quadratic program












0












$begingroup$


I have the following quadratic program



$$min_x x^TAx qquad text{s.t} quad Ax in [a,b]^m$$



where matrix $A$ is positive semidefinite, and is similar both the objective function and in the constraint. I would like to solve this problem for large-scale input matrix $A$ and am looking for a swift solution.



Is there any specific recommendation from different optimization methods such as projection gradient, proximal gradient, trust region active set, so forth? Any comparison between these methods (and others) are appreciated.










share|cite|improve this question











$endgroup$












  • $begingroup$
    Since $Ax$ is a vector, what do you mean by $Ax in [a,b]$?
    $endgroup$
    – littleO
    Dec 16 '18 at 9:04






  • 1




    $begingroup$
    How large is $A$? Is it sparse or does it have any other structure?
    $endgroup$
    – littleO
    Dec 16 '18 at 9:16










  • $begingroup$
    That is right, Ax is vector! I edited the question.
    $endgroup$
    – Majid Mohammadi
    Dec 16 '18 at 9:59










  • $begingroup$
    The only thing we know is that A is positive semi definite! Nothing more is available
    $endgroup$
    – Majid Mohammadi
    Dec 16 '18 at 10:00






  • 1




    $begingroup$
    @LinAlg I don't have much experience with L-BFGS, is L-BFGS-B able to handle the hard constraint that $Ax in [a,b]^m$? I've read that it can handle box constraints like $x in [a,b]^n$, but $Ax in [a,b]^m$ is more difficult.
    $endgroup$
    – littleO
    Dec 16 '18 at 20:56


















0












$begingroup$


I have the following quadratic program



$$min_x x^TAx qquad text{s.t} quad Ax in [a,b]^m$$



where matrix $A$ is positive semidefinite, and is similar both the objective function and in the constraint. I would like to solve this problem for large-scale input matrix $A$ and am looking for a swift solution.



Is there any specific recommendation from different optimization methods such as projection gradient, proximal gradient, trust region active set, so forth? Any comparison between these methods (and others) are appreciated.










share|cite|improve this question











$endgroup$












  • $begingroup$
    Since $Ax$ is a vector, what do you mean by $Ax in [a,b]$?
    $endgroup$
    – littleO
    Dec 16 '18 at 9:04






  • 1




    $begingroup$
    How large is $A$? Is it sparse or does it have any other structure?
    $endgroup$
    – littleO
    Dec 16 '18 at 9:16










  • $begingroup$
    That is right, Ax is vector! I edited the question.
    $endgroup$
    – Majid Mohammadi
    Dec 16 '18 at 9:59










  • $begingroup$
    The only thing we know is that A is positive semi definite! Nothing more is available
    $endgroup$
    – Majid Mohammadi
    Dec 16 '18 at 10:00






  • 1




    $begingroup$
    @LinAlg I don't have much experience with L-BFGS, is L-BFGS-B able to handle the hard constraint that $Ax in [a,b]^m$? I've read that it can handle box constraints like $x in [a,b]^n$, but $Ax in [a,b]^m$ is more difficult.
    $endgroup$
    – littleO
    Dec 16 '18 at 20:56
















0












0








0


2



$begingroup$


I have the following quadratic program



$$min_x x^TAx qquad text{s.t} quad Ax in [a,b]^m$$



where matrix $A$ is positive semidefinite, and is similar both the objective function and in the constraint. I would like to solve this problem for large-scale input matrix $A$ and am looking for a swift solution.



Is there any specific recommendation from different optimization methods such as projection gradient, proximal gradient, trust region active set, so forth? Any comparison between these methods (and others) are appreciated.










share|cite|improve this question











$endgroup$




I have the following quadratic program



$$min_x x^TAx qquad text{s.t} quad Ax in [a,b]^m$$



where matrix $A$ is positive semidefinite, and is similar both the objective function and in the constraint. I would like to solve this problem for large-scale input matrix $A$ and am looking for a swift solution.



Is there any specific recommendation from different optimization methods such as projection gradient, proximal gradient, trust region active set, so forth? Any comparison between these methods (and others) are appreciated.







optimization convex-optimization numerical-optimization quadratic-programming






share|cite|improve this question















share|cite|improve this question













share|cite|improve this question




share|cite|improve this question








edited Jan 20 at 16:04









Rodrigo de Azevedo

13k41958




13k41958










asked Dec 16 '18 at 8:05









Majid MohammadiMajid Mohammadi

275




275












  • $begingroup$
    Since $Ax$ is a vector, what do you mean by $Ax in [a,b]$?
    $endgroup$
    – littleO
    Dec 16 '18 at 9:04






  • 1




    $begingroup$
    How large is $A$? Is it sparse or does it have any other structure?
    $endgroup$
    – littleO
    Dec 16 '18 at 9:16










  • $begingroup$
    That is right, Ax is vector! I edited the question.
    $endgroup$
    – Majid Mohammadi
    Dec 16 '18 at 9:59










  • $begingroup$
    The only thing we know is that A is positive semi definite! Nothing more is available
    $endgroup$
    – Majid Mohammadi
    Dec 16 '18 at 10:00






  • 1




    $begingroup$
    @LinAlg I don't have much experience with L-BFGS, is L-BFGS-B able to handle the hard constraint that $Ax in [a,b]^m$? I've read that it can handle box constraints like $x in [a,b]^n$, but $Ax in [a,b]^m$ is more difficult.
    $endgroup$
    – littleO
    Dec 16 '18 at 20:56




















  • $begingroup$
    Since $Ax$ is a vector, what do you mean by $Ax in [a,b]$?
    $endgroup$
    – littleO
    Dec 16 '18 at 9:04






  • 1




    $begingroup$
    How large is $A$? Is it sparse or does it have any other structure?
    $endgroup$
    – littleO
    Dec 16 '18 at 9:16










  • $begingroup$
    That is right, Ax is vector! I edited the question.
    $endgroup$
    – Majid Mohammadi
    Dec 16 '18 at 9:59










  • $begingroup$
    The only thing we know is that A is positive semi definite! Nothing more is available
    $endgroup$
    – Majid Mohammadi
    Dec 16 '18 at 10:00






  • 1




    $begingroup$
    @LinAlg I don't have much experience with L-BFGS, is L-BFGS-B able to handle the hard constraint that $Ax in [a,b]^m$? I've read that it can handle box constraints like $x in [a,b]^n$, but $Ax in [a,b]^m$ is more difficult.
    $endgroup$
    – littleO
    Dec 16 '18 at 20:56


















$begingroup$
Since $Ax$ is a vector, what do you mean by $Ax in [a,b]$?
$endgroup$
– littleO
Dec 16 '18 at 9:04




$begingroup$
Since $Ax$ is a vector, what do you mean by $Ax in [a,b]$?
$endgroup$
– littleO
Dec 16 '18 at 9:04




1




1




$begingroup$
How large is $A$? Is it sparse or does it have any other structure?
$endgroup$
– littleO
Dec 16 '18 at 9:16




$begingroup$
How large is $A$? Is it sparse or does it have any other structure?
$endgroup$
– littleO
Dec 16 '18 at 9:16












$begingroup$
That is right, Ax is vector! I edited the question.
$endgroup$
– Majid Mohammadi
Dec 16 '18 at 9:59




$begingroup$
That is right, Ax is vector! I edited the question.
$endgroup$
– Majid Mohammadi
Dec 16 '18 at 9:59












$begingroup$
The only thing we know is that A is positive semi definite! Nothing more is available
$endgroup$
– Majid Mohammadi
Dec 16 '18 at 10:00




$begingroup$
The only thing we know is that A is positive semi definite! Nothing more is available
$endgroup$
– Majid Mohammadi
Dec 16 '18 at 10:00




1




1




$begingroup$
@LinAlg I don't have much experience with L-BFGS, is L-BFGS-B able to handle the hard constraint that $Ax in [a,b]^m$? I've read that it can handle box constraints like $x in [a,b]^n$, but $Ax in [a,b]^m$ is more difficult.
$endgroup$
– littleO
Dec 16 '18 at 20:56






$begingroup$
@LinAlg I don't have much experience with L-BFGS, is L-BFGS-B able to handle the hard constraint that $Ax in [a,b]^m$? I've read that it can handle box constraints like $x in [a,b]^n$, but $Ax in [a,b]^m$ is more difficult.
$endgroup$
– littleO
Dec 16 '18 at 20:56












1 Answer
1






active

oldest

votes


















2












$begingroup$

One option which I think is promising is to formulate your problem as minimizing $f(x) + g(x) + h(Ax)$, where $f(x) = x^T A x$ and $g(x) = 0$ and $h(y) = I_{[a,b]}(y)$. (That is, $h$ is the convex indicator function of the set denoted here as $[a,b]$.) The function $f$ is differentiable and $g$ and $h$ have easy proximal operators, so you can solve this optimization problem using the PD3O method (which is a recent three-operator splitting method).



In this approach, we never have to solve a linear system involving $A$, which is nice because $A$ is large.






share|cite|improve this answer









$endgroup$













  • $begingroup$
    Thanks for response! I have seen some three operator method using different proximal gradient! Can you a bit explain the method?
    $endgroup$
    – Majid Mohammadi
    Dec 16 '18 at 10:01











Your Answer





StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");

StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);

StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});

function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});


}
});














draft saved

draft discarded


















StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3042354%2flinearly-constrained-quadratic-program%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown

























1 Answer
1






active

oldest

votes








1 Answer
1






active

oldest

votes









active

oldest

votes






active

oldest

votes









2












$begingroup$

One option which I think is promising is to formulate your problem as minimizing $f(x) + g(x) + h(Ax)$, where $f(x) = x^T A x$ and $g(x) = 0$ and $h(y) = I_{[a,b]}(y)$. (That is, $h$ is the convex indicator function of the set denoted here as $[a,b]$.) The function $f$ is differentiable and $g$ and $h$ have easy proximal operators, so you can solve this optimization problem using the PD3O method (which is a recent three-operator splitting method).



In this approach, we never have to solve a linear system involving $A$, which is nice because $A$ is large.






share|cite|improve this answer









$endgroup$













  • $begingroup$
    Thanks for response! I have seen some three operator method using different proximal gradient! Can you a bit explain the method?
    $endgroup$
    – Majid Mohammadi
    Dec 16 '18 at 10:01
















2












$begingroup$

One option which I think is promising is to formulate your problem as minimizing $f(x) + g(x) + h(Ax)$, where $f(x) = x^T A x$ and $g(x) = 0$ and $h(y) = I_{[a,b]}(y)$. (That is, $h$ is the convex indicator function of the set denoted here as $[a,b]$.) The function $f$ is differentiable and $g$ and $h$ have easy proximal operators, so you can solve this optimization problem using the PD3O method (which is a recent three-operator splitting method).



In this approach, we never have to solve a linear system involving $A$, which is nice because $A$ is large.






share|cite|improve this answer









$endgroup$













  • $begingroup$
    Thanks for response! I have seen some three operator method using different proximal gradient! Can you a bit explain the method?
    $endgroup$
    – Majid Mohammadi
    Dec 16 '18 at 10:01














2












2








2





$begingroup$

One option which I think is promising is to formulate your problem as minimizing $f(x) + g(x) + h(Ax)$, where $f(x) = x^T A x$ and $g(x) = 0$ and $h(y) = I_{[a,b]}(y)$. (That is, $h$ is the convex indicator function of the set denoted here as $[a,b]$.) The function $f$ is differentiable and $g$ and $h$ have easy proximal operators, so you can solve this optimization problem using the PD3O method (which is a recent three-operator splitting method).



In this approach, we never have to solve a linear system involving $A$, which is nice because $A$ is large.






share|cite|improve this answer









$endgroup$



One option which I think is promising is to formulate your problem as minimizing $f(x) + g(x) + h(Ax)$, where $f(x) = x^T A x$ and $g(x) = 0$ and $h(y) = I_{[a,b]}(y)$. (That is, $h$ is the convex indicator function of the set denoted here as $[a,b]$.) The function $f$ is differentiable and $g$ and $h$ have easy proximal operators, so you can solve this optimization problem using the PD3O method (which is a recent three-operator splitting method).



In this approach, we never have to solve a linear system involving $A$, which is nice because $A$ is large.







share|cite|improve this answer












share|cite|improve this answer



share|cite|improve this answer










answered Dec 16 '18 at 9:12









littleOlittleO

29.8k646109




29.8k646109












  • $begingroup$
    Thanks for response! I have seen some three operator method using different proximal gradient! Can you a bit explain the method?
    $endgroup$
    – Majid Mohammadi
    Dec 16 '18 at 10:01


















  • $begingroup$
    Thanks for response! I have seen some three operator method using different proximal gradient! Can you a bit explain the method?
    $endgroup$
    – Majid Mohammadi
    Dec 16 '18 at 10:01
















$begingroup$
Thanks for response! I have seen some three operator method using different proximal gradient! Can you a bit explain the method?
$endgroup$
– Majid Mohammadi
Dec 16 '18 at 10:01




$begingroup$
Thanks for response! I have seen some three operator method using different proximal gradient! Can you a bit explain the method?
$endgroup$
– Majid Mohammadi
Dec 16 '18 at 10:01


















draft saved

draft discarded




















































Thanks for contributing an answer to Mathematics Stack Exchange!


  • Please be sure to answer the question. Provide details and share your research!

But avoid



  • Asking for help, clarification, or responding to other answers.

  • Making statements based on opinion; back them up with references or personal experience.


Use MathJax to format equations. MathJax reference.


To learn more, see our tips on writing great answers.




draft saved


draft discarded














StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3042354%2flinearly-constrained-quadratic-program%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown





















































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown

































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown







Popular posts from this blog

Quarter-circle Tiles

build a pushdown automaton that recognizes the reverse language of a given pushdown automaton?

Mont Emei