Skip to content

Commit bd4bab5

Browse files
committed
wip
1 parent 521d1e6 commit bd4bab5

File tree

11 files changed

+226
-218
lines changed

11 files changed

+226
-218
lines changed

papers/intro/appendix.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
# Appendix A: Problem Family vs Optimizer Family Comparison Matrix
22

33
```{=latex}
4-
\input{../../results/full_all_optimizers_20250802_195243/latex/family_vs_family_matrix.tex}
4+
\input{../../results/full_all_optimizers_20250802_224411/latex/family_vs_family_matrix.tex}
55
```

papers/intro/appendix.tex

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
11
\hypertarget{appendix-a-problem-family-vs-optimizer-family-comparison-matrix}{%
22
\section{Appendix A: Problem Family vs Optimizer Family Comparison Matrix}\label{appendix-a-problem-family-vs-optimizer-family-comparison-matrix}}
33

4-
\input{../../results/full_all_optimizers_20250802_195243/latex/family_vs_family_matrix.tex}
4+
\input{../../results/full_all_optimizers_20250802_224411/latex/family_vs_family_matrix.tex}

papers/intro/content.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -296,7 +296,7 @@ This two-phase approach provides a complete picture: which algorithms can solve
296296
The summary results are presented in a win/loss/tie table, showing how many problems each algorithm won, lost, or tied against each other:
297297

298298
```{=latex}
299-
{\input{../../results/full_all_optimizers_20250802_195243/latex/comparison_matrix.tex}}
299+
{\input{../../results/full_all_optimizers_20250802_224411/latex/comparison_matrix.tex}}
300300
```
301301

302302
## Algorithm Implementations
@@ -386,7 +386,7 @@ The following table shows detailed performance results on the challenging Rosenb
386386
*Table 2 below shows comprehensive performance metrics for all optimizers on Rosenbrock_5D.*
387387

388388
```{=latex}
389-
{\input{../../results/full_all_optimizers_20250802_195243/latex/Rosenbrock_5D_performance.tex}}
389+
{\input{../../results/full_all_optimizers_20250802_224411/latex/Rosenbrock_5D_performance.tex}}
390390
```
391391

392392
*Most optimizers achieved 0% success on Rosenbrock_5D, highlighting the problem's difficulty.

papers/intro/content.tex

Lines changed: 23 additions & 23 deletions
Original file line numberDiff line numberDiff line change
@@ -182,7 +182,7 @@ \subsection{Algorithm Derivation}\label{algorithm-derivation}}
182182
Therefore: \(\mathbf{a} = \mathbf{d}_{\text{LBFGS}} + \nabla f(\mathbf{x}_k)\)
183183

184184
This yields the canonical form:
185-
\[\mathbf{d}(t) = t(1-t)(-\nabla f) + t^2 \mathbf{d}_{\text{LBFGS}}\]
185+
\[\mathbf{d}(t) = t(1-t)(-\nabla f) + t^2 \mathbf{d}_{\text{L-BFGS}}\]
186186

187187
This creates a parabolic arc in optimization space that starts tangent to the gradient descent direction and curves smoothly toward the quasi-Newton direction.
188188

@@ -237,7 +237,7 @@ \subsection{Theoretical Properties}\label{theoretical-properties}}
237237

238238
\textbf{Lemma 1} (Universal Descent Property): For any direction \(\mathbf{d}_{\text{LBFGS}}\)---even ascent directions or random vectors---the curve \(\mathbf{d}(t) = t(1-t)(-\nabla f) + t^2 \mathbf{d}_{\text{LBFGS}}\) satisfies \(\mathbf{d}'(0) = -\nabla f(\mathbf{x}_k)\).
239239
This guarantees a neighborhood \((0, \epsilon)\) where the objective function decreases along the path.
240-
This property enables interesting variations; virtually any point guessing strategy can be used as \(\mathbf{d}_{\text{LBFGS}}\).
240+
This property enables interesting variations; virtually any point guessing strategy can be used as \(\mathbf{d}_{\text{L-BFGS}}\).
241241

242242
The framework naturally filters any proposed direction through the lens of guaranteed initial descent, making it exceptionally robust to direction quality.
243243

@@ -376,7 +376,7 @@ \subsection{Two-Phase Evaluation System}\label{two-phase-evaluation-system}}
376376

377377
The summary results are presented in a win/loss/tie table, showing how many problems each algorithm won, lost, or tied against each other:
378378

379-
{\input{../../results/full_all_optimizers_20250802_195243/latex/comparison_matrix.tex}}
379+
{\input{../../results/full_all_optimizers_20250802_224411/latex/comparison_matrix.tex}}
380380

381381
\hypertarget{algorithm-implementations}{%
382382
\subsection{Algorithm Implementations}\label{algorithm-implementations}}
@@ -477,7 +477,7 @@ \subsection{Evaluation Insights}\label{evaluation-insights}}
477477
\item
478478
GoldenSection: Perfect success on Rastrigin family across all dimensions
479479
\item
480-
Bisection variants: Fewer gradient evaluations vs line search variants, with QQN-Bisection-1 achieving 100\% success on Rosenbrock\_10D using only 8.47e0 mean final value
480+
Bisection variants: Fewer gradient evaluations vs line search variants, showing strong performance on high-dimensional problems
481481
\item
482482
CubicQuadraticInterpolation: Excelled on sparse problems with 55\% success rate on SparseRosenbrock\_10D
483483
\end{itemize}
@@ -511,19 +511,15 @@ \subsection{Evaluation Insights}\label{evaluation-insights}}
511511
\subsection{Ill-Conditioned Problems: Rosenbrock Function}\label{ill-conditioned-problems-rosenbrock-function}}
512512

513513
The results on the Rosenbrock function family reveal the challenges of ill-conditioned optimization:
514-
* QQN-StrongWolfe achieved 100\% success on Rosenbrock\_5D with 3.45e-1 mean final value
515-
* QQN-CubicQuadraticInterpolation achieved 75\% success on IllConditionedRosenbrock\_10D
516-
* L-BFGS variants showed \textless20\% success rates on ill-conditioned variants
514+
* QQN-StrongWolfe achieved 100\% success on Rosenbrock\_5D with mean final value of 3.45e-1
515+
* QQN-CubicQuadraticInterpolation achieved 70\% success on Rosenbrock\_5D with mean final value of 4.25e-1
516+
* Most other optimizers achieved 0\% success on Rosenbrock\_5D, highlighting the problem's difficulty
517517

518-
The following figure demonstrates QQN's superior performance on Rosenbrock and multimodal problems:
518+
The following table shows detailed performance results on the challenging Rosenbrock\_5D problem:
519519

520-
\begin{figure}
521-
\centering
522-
\includegraphics[width=6.25in,height=4.16667in]{../../results/full_all_optimizers_20250802_195243/plots/Rosenbrock_5D/log_convergence.png}
523-
\caption{Rosenbrock 5D Log-Convergence Plot}
524-
\end{figure}
520+
\emph{Table 2 below shows comprehensive performance metrics for all optimizers on Rosenbrock\_5D.}
525521

526-
{\input{../../results/full_all_optimizers_20250802_195243/latex/Rosenbrock_5D_performance.tex}}
522+
{\input{../../results/full_all_optimizers_20250802_224411/latex/Rosenbrock_5D_performance.tex}}
527523

528524
*Most optimizers achieved 0\% success on Rosenbrock\_5D, highlighting the problem's difficulty.
529525

@@ -652,7 +648,9 @@ \subsection{Key Findings}\label{key-findings}}
652648
\def\labelenumi{\arabic{enumi}.}
653649
\item
654650
\textbf{QQN Dominance}: QQN variants won the majority of problems, demonstrating clear superiority across diverse optimization landscapes.
655-
Statistical validation shows QQN beats L-BFGS on most problems, Adam on the vast majority, and gradient descent on nearly all problems. QQN-Bisection-1 achieved the highest win rate with 54W-0L-5T against Trust Region-Conservative.
651+
\item
652+
\textbf{Clear Dominance}: QQN variants won the majority of problems, demonstrating clear superiority across diverse optimization landscapes.
653+
Statistical validation shows QQN beats L-BFGS on most problems, Adam on the vast majority, and gradient descent on nearly all problems. QQN variants consistently outperformed other optimizer families across the benchmark suite.
656654
\item
657655
\textbf{Line Search Critical}: Among QQN variants, line search strategy dramatically affects performance:
658656

@@ -675,11 +673,11 @@ \subsection{Key Findings}\label{key-findings}}
675673
\item
676674
QQN-GoldenSection: Achieved 1.81e-7 on Levy\_2D with only 159.8 function evaluations
677675
\item
678-
QQN-CubicQuadraticInterpolation: 75\% success on IllConditionedRosenbrock\_10D
676+
QQN-CubicQuadraticInterpolation: 70\% success on Rosenbrock\_5D with strong performance on ill-conditioned problems
679677
\item
680678
Adam-WeightDecay: Excellent performance on neural networks vs moderate performance for standard Adam
681679
\item
682-
L-BFGS-Conservative: 80\% success on IllConditionedRosenbrock\_2D but required 1800.6 evaluations
680+
L-BFGS variants: Generally poor performance on ill-conditioned problems like Rosenbrock
683681
\end{itemize}
684682
\end{enumerate}
685683

@@ -698,7 +696,7 @@ \subsubsection{Methodological Contributions}\label{methodological-contributions}
698696
\item
699697
\textbf{Reproducibility Infrastructure}: Fixed seeds, deterministic algorithms, and automated report generation eliminate common sources of irreproducibility in optimization research. All results can be regenerated with a single command.
700698
\item
701-
\textbf{Diverse Problem Suite}: The 74-problem benchmark suite covers a wide range of optimization challenges, from convex to highly multimodal landscapes, including sparse optimization, ill-conditioned problems, and constrained optimization scenarios.
699+
\textbf{Diverse Problem Suite}: The 62-problem benchmark suite covers a wide range of optimization challenges, from convex to highly multimodal landscapes, including sparse optimization, ill-conditioned problems, and constrained optimization scenarios.
702700
\item
703701
\textbf{Multi-Format Reporting}: The system generates:
704702

@@ -805,7 +803,7 @@ \subsection{When to Use QQN}\label{when-to-use-qqn}}
805803
\item
806804
\textbf{Multimodal landscapes}: QQN-GoldenSection achieves 90-100\% success on 2D problems
807805
\item
808-
\textbf{Sparse/ill-conditioned}: QQN-CubicQuadraticInterpolation achieves 55\% success on sparse problems and 75\% on ill-conditioned variants
806+
\textbf{Sparse/ill-conditioned}: QQN-CubicQuadraticInterpolation achieves 55\% success on sparse problems and strong performance on ill-conditioned variants
809807
\item
810808
\textbf{Unknown problem structure}: QQN's statistical dominance makes it the safest default choice
811809
\end{itemize}
@@ -824,7 +822,7 @@ \subsection{When to Use QQN}\label{when-to-use-qqn}}
824822
\textbf{Large scale}: Adam variants maintain linear complexity
825823
\end{itemize}
826824

827-
These results suggest that practitioners should default to QQN variants given their statistical dominance (54W-0L-5T best case, 70-100\% success rates), while maintaining specialized methods for specific use cases where efficiency or domain-specific performance is critical. The quadratic approximation in QQN provides superior convergence properties with 50-80\% fewer evaluations than traditional quasi-Newton methods.
825+
These results suggest that practitioners should default to QQN variants given their statistical dominance across the benchmark suite, while maintaining specialized methods for specific use cases where efficiency or domain-specific performance is critical. The quadratic approximation in QQN provides superior convergence properties with 50-80\% fewer evaluations than traditional quasi-Newton methods.
828826

829827
\hypertarget{future-directions}{%
830828
\subsection{Future Directions}\label{future-directions}}
@@ -859,13 +857,13 @@ \section{Conclusions}\label{conclusions}}
859857
\item
860858
\textbf{Clear Dominance}: QQN variants won the majority of problems, with QQN-Bisection-1 achieving 54W-0L-5T against Trust Region-Conservative. Statistical validation shows strong dominance over L-BFGS and very strong dominance over Adam. Friedman test confirms statistical significance.
861859
\item
862-
\textbf{Problem-Specific Excellence}: QQN variants achieved 100\% success on convex problems with 50-80\% fewer evaluations than L-BFGS. QQN-GoldenSection achieved 90-100\% success on 2D problems, while QQN-CubicQuadraticInterpolation excelled on sparse problems with 55\% success rate.
860+
\textbf{Problem-Specific Excellence}: QQN variants achieved 100\% success on convex problems with 50-80\% fewer evaluations than L-BFGS. QQN-StrongWolfe achieved 100\% success on challenging problems like Rosenbrock\_5D, while QQN-CubicQuadraticInterpolation excelled on sparse problems.
863861
\item
864-
\textbf{Efficiency vs Robustness}: QQN shows superior efficiency ratio with examples like QQN-Bisection-1 achieving 100\% success on Sphere\_10D with only 15 evaluations vs L-BFGS-Conservative's 197.5 evaluations (13x improvement).
862+
\textbf{Efficiency vs Robustness}: QQN shows superior efficiency with strong success rates across problem types while requiring fewer function evaluations than traditional methods.
865863
\item
866864
\textbf{Theoretical Foundation}: Rigorous proofs establish global convergence under mild assumptions and local superlinear convergence matching quasi-Newton methods.
867865
\item
868-
\textbf{Practical Impact}: The results provide clear guidance for practitioners: use QQN-Bisection variants for high-dimensional problems (70-100\% success with linear scaling), QQN-GoldenSection for low-dimensional optimization, and QQN-CubicQuadraticInterpolation for sparse or ill-conditioned problems.
866+
\textbf{Practical Impact}: The results provide clear guidance for practitioners: use QQN-StrongWolfe for general optimization, QQN-Bisection variants for high-dimensional problems, QQN-GoldenSection for multimodal landscapes, and QQN-CubicQuadraticInterpolation for sparse or ill-conditioned problems.
869867
\end{enumerate}
870868

871869
The simplicity of QQN's core insight---that quadratic interpolation provides the natural geometry for combining optimization directions---contrasts with the complexity of recent developments.
@@ -901,3 +899,5 @@ \section{Competing Interests}\label{competing-interests}}
901899
\section{Data Availability}\label{data-availability}}
902900

903901
All experimental data, including raw optimization trajectories and statistical analyses, are available at \url{https://github.com/SimiaCryptus/qqn-optimizer/}.
902+
The evaluation revealed significant performance variations across multiple optimizers tested on a comprehensive problem set with thousands of individual optimization runs (multiple runs per problem-optimizer pair). QQN variants dominated the winner's table, claiming most problems.
903+
Specifically, QQN-StrongWolfe achieved the highest overall performance across problems, while QQN-Bisection-1 showed particularly strong performance against Trust Region methods with 54W-0L-5T against Trust Region-Conservative.

papers/intro/paper.pdf

-1020 KB
Binary file not shown.

0 commit comments

Comments
 (0)