diff --git a/unused/deprecated/comment.tex b/unused/deprecated/comment.tex index bb6120a..a51ff15 100644 --- a/unused/deprecated/comment.tex +++ b/unused/deprecated/comment.tex @@ -2,4 +2,4 @@ It's difficult to write out. Essentially, there's two variables when collecting the data. The {\em start} which is a time marker and {\em count} which is a count of 2 week collection of records. The way how the code works is it adds both the start and seek together to identify where to begin collecting the data. After this point is set, the program steps forward in 2 week groupings to collect the number of {\em count} records to include. So the 2wk+30AVG results in 30 data points. The starting date is what makes this difficult as theres common language mixed in (1 year back; 3M; ...). I agree this is complicated, but I'm not sure there's an easier way to represent this without going too deep and distracting the reader. --Rob -I tried to explain a in a bit more detail. Hopefully this helps. Basically the whole data is split into 2 week groupings. The model steps through all of these groupings in sequence to determine predictions. It uses up to 1 year, $<=26$ data points, before the current 2 week grouping run on the model as training data and up to 4 years, $<=104$ data points, after the current 2 week grouping. So say the model is taking the first 2 weeks of January 2000, it would take $<=26$ 2 week groupings before that point as training data, so all of 1999 data would be 26 data points, and $<=104$ data points after that point as validation data,, so all of 2000-2003 data and first 2 weeks of January 2004 minus the first 2 weeks in January 2000. Then it repeats this whole sequence again for 2 week grouping (week 3 and 4 of January 2000) until all data points are run this way. The notation says how much training data is used but not how much validation data is used. --Tom} \ No newline at end of file +I tried to explain a in a bit more detail. Hopefully, this helps. Basically the whole data is split into 2 week groupings. The model steps through all of these groupings in sequence to determine predictions. It uses up to 1 year, $<=26$ data points, before the current 2 week grouping run on the model as training data and up to 4 years, $<=104$ data points, after the current 2 week grouping. So say the model is taking the first 2 weeks of January 2000, it would take $<=26$ 2 week groupings before that point as training data, so all of 1999 data would be 26 data points, and $<=104$ data points after that point as validation data,, so all of 2000-2003 data and first 2 weeks of January 2004 minus the first 2 weeks in January 2000. Then it repeats this whole sequence again for 2 week grouping (week 3 and 4 of January 2000) until all data points are run this way. The notation says how much training data is used but not how much validation data is used. --Tom} \ No newline at end of file diff --git a/unused/vonLaszewski-references.bib b/unused/vonLaszewski-references.bib index b1b64cd..92c5d44 100644 --- a/unused/vonLaszewski-references.bib +++ b/unused/vonLaszewski-references.bib @@ -1018,7 +1018,6 @@ @article{mlperf-training information sciences, FOS: Computer and information sciences}, title = {{MLPerf Training Benchmark}}, - publisher = {arXiv}, journal = {arXiv}, year = 2019, copyright = {arXiv.org perpetual, non-exclusive license} diff --git a/vonLaszewski-frontiers-citations.bib b/vonLaszewski-frontiers-citations.bib index 476b646..795f7a4 100644 --- a/vonLaszewski-frontiers-citations.bib +++ b/vonLaszewski-frontiers-citations.bib @@ -343,4 +343,3 @@ @inproceedings{green500 year = {2007}, doi = {10.1109/SC.2007.52}, } - diff --git a/vonLaszewski-frontiers.tex b/vonLaszewski-frontiers.tex index 8c67e41..0e4895a 100644 --- a/vonLaszewski-frontiers.tex +++ b/vonLaszewski-frontiers.tex @@ -390,7 +390,7 @@ \subsubsection{Earthquake Data} {\bf Objectives} & \multicolumn{2}{l|}{Improve the quality of Earthquake forecasting in a region of Southern California.}\\ \hline -{\bf Metrics} & \multicolumn{2}{l|}{Normalized Nash-Sutcliffe model efficiency coefficient (NNSE)with $0.8\leq NNSE\leq 0.99$}\\ +{\bf Metrics} & \multicolumn{2}{l|}{Normalized Nash-Sutcliffe model efficiency coefficient (NNSE) with $0.8\leq NNSE\leq 0.99$}\\ \hline {\bf Data} & Type: & Richter Measurements with spatial and temporal information (Events). \\ & Input: & Earthquakes since 1950.\\ @@ -651,7 +651,7 @@ \subsubsection{Workflow Compute Coordinator} Figure \ref{fig:cc-3}A shows the REST specification and \ref{fig:cc-3}B shows graphical user interface. -\subsubsection{Parameterized Workflow Job Generator} +\subsubsection{Parameterized Experiment Workflow Job Generator} \label{sec:workflow-sbatch} In traditional machine learning workflows, hyperparameter tuning and configuration are key elements in assessing and optimizing the performance of models. However, scaling hyperparameters for highly parallel execution with heterogeneous hardware is complex.