From 3063961c81b2fd09a39ceb6a8dbedf53d72d2823 Mon Sep 17 00:00:00 2001 From: feiyun0112 Date: Sat, 29 May 2021 19:56:12 +0800 Subject: [PATCH] Remove redundant jump --- src/Microsoft.ML.Core/Data/ModelSaving.cs | 1 - src/Microsoft.ML.Core/Utilities/TextReaderStream.cs | 1 - src/Microsoft.ML.Data/Data/Conversion.cs | 1 - src/Microsoft.ML.FastTree/FastTree.cs | 2 -- .../AcceleratedGradientDescent.cs | 7 +------ .../Training/Parallel/SingleTrainer.cs | 11 ----------- src/Microsoft.ML.TimeSeries/TimeSeriesUtils.cs | 4 ---- 7 files changed, 1 insertion(+), 26 deletions(-) diff --git a/src/Microsoft.ML.Core/Data/ModelSaving.cs b/src/Microsoft.ML.Core/Data/ModelSaving.cs index 6023f86bc0..51e0073079 100644 --- a/src/Microsoft.ML.Core/Data/ModelSaving.cs +++ b/src/Microsoft.ML.Core/Data/ModelSaving.cs @@ -51,7 +51,6 @@ internal static void SaveModel(RepositoryWriter rep, T value, string path) { sb.SaveAsBinary(writer); } - return; } } diff --git a/src/Microsoft.ML.Core/Utilities/TextReaderStream.cs b/src/Microsoft.ML.Core/Utilities/TextReaderStream.cs index 8a828b8f57..d647170943 100644 --- a/src/Microsoft.ML.Core/Utilities/TextReaderStream.cs +++ b/src/Microsoft.ML.Core/Utilities/TextReaderStream.cs @@ -95,7 +95,6 @@ protected override void Dispose(bool disposing) public override void Flush() { - return; } /// diff --git a/src/Microsoft.ML.Data/Data/Conversion.cs b/src/Microsoft.ML.Data/Data/Conversion.cs index 57b27c9184..a338735a0e 100644 --- a/src/Microsoft.ML.Data/Data/Conversion.cs +++ b/src/Microsoft.ML.Data/Data/Conversion.cs @@ -1365,7 +1365,6 @@ private void TryParseSigned(long max, in TX text, out long? result) result = (long)sVal; Contracts.Assert(0 <= result && result <= long.MaxValue); - return; } /// diff --git a/src/Microsoft.ML.FastTree/FastTree.cs b/src/Microsoft.ML.FastTree/FastTree.cs index b7f33239fe..6e43925e24 100644 --- a/src/Microsoft.ML.FastTree/FastTree.cs +++ b/src/Microsoft.ML.FastTree/FastTree.cs @@ -330,8 +330,6 @@ private protected void PrintTestGraph(IChannel ch) ch.Info(GetTestGraphHeader()); else ch.Info(GetTestGraphLine()); - - return; } private protected virtual void Initialize(IChannel ch) diff --git a/src/Microsoft.ML.FastTree/Training/OptimizationAlgorithms/AcceleratedGradientDescent.cs b/src/Microsoft.ML.FastTree/Training/OptimizationAlgorithms/AcceleratedGradientDescent.cs index 4c796dcc67..0db70a848b 100644 --- a/src/Microsoft.ML.FastTree/Training/OptimizationAlgorithms/AcceleratedGradientDescent.cs +++ b/src/Microsoft.ML.FastTree/Training/OptimizationAlgorithms/AcceleratedGradientDescent.cs @@ -54,12 +54,7 @@ internal override InternalRegressionTree TrainingIteration(IChannel ch, bool[] a internal override void UpdateScores(ScoreTracker t, InternalRegressionTree tree) { - if (t == TrainingScores) - { - return; - //Special optimized routine for updating TrainingScores is implemented as part of TrainingItearation - } - else + if (t != TrainingScores) base.UpdateScores(t, tree); } diff --git a/src/Microsoft.ML.FastTree/Training/Parallel/SingleTrainer.cs b/src/Microsoft.ML.FastTree/Training/Parallel/SingleTrainer.cs index 213e6722e7..644452ac6a 100644 --- a/src/Microsoft.ML.FastTree/Training/Parallel/SingleTrainer.cs +++ b/src/Microsoft.ML.FastTree/Training/Parallel/SingleTrainer.cs @@ -23,7 +23,6 @@ internal sealed class SingleTrainer : IParallelTraining { void IParallelTraining.CacheHistogram(bool isSmallerLeaf, int featureIdx, int subfeature, SufficientStatsBase sufficientStatsBase, bool hasWeights) { - return; } bool IParallelTraining.IsNeedFindLocalBestSplit() @@ -36,12 +35,10 @@ void IParallelTraining.FindGlobalBestSplit(LeafSplitCandidates smallerChildSplit FindBestThresholdFromRawArrayFun findFunction, SplitInfo[] bestSplits) { - return; } void IParallelTraining.GetGlobalDataCountInLeaf(int leafIdx, ref int cnt) { - return; } bool[] IParallelTraining.GetLocalBinConstructionFeatures(int numFeatures) @@ -61,42 +58,34 @@ double[] IParallelTraining.GlobalMean(Dataset dataset, InternalRegressionTree tr void IParallelTraining.PerformGlobalSplit(int leaf, int lteChild, int gtChild, SplitInfo splitInfo) { - return; } void IParallelTraining.InitIteration(ref bool[] activeFeatures) { - return; } void IParallelTraining.InitEnvironment() { - return; } void IParallelTraining.InitTreeLearner(Dataset trainData, int maxNumLeaves, int maxCatSplitPoints, ref int minDocInLeaf) { - return; } void IParallelTraining.SyncGlobalBoundary(int numFeatures, int maxBin, Double[][] binUpperBounds) { - return; } void IParallelTraining.FinalizeEnvironment() { - return; } void IParallelTraining.FinalizeTreeLearner() { - return; } void IParallelTraining.FinalizeIteration() { - return; } bool IParallelTraining.IsSkipNonSplittableHistogram() diff --git a/src/Microsoft.ML.TimeSeries/TimeSeriesUtils.cs b/src/Microsoft.ML.TimeSeries/TimeSeriesUtils.cs index 1ea7efe3c7..810840f38a 100644 --- a/src/Microsoft.ML.TimeSeries/TimeSeriesUtils.cs +++ b/src/Microsoft.ML.TimeSeries/TimeSeriesUtils.cs @@ -21,8 +21,6 @@ internal static void SerializeFixedSizeQueue(FixedSizeQueue queue, Binar writer.Write(queue.Count); for (int index = 0; index < queue.Count; index++) writer.Write(queue[index]); - - return; } internal static FixedSizeQueue DeserializeFixedSizeQueueSingle(BinaryReader reader, IHost host) @@ -52,8 +50,6 @@ internal static void SerializeFixedSizeQueue(FixedSizeQueue queue, Binar writer.Write(queue.Count); for (int index = 0; index < queue.Count; index++) writer.Write(queue[index]); - - return; } internal static FixedSizeQueue DeserializeFixedSizeQueueDouble(BinaryReader reader, IHost host)