Skip to content

Add Regularizers 1 #216

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 29 commits into from
May 2, 2021
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
29 commits
Select commit Hold shift + click to select a range
c57a2e7
Merge pull request #3 from tensorflow/master
JimClarke5 Oct 8, 2020
09fc07e
Merge pull request #4 from tensorflow/master
JimClarke5 Oct 27, 2020
a99dcb4
Merge pull request #5 from tensorflow/master
JimClarke5 Nov 17, 2020
ba294ea
Merge pull request #6 from tensorflow/master
JimClarke5 Nov 19, 2020
04f419a
Merge pull request #7 from tensorflow/master
JimClarke5 Dec 30, 2020
02e7ebf
Merge pull request #8 from tensorflow/master
JimClarke5 Jan 29, 2021
e0c9ed8
Merge pull request #9 from tensorflow/master
JimClarke5 Feb 1, 2021
5b0374b
Merge pull request #10 from tensorflow/master
JimClarke5 Feb 11, 2021
ccc7820
Initial Checkin
JimClarke5 Jan 15, 2021
05ec6e8
Clean up JavaDoc
JimClarke5 Jan 26, 2021
b446618
Fix to match the lates version of losses.Loss
JimClarke5 Feb 13, 2021
b5c7c78
Updates based on comments from PR.
JimClarke5 Feb 13, 2021
a3ccf61
Add JavDoc to new method l1_l2
JimClarke5 Feb 14, 2021
8c79214
change l1_l2 to create.
JimClarke5 Feb 14, 2021
1af4552
delete class L1_L2
JimClarke5 Feb 14, 2021
e038bbd
Merge pull request #11 from tensorflow/master
JimClarke5 Feb 23, 2021
def3051
Merge pull request #13 from tensorflow/master
JimClarke5 Mar 3, 2021
11748ae
Merge pull request #15 from tensorflow/master
JimClarke5 Mar 21, 2021
a9412ea
Merge pull request #16 from tensorflow/master
JimClarke5 Apr 9, 2021
2ff8dfe
Merge pull request #17 from tensorflow/master
JimClarke5 Apr 22, 2021
ee5e38a
Merge pull request #18 from tensorflow/master
JimClarke5 May 1, 2021
54f1802
Rebase with tensorflow Master
JimClarke5 May 1, 2021
bbd3bc3
Updating fixed local copy to repair broken remote copy
JimClarke5 May 2, 2021
6c48131
Clean up JavaDoc
JimClarke5 Jan 26, 2021
3c45a87
Updates based on comments from PR.
JimClarke5 Feb 13, 2021
2bd80b3
Add JavDoc to new method l1_l2
JimClarke5 Feb 14, 2021
da7a10b
change l1_l2 to create.
JimClarke5 Feb 14, 2021
9ea1d9a
delete class L1_L2
JimClarke5 Feb 14, 2021
1a93bdc
Merge remote-tracking branch 'origin/Regularizers_1' into Regularizers_1
JimClarke5 May 2, 2021
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
@@ -0,0 +1,47 @@
/* Copyright 2020 The TensorFlow Authors. All Rights Reserved.

Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at

http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
=======================================================================*/
package org.tensorflow.framework.regularizers;

import org.tensorflow.op.Ops;

/**
* A regularizer that applies an L1 or Lasso(least absolute shrinkage and selection operator)
* Regression, regularization penalty.
*
* <p>The L1 regularization penalty is computed as: <code>loss = l1 * reduceSum(abs(x))</code>
*/
public class L1 extends L1L2 {

/**
* Create a regularizer that applies an L1 regularization penalty of {@link
* #DEFAULT_REGULARIZATION_PENALTY}
*
* @param tf the TensorFlow Ops
*/
public L1(Ops tf) {
this(tf, DEFAULT_REGULARIZATION_PENALTY);
}

/**
* Create a regularizer that applies an L1 regularization penalty
*
* @param tf the TensorFlow Ops
* @param l1 the L1 regularization penalty
* @throws IllegalArgumentException if the l1 regularization factor is NaN or is infinite.
*/
public L1(Ops tf, float l1) {
super(tf, l1, 0f);
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,120 @@
/* Copyright 2020 The TensorFlow Authors. All Rights Reserved.

Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at

http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
=======================================================================*/
package org.tensorflow.framework.regularizers;

import org.tensorflow.Operand;
import org.tensorflow.framework.losses.impl.LossesHelper;
import org.tensorflow.op.Ops;
import org.tensorflow.types.family.TNumber;

/**
* A regularizer that applies both L1 and L2 regularization penalties.
*
* <p>The L1 regularization penalty is computed as:
*
* <pre>loss = l1 * reduceSum(abs(x))</pre>
*
* <p>The L2 regularization penalty is computed as
*
* <pre>loss = l2 * reduceSum(square(x))</pre>
*
*/
public class L1L2 extends Regularizer {

private final float l1;
private final float l2;

/**
* Creates an L1L2 regularizer with no l1 or l2 penalty with zero penalty
*
* @param tf the TensorFlow Ops
*/
public L1L2(Ops tf) {
this(tf, DEFAULT_REGULARIZATION_PENALTY, DEFAULT_REGULARIZATION_PENALTY);
}

/**
* Creates an L1L2 regularizer
*
* @param tf the TensorFlow Ops
* @param l1 L1 regularization factor, if null it is set to 0.
* @param l2 L2 regularization factor, if null it is set to 0.
* @throws IllegalArgumentException if the l1 or l2 regularization factor is {@link Float#isNaN}
* of {@link Float#isInfinite}
*/
public L1L2(Ops tf, float l1, float l2) {
super(tf);
if (Float.isNaN(l1) || Float.isInfinite(l1)) {
throw new IllegalArgumentException(
String.format(
"L1 Value: %f is not a valid regularization penalty number, a positive/negative infinity or NaN is not a property value",
l1));
}
this.l1 = l1;

if (Float.isNaN(l2) || Float.isInfinite(l2)) {
throw new IllegalArgumentException(
String.format(
"L2 Value: %f is not a valid regularization penalty number, a positive/negative infinity or NaN is not a property value",
l2));
}
this.l2 = l2;
}


/** {@inheritDoc} */
@Override
public <R extends TNumber> Operand<R> call(Operand<R> input) {
Ops tf = getTF();
if (this.getL1() == 0f && this.getL2() == 0f) {
return tf.dtypes.cast(tf.constant(0), input.type());
}
Operand<R> regularization = tf.dtypes.cast(tf.constant(0), input.type());

if (this.getL1() != 0.f) {
Operand<R> l1Op = tf.dtypes.cast(tf.constant(this.getL1()), input.type());
Operand<R> abs = tf.math.abs(input);
Operand<R> reduceSum = tf.reduceSum(abs, LossesHelper.allAxes(tf, input));
regularization = tf.math.add(regularization, tf.math.mul(l1Op, reduceSum));
}

if (this.getL2() != 0.f) {
Operand<R> l2Op = tf.dtypes.cast(tf.constant(this.getL2()), input.type());
Operand<R> sqr = tf.math.square(input);
Operand<R> reduceSum = tf.reduceSum(sqr, LossesHelper.allAxes(tf, input));
regularization = tf.math.add(regularization, tf.math.mul(l2Op, reduceSum));
}

return regularization;
}

/**
* Gets the L1 regularization factor
*
* @return the L1 regularization factor
*/
public float getL1() {
return l1;
}

/**
* Gets the L2 regularization factor
*
* @return the L2 regularization factor
*/
public float getL2() {
return l2;
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,46 @@
/* Copyright 2020 The TensorFlow Authors. All Rights Reserved.

Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at

http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
=======================================================================*/
package org.tensorflow.framework.regularizers;

import org.tensorflow.op.Ops;

/**
* A regularizer that applies a L2 (Ridge Regression) regularization penalty.
*
* <p>The L2 regularization penalty is computed as: <code>loss = l2 * reduceSum(square(x))</code>
*/
public class L2 extends L1L2 {

/**
* Create a regularizer that applies an L2 regularization penalty of {@link
* #DEFAULT_REGULARIZATION_PENALTY}
*
* @param tf the TensorFlow Ops
*/
public L2(Ops tf) {
this(tf, DEFAULT_REGULARIZATION_PENALTY);
}

/**
* Create a regularizer that applies an L1 regularization penalty
*
* @param tf the TensorFlow Ops
* @param l2 the L2 regularization penalty
* @throws IllegalArgumentException if the l2 regularization factor is NaN or is infinite.
*/
public L2(Ops tf, float l2) {
super(tf, 0f, l2);
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,91 @@
/* Copyright 2020 The TensorFlow Authors. All Rights Reserved.

Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at

http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
=======================================================================*/
package org.tensorflow.framework.regularizers;

import org.tensorflow.Operand;
import org.tensorflow.framework.losses.Loss;
import org.tensorflow.op.Ops;
import org.tensorflow.types.family.TNumber;

/**
* Base class for Regularizers
*
* <p>Regularizers allow you to apply penalties on layer parameters or layer activity during
* optimization. These penalties are summed into the loss function that the network optimizes.
*/
public abstract class Regularizer {

public static final float DEFAULT_REGULARIZATION_PENALTY = 0.01f;

private final Ops tf;
private final String name;

/**
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

To me, this is one of those header docs that should be either elaborated or omitted.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

👍 done

* Creates a Regularizer, using {@link Class#getSimpleName()} for the name
*
* @param tf the TensorFlow ops.
*/
protected Regularizer(Ops tf) {
this(tf, null);
}
/**
* Creates a Regularizer
*
* @param tf the TensorFlow ops.
* @param name the name of this regularizer, if null use {@link Class#getSimpleName()} for the
* name.
*/
protected Regularizer(Ops tf, String name) {
this.tf = tf;
this.name = name == null ? this.getClass().getSimpleName() : name;
}

/**
* Returns this Regularizer as a Loss This is a convenience to use regularize a loss. Only
* sampleWeights are applied to the regularizer.
*
* @return this Regularizer as a Loss
*/
public Loss asLoss() {
return new RegularizerLoss(this.tf, this);
}

/**
* Computes a regularization penalty from an input.
*
* @param input the weighted input
* @return the result of computing the regularization penalty
* @param <R> the data type of the input and result
*/
public abstract <R extends TNumber> Operand<R> call(Operand<R> input);

/**
* Gets the TensorFlow Ops
*
* @return the TensorFlow Ops
*/
public Ops getTF() {
return tf;
}

/**
* Gets the name for this regularizer
*
* @return the name for this regularizer
*/
public String getName() {
return name;
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,64 @@
/* Copyright 2020 The TensorFlow Authors. All Rights Reserved.

Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at

http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
=======================================================================*/
package org.tensorflow.framework.regularizers;

import org.tensorflow.Operand;
import org.tensorflow.framework.losses.Loss;
import org.tensorflow.op.Ops;
import org.tensorflow.types.family.TNumber;

/**
* A Regularizer call wrapped as a Loss instance
*
* <p>This class facilitates using a regularizer as a loss, only <code>sampleWeights</code> are
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I can't find the idea of regularizing sample weights on the web anywhere. Just double-checking that it's what's intended here.

* regularized.
*/
class RegularizerLoss extends Loss {

private final Regularizer regularizer;

/**
* Creates a Loss using {@link Class#getSimpleName()} as the name and a Loss Reduction of {@link
* Loss#REDUCTION_DEFAULT}
*
* @param tf the TensorFlow Ops
* @param regularizer the regularizer used to calculate the loss
*/
public RegularizerLoss(Ops tf, Regularizer regularizer) {
this(tf, null, regularizer);
}

/**
* Creates a Loss using a Loss Reduction of {@link Loss#REDUCTION_DEFAULT}
*
* @param tf the TensorFlow Ops
* @param name the name of this Loss, if null the name will be {@link Class#getSimpleName()}.
* @param regularizer the regularizer used to calculate the loss
*/
public RegularizerLoss(Ops tf, String name, Regularizer regularizer) {
super(tf, name);
this.regularizer = regularizer;
}

/** {@inheritDoc} */
@Override
public <T extends TNumber> Operand<T> call(
Operand<? extends TNumber> labels, Operand<T> predictions, Operand<T> sampleWeights) {
if (sampleWeights == null) {
throw new IllegalArgumentException("sampleWeights cannot be null");
}
return regularizer.call(sampleWeights);
}
}
Loading