trait Optimizer extends AnyRef
Linear Supertypes
Known Subclasses
Ordering
- Alphabetic
- By Inheritance
Inherited
- Optimizer
- AnyRef
- Any
- Hide All
- Show All
Visibility
- Public
- All
Abstract Value Members
- abstract def applyDense[T, I](gradient: Output[T], variable: variables.Variable[T], iteration: Option[variables.Variable[I]])(implicit arg0: core.types.TF[T], arg1: core.types.IsNotQuantized[T], arg2: core.types.TF[I], arg3: core.types.IsIntOrLong[I]): UntypedOp
- abstract def applySparse[T, I](gradient: OutputIndexedSlices[T], variable: variables.Variable[T], iteration: Option[variables.Variable[I]])(implicit arg0: core.types.TF[T], arg1: core.types.IsNotQuantized[T], arg2: core.types.TF[I], arg3: core.types.IsIntOrLong[I]): UntypedOp
- abstract val ignoreDuplicateSparseIndices: Boolean
- abstract val name: String
- abstract val useLocking: Boolean
Concrete Value Members
-
final
def
!=(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
final
def
##(): Int
- Definition Classes
- AnyRef → Any
-
final
def
==(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
def
applyGradients[T, I](gradientsAndVariables: Seq[(OutputLike[T], variables.Variable[Any])], iteration: Option[variables.Variable[I]] = None, name: String = this.name)(implicit arg0: core.types.TF[T], arg1: LongDefault[I], arg2: core.types.TF[I], arg3: core.types.IsIntOrLong[I]): UntypedOp
- Annotations
- @throws( ... )
- def applySparseDuplicateIndices[T, I](gradient: OutputIndexedSlices[T], variable: variables.Variable[T], iteration: Option[variables.Variable[I]])(implicit arg0: core.types.TF[T], arg1: core.types.IsNotQuantized[T], arg2: core.types.TF[I], arg3: core.types.IsIntOrLong[I]): UntypedOp
-
final
def
asInstanceOf[T0]: T0
- Definition Classes
- Any
-
def
clone(): AnyRef
- Attributes
- protected[java.lang]
- Definition Classes
- AnyRef
- Annotations
- @native() @throws( ... )
-
def
computeGradients[T](loss: Output[T], lossGradients: Seq[OutputLike[T]] = null, variables: Set[variables.Variable[Any]] = null, gradientsGatingMethod: GatingMethod = Gradients.OpGating, gradientsAggregationMethod: AggregationMethod = Gradients.AddAggregationMethod, colocateGradientsWithOps: Boolean = false)(implicit arg0: core.types.TF[T], arg1: core.types.IsFloatOrDouble[T]): Seq[(OutputLike[T], variables.Variable[Any])]
- Annotations
- @throws( ... )
- def createSlots(variables: Seq[variables.Variable[Any]]): Unit
-
final
def
eq(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
def
equals(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
def
finalize(): Unit
- Attributes
- protected[java.lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( classOf[java.lang.Throwable] )
- def finish(updateOps: Set[UntypedOp], nameScope: String): UntypedOp
-
final
def
getClass(): Class[_]
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
-
final
def
getNonSlotVariable[T](name: String, graph: core.Graph = null)(implicit arg0: core.types.TF[T]): variables.Variable[T]
- Attributes
- protected
-
final
def
getNonSlotVariables: Iterable[variables.Variable[Any]]
- Attributes
- protected
-
final
def
getOrCreateNonSlotVariable[T](name: String, initialValue: tensors.Tensor[T], colocationOps: Set[UntypedOp] = Set.empty, ignoreExisting: Boolean = false)(implicit arg0: core.types.TF[T]): variables.Variable[T]
- Attributes
- protected
-
final
def
getSlot[T, R](name: String, variable: variables.Variable[T])(implicit arg0: core.types.TF[T], arg1: core.types.TF[R]): variables.Variable[R]
- Attributes
- protected
-
final
def
getSlot[T, R](name: String, variable: variables.Variable[T], dataType: core.types.DataType[R], initializer: Initializer, shape: core.Shape, variableScope: String)(implicit arg0: core.types.TF[T], arg1: core.types.TF[R]): variables.Variable[R]
- Attributes
- protected
-
def
hashCode(): Int
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
-
final
def
isInstanceOf[T0]: Boolean
- Definition Classes
- Any
-
def
minimize[T, I](loss: Output[T], lossGradients: Seq[OutputLike[T]] = null, variables: Set[variables.Variable[Any]] = null, gradientsGatingMethod: GatingMethod = Gradients.OpGating, gradientsAggregationMethod: AggregationMethod = Gradients.AddAggregationMethod, colocateGradientsWithOps: Boolean = false, iteration: Option[variables.Variable[I]] = None, name: String = "Minimize")(implicit arg0: core.types.TF[T], arg1: core.types.IsFloatOrDouble[T], arg2: LongDefault[I], arg3: core.types.TF[I], arg4: core.types.IsIntOrLong[I]): UntypedOp
- Annotations
- @throws( ... )
-
final
def
ne(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
final
val
nonSlotVariables: Map[(String, Option[core.Graph]), variables.Variable[Any]]
- Attributes
- protected
-
final
def
notify(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
-
final
def
notifyAll(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
- def prepare[I](iteration: Option[variables.Variable[I]])(implicit arg0: core.types.TF[I], arg1: core.types.IsIntOrLong[I]): Unit
-
final
def
slotNames: Set[String]
- Attributes
- protected
-
final
val
slots: Map[String, Map[variables.Variable[Any], variables.Variable[Any]]]
- Attributes
- protected
- final def state: Seq[variables.Variable[Any]]
-
final
def
synchronized[T0](arg0: ⇒ T0): T0
- Definition Classes
- AnyRef
-
def
toString(): String
- Definition Classes
- AnyRef → Any
-
final
def
wait(): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long, arg1: Int): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long): Unit
- Definition Classes
- AnyRef
- Annotations
- @native() @throws( ... )
-
final
def
zerosSlot[T](name: String, variable: variables.Variable[T], variableScope: String)(implicit arg0: core.types.TF[T]): variables.Variable[T]
- Attributes
- protected