Skip to content

Commit

Permalink
Add @Name after bytedeco/javacpp#700
Browse files Browse the repository at this point in the history
  • Loading branch information
HGuillemet committed Sep 12, 2023
1 parent 9835913 commit 002b031
Show file tree
Hide file tree
Showing 324 changed files with 1,260 additions and 1,014 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -44,6 +44,9 @@ private native void allocate(
@Const @ByRef AdaptiveAvgPool1dOptions options_);
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveAvgPool1dImpl(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveAvgPool1dImpl(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::AdaptiveAvgPool1dImpl, torch::nn::Module>") void allocate(@SharedPtr Module pointer);


public native @ByVal Tensor forward(@Const @ByRef Tensor input);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -26,6 +26,9 @@ public class AdaptiveAvgPool1dImplBase extends AdaptiveAvgPool1dImplCloneable {
static { Loader.load(); }
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveAvgPool1dImplBase(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveAvgPool1dImplBase(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::AdaptiveAvgPoolImpl<1,torch::ExpandingArray<1>,torch::nn::AdaptiveAvgPool1dImpl>, torch::nn::Module>") void allocate(@SharedPtr Module pointer);

public AdaptiveAvgPool1dImplBase(@ByVal @Cast("torch::ExpandingArray<1>*") LongPointer output_size) { super((Pointer)null); allocate(output_size); }
private native void allocate(@ByVal @Cast("torch::ExpandingArray<1>*") LongPointer output_size);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,9 @@ public class AdaptiveAvgPool1dImplCloneable extends Module {
static { Loader.load(); }
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveAvgPool1dImplCloneable(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveAvgPool1dImplCloneable(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::Cloneable<torch::nn::AdaptiveAvgPool1dImpl>, torch::nn::Module>") void allocate(@SharedPtr Module pointer);
@Override public Module asModule() { return asModule(this); }
@Namespace public static native @SharedPtr @Name("SHARED_PTR_NAMESPACE::static_pointer_cast<torch::nn::Module, torch::nn::Cloneable<torch::nn::AdaptiveAvgPool1dImpl>>") Module asModule(@SharedPtr AdaptiveAvgPool1dImplCloneable pointer);

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -44,6 +44,9 @@ private native void allocate(
@Const @ByRef AdaptiveAvgPool2dOptions options_);
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveAvgPool2dImpl(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveAvgPool2dImpl(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::AdaptiveAvgPool2dImpl, torch::nn::Module>") void allocate(@SharedPtr Module pointer);


public native @ByVal Tensor forward(@Const @ByRef Tensor input);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,9 @@ public class AdaptiveAvgPool2dImplBase extends AdaptiveAvgPool2dImplCloneable {
static { Loader.load(); }
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveAvgPool2dImplBase(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveAvgPool2dImplBase(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::AdaptiveAvgPoolImpl<2,torch::ExpandingArrayWithOptionalElem<2>,torch::nn::AdaptiveAvgPool2dImpl>, torch::nn::Module>") void allocate(@SharedPtr Module pointer);

public AdaptiveAvgPool2dImplBase(@ByVal @Cast("torch::ExpandingArrayWithOptionalElem<2>*") LongOptional output_size) { super((Pointer)null); allocate(output_size); }
private native void allocate(@ByVal @Cast("torch::ExpandingArrayWithOptionalElem<2>*") LongOptional output_size);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,9 @@ public class AdaptiveAvgPool2dImplCloneable extends Module {
static { Loader.load(); }
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveAvgPool2dImplCloneable(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveAvgPool2dImplCloneable(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::Cloneable<torch::nn::AdaptiveAvgPool2dImpl>, torch::nn::Module>") void allocate(@SharedPtr Module pointer);
@Override public Module asModule() { return asModule(this); }
@Namespace public static native @SharedPtr @Name("SHARED_PTR_NAMESPACE::static_pointer_cast<torch::nn::Module, torch::nn::Cloneable<torch::nn::AdaptiveAvgPool2dImpl>>") Module asModule(@SharedPtr AdaptiveAvgPool2dImplCloneable pointer);

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -44,6 +44,9 @@ private native void allocate(
@Const @ByRef AdaptiveAvgPool3dOptions options_);
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveAvgPool3dImpl(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveAvgPool3dImpl(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::AdaptiveAvgPool3dImpl, torch::nn::Module>") void allocate(@SharedPtr Module pointer);


public native @ByVal Tensor forward(@Const @ByRef Tensor input);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,9 @@ public class AdaptiveAvgPool3dImplBase extends AdaptiveAvgPool3dImplCloneable {
static { Loader.load(); }
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveAvgPool3dImplBase(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveAvgPool3dImplBase(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::AdaptiveAvgPoolImpl<3,torch::ExpandingArrayWithOptionalElem<3>,torch::nn::AdaptiveAvgPool3dImpl>, torch::nn::Module>") void allocate(@SharedPtr Module pointer);

public AdaptiveAvgPool3dImplBase(@ByVal @Cast("torch::ExpandingArrayWithOptionalElem<3>*") LongOptional output_size) { super((Pointer)null); allocate(output_size); }
private native void allocate(@ByVal @Cast("torch::ExpandingArrayWithOptionalElem<3>*") LongOptional output_size);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,9 @@ public class AdaptiveAvgPool3dImplCloneable extends Module {
static { Loader.load(); }
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveAvgPool3dImplCloneable(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveAvgPool3dImplCloneable(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::Cloneable<torch::nn::AdaptiveAvgPool3dImpl>, torch::nn::Module>") void allocate(@SharedPtr Module pointer);
@Override public Module asModule() { return asModule(this); }
@Namespace public static native @SharedPtr @Name("SHARED_PTR_NAMESPACE::static_pointer_cast<torch::nn::Module, torch::nn::Cloneable<torch::nn::AdaptiveAvgPool3dImpl>>") Module asModule(@SharedPtr AdaptiveAvgPool3dImplCloneable pointer);

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -41,19 +41,22 @@ public class AdaptiveLogSoftmaxWithLossImpl extends AdaptiveLogSoftmaxWithLossIm
static { Loader.load(); }
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveLogSoftmaxWithLossImpl(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveLogSoftmaxWithLossImpl(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::AdaptiveLogSoftmaxWithLossImpl, torch::nn::Module>") void allocate(@SharedPtr Module pointer);

public AdaptiveLogSoftmaxWithLossImpl(
@Cast("int64_t") long in_features,
@Cast("int64_t") long n_classes,
@ByVal @Cast("std::vector<int64_t>*") LongVector cutoffs) { super((Pointer)null); allocate(in_features, n_classes, cutoffs); }
@SharedPtr private native void allocate(
@SharedPtr @Name("std::make_shared<torch::nn::AdaptiveLogSoftmaxWithLossImpl>") private native void allocate(
@Cast("int64_t") long in_features,
@Cast("int64_t") long n_classes,
@ByVal @Cast("std::vector<int64_t>*") LongVector cutoffs);

public AdaptiveLogSoftmaxWithLossImpl(
@ByVal AdaptiveLogSoftmaxWithLossOptions options_) { super((Pointer)null); allocate(options_); }
@SharedPtr private native void allocate(
@SharedPtr @Name("std::make_shared<torch::nn::AdaptiveLogSoftmaxWithLossImpl>") private native void allocate(
@ByVal AdaptiveLogSoftmaxWithLossOptions options_);

public native @ByVal ASMoutput forward(@Const @ByRef Tensor input, @Const @ByRef Tensor target);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,9 @@ public class AdaptiveLogSoftmaxWithLossImplCloneable extends Module {
static { Loader.load(); }
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveLogSoftmaxWithLossImplCloneable(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveLogSoftmaxWithLossImplCloneable(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::Cloneable<torch::nn::AdaptiveLogSoftmaxWithLossImpl>, torch::nn::Module>") void allocate(@SharedPtr Module pointer);
@Override public Module asModule() { return asModule(this); }
@Namespace public static native @SharedPtr @Name("SHARED_PTR_NAMESPACE::static_pointer_cast<torch::nn::Module, torch::nn::Cloneable<torch::nn::AdaptiveLogSoftmaxWithLossImpl>>") Module asModule(@SharedPtr AdaptiveLogSoftmaxWithLossImplCloneable pointer);

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -44,6 +44,9 @@ private native void allocate(
@Const @ByRef AdaptiveMaxPool1dOptions options_);
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveMaxPool1dImpl(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveMaxPool1dImpl(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::AdaptiveMaxPool1dImpl, torch::nn::Module>") void allocate(@SharedPtr Module pointer);


public native @ByVal Tensor forward(@Const @ByRef Tensor input);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -26,6 +26,9 @@ public class AdaptiveMaxPool1dImplBase extends AdaptiveMaxPool1dImplCloneable {
static { Loader.load(); }
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveMaxPool1dImplBase(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveMaxPool1dImplBase(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::AdaptiveMaxPoolImpl<1,torch::ExpandingArray<1>,torch::nn::AdaptiveMaxPool1dImpl>, torch::nn::Module>") void allocate(@SharedPtr Module pointer);

public AdaptiveMaxPool1dImplBase(@ByVal @Cast("torch::ExpandingArray<1>*") LongPointer output_size) { super((Pointer)null); allocate(output_size); }
private native void allocate(@ByVal @Cast("torch::ExpandingArray<1>*") LongPointer output_size);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,9 @@ public class AdaptiveMaxPool1dImplCloneable extends Module {
static { Loader.load(); }
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveMaxPool1dImplCloneable(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveMaxPool1dImplCloneable(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::Cloneable<torch::nn::AdaptiveMaxPool1dImpl>, torch::nn::Module>") void allocate(@SharedPtr Module pointer);
@Override public Module asModule() { return asModule(this); }
@Namespace public static native @SharedPtr @Name("SHARED_PTR_NAMESPACE::static_pointer_cast<torch::nn::Module, torch::nn::Cloneable<torch::nn::AdaptiveMaxPool1dImpl>>") Module asModule(@SharedPtr AdaptiveMaxPool1dImplCloneable pointer);

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -44,6 +44,9 @@ private native void allocate(
@Const @ByRef AdaptiveMaxPool2dOptions options_);
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveMaxPool2dImpl(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveMaxPool2dImpl(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::AdaptiveMaxPool2dImpl, torch::nn::Module>") void allocate(@SharedPtr Module pointer);


public native @ByVal Tensor forward(@Const @ByRef Tensor input);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,9 @@ public class AdaptiveMaxPool2dImplBase extends AdaptiveMaxPool2dImplCloneable {
static { Loader.load(); }
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveMaxPool2dImplBase(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveMaxPool2dImplBase(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::AdaptiveMaxPoolImpl<2,torch::ExpandingArrayWithOptionalElem<2>,torch::nn::AdaptiveMaxPool2dImpl>, torch::nn::Module>") void allocate(@SharedPtr Module pointer);

public AdaptiveMaxPool2dImplBase(@ByVal @Cast("torch::ExpandingArrayWithOptionalElem<2>*") LongOptional output_size) { super((Pointer)null); allocate(output_size); }
private native void allocate(@ByVal @Cast("torch::ExpandingArrayWithOptionalElem<2>*") LongOptional output_size);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,9 @@ public class AdaptiveMaxPool2dImplCloneable extends Module {
static { Loader.load(); }
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveMaxPool2dImplCloneable(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveMaxPool2dImplCloneable(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::Cloneable<torch::nn::AdaptiveMaxPool2dImpl>, torch::nn::Module>") void allocate(@SharedPtr Module pointer);
@Override public Module asModule() { return asModule(this); }
@Namespace public static native @SharedPtr @Name("SHARED_PTR_NAMESPACE::static_pointer_cast<torch::nn::Module, torch::nn::Cloneable<torch::nn::AdaptiveMaxPool2dImpl>>") Module asModule(@SharedPtr AdaptiveMaxPool2dImplCloneable pointer);

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -44,6 +44,9 @@ private native void allocate(
@Const @ByRef AdaptiveMaxPool3dOptions options_);
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveMaxPool3dImpl(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveMaxPool3dImpl(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::AdaptiveMaxPool3dImpl, torch::nn::Module>") void allocate(@SharedPtr Module pointer);


public native @ByVal Tensor forward(@Const @ByRef Tensor input);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,9 @@ public class AdaptiveMaxPool3dImplBase extends AdaptiveMaxPool3dImplCloneable {
static { Loader.load(); }
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveMaxPool3dImplBase(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveMaxPool3dImplBase(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::AdaptiveMaxPoolImpl<3,torch::ExpandingArrayWithOptionalElem<3>,torch::nn::AdaptiveMaxPool3dImpl>, torch::nn::Module>") void allocate(@SharedPtr Module pointer);

public AdaptiveMaxPool3dImplBase(@ByVal @Cast("torch::ExpandingArrayWithOptionalElem<3>*") LongOptional output_size) { super((Pointer)null); allocate(output_size); }
private native void allocate(@ByVal @Cast("torch::ExpandingArrayWithOptionalElem<3>*") LongOptional output_size);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,9 @@ public class AdaptiveMaxPool3dImplCloneable extends Module {
static { Loader.load(); }
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AdaptiveMaxPool3dImplCloneable(Pointer p) { super(p); }
/** Downcast constructor. */
public AdaptiveMaxPool3dImplCloneable(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::Cloneable<torch::nn::AdaptiveMaxPool3dImpl>, torch::nn::Module>") void allocate(@SharedPtr Module pointer);
@Override public Module asModule() { return asModule(this); }
@Namespace public static native @SharedPtr @Name("SHARED_PTR_NAMESPACE::static_pointer_cast<torch::nn::Module, torch::nn::Cloneable<torch::nn::AdaptiveMaxPool3dImpl>>") Module asModule(@SharedPtr AdaptiveMaxPool3dImplCloneable pointer);

Expand Down
12 changes: 3 additions & 9 deletions pytorch/src/gen/java/org/bytedeco/pytorch/AlphaDropoutImpl.java
Original file line number Diff line number Diff line change
Expand Up @@ -45,15 +45,9 @@ public class AlphaDropoutImpl extends AlphaDropoutImplBase {
private native void allocate();
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AlphaDropoutImpl(Pointer p) { super(p); }
/** Native array allocator. Access with {@link Pointer#position(long)}. */
public AlphaDropoutImpl(long size) { super((Pointer)null); allocateArray(size); }
private native void allocateArray(long size);
@Override public AlphaDropoutImpl position(long position) {
return (AlphaDropoutImpl)super.position(position);
}
@Override public AlphaDropoutImpl getPointer(long i) {
return new AlphaDropoutImpl((Pointer)this).offsetAddress(i);
}
/** Downcast constructor. */
public AlphaDropoutImpl(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::AlphaDropoutImpl, torch::nn::Module>") void allocate(@SharedPtr Module pointer);


public native @ByVal Tensor forward(@Const @ByRef Tensor input);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,9 @@ public class AlphaDropoutImplBase extends AlphaDropoutImplCloneable {
static { Loader.load(); }
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AlphaDropoutImplBase(Pointer p) { super(p); }
/** Downcast constructor. */
public AlphaDropoutImplBase(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::detail::_DropoutNd<torch::nn::AlphaDropoutImpl>, torch::nn::Module>") void allocate(@SharedPtr Module pointer);

public AlphaDropoutImplBase(double p) { super((Pointer)null); allocate(p); }
private native void allocate(double p);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,9 @@ public class AlphaDropoutImplCloneable extends Module {
static { Loader.load(); }
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public AlphaDropoutImplCloneable(Pointer p) { super(p); }
/** Downcast constructor. */
public AlphaDropoutImplCloneable(Module pointer) { super((Pointer)null); allocate(pointer); }
@Namespace private native @SharedPtr @Name("SHARED_PTR_NAMESPACE::dynamic_pointer_cast<torch::nn::Cloneable<torch::nn::AlphaDropoutImpl>, torch::nn::Module>") void allocate(@SharedPtr Module pointer);
@Override public Module asModule() { return asModule(this); }
@Namespace public static native @SharedPtr @Name("SHARED_PTR_NAMESPACE::static_pointer_cast<torch::nn::Module, torch::nn::Cloneable<torch::nn::AlphaDropoutImpl>>") Module asModule(@SharedPtr AlphaDropoutImplCloneable pointer);

Expand Down
Loading

0 comments on commit 002b031

Please sign in to comment.