gotch/example/mnist/nn.go

67 lines
1.4 KiB
Go
Raw Normal View History

2020-06-18 08:14:48 +01:00
package main
import (
"fmt"
"log"
"github.com/sugarme/gotch"
"github.com/sugarme/gotch/nn"
ts "github.com/sugarme/gotch/tensor"
"github.com/sugarme/gotch/vision"
)
const (
ImageDimNN int64 = 784
HiddenNodesNN int64 = 128
LabelNN int64 = 10
MnistDirNN string = "../../data/mnist"
epochsNN = 50
2020-06-18 08:14:48 +01:00
batchSizeNN = 256
LrNN = 1e-3
)
2020-06-18 16:37:13 +01:00
var l nn.Linear
2020-06-18 08:14:48 +01:00
func netInit(vs nn.Path) ts.Module {
n := nn.Seq()
2020-06-18 16:37:13 +01:00
l = nn.NewLinear(vs.Sub("layer1"), ImageDimNN, HiddenNodesNN, nn.DefaultLinearConfig())
n.Add(l)
n.AddFn(nn.ForwardWith(func(xs ts.Tensor) ts.Tensor {
2020-06-18 08:14:48 +01:00
return xs.MustRelu()
2020-06-18 16:37:13 +01:00
}))
2020-06-18 08:14:48 +01:00
n.Add(nn.NewLinear(vs, HiddenNodesNN, LabelNN, nn.DefaultLinearConfig()))
return n
}
func runNN() {
var ds vision.Dataset
ds = vision.LoadMNISTDir(MnistDirNN)
vs := nn.NewVarStore(gotch.CPU)
net := netInit(vs.Root())
2020-06-19 12:39:34 +01:00
2020-06-18 08:14:48 +01:00
opt, err := nn.DefaultAdamConfig().Build(vs, LrNN)
if err != nil {
log.Fatal(err)
}
for epoch := 0; epoch < epochsNN; epoch++ {
loss := net.Forward(ds.TrainImages).CrossEntropyForLogits(ds.TrainLabels)
2020-06-18 16:37:13 +01:00
opt.BackwardStep(loss)
2020-06-18 16:37:13 +01:00
lossVal := loss.MustShallowClone().MustView([]int64{-1}).MustFloat64Value([]int64{0})
2020-06-18 08:14:48 +01:00
testAccuracy := net.Forward(ds.TestImages).AccuracyForLogits(ds.TestLabels).MustView([]int64{-1}).MustFloat64Value([]int64{0})
2020-06-19 12:39:34 +01:00
fmt.Printf("Epoch: %v \t Loss: %.3f \t Test accuracy: %.2f%%\n", epoch, lossVal, testAccuracy*100)
fmt.Printf("Loss: %v\n", lossVal)
2020-06-18 08:14:48 +01:00
}
}