neat alternatives and similar packages
Based on the "Machine Learning" category.
Alternatively, view neat alternatives based on common mentions on social networks and blogs.
-
Gorgonia
Gorgonia is a library that helps facilitate machine learning in Go. -
m2cgen
Transform ML models into a native code (Java, C, Python, Go, JavaScript, Visual Basic, C#, R, PowerShell, PHP, Dart, Haskell, Ruby, F#, Rust) with zero dependencies -
gosseract
Go package for OCR (Optical Character Recognition), by using Tesseract C++ library -
gago
:four_leaf_clover: Evolutionary optimization library for Go (genetic algorithm, partical swarm optimization, differential evolution) -
ocrserver
A simple OCR API server, seriously easy to be deployed by Docker, on Heroku as well -
onnx-go
onnx-go gives the ability to import a pre-trained neural network within Go without being linked to a framework or library. -
shield
Bayesian text classifier with flexible tokenizers and storage backends for Go -
go-featureprocessing
🔥 Fast, simple sklearn-like feature processing for Go -
neural-go
A multilayer perceptron network implemented in Go, with training via backpropagation. -
go-cluster
k-modes and k-prototypes clustering algorithms implementation in Go
Access the most powerful time series database as a service
Do you think we are missing an alternative of neat or a related project?
Popular Comparisons
README
CURRENTLY NOT WORKING! There will be a further notice when it's updated.
NEAT (NeuroEvolution of Augmenting Topologies) is a neuroevolution algorithm by Dr. Kenneth O. Stanley which evolves not only neural networks' weights but also their topologies. This method starts the evolution process with genomes with minimal structure, then complexifies the structure of each genome as it progresses. You can read the original paper from here.
Installation
To install neat
run the following:
$ go get -u github.com/jinyeom/neat
Usage
This NEAT package is as simple as plug and play. All you have to do is to create a new instance of NEAT, given the configuration from a JSON file, for which the template is provided below, and an evaluation method of a neural network, and run.
{
"experimentName": "XOR Test",
"verbose": true,
"numInputs": 3,
"numOutputs": 1,
"fullyConnected": false,
"numGenerations": 50,
"populationSize": 100,
"initFitness": 9999.0,
"minimizeFitness": true,
"survivalRate": 0.5,
"stagnationLimit": 5,
"ratePerturb": 0.2,
"rateAddNode": 0.2,
"rateAddConn": 0.2,
"rateMutateChild": 0.5,
"distanceThreshold": 20.0,
"coeffUnmatching": 1.0,
"coeffMatching": 1.0,
"cppnActivations": [],
}
Now that you have the configuration JSON file is ready as config.json
, we can
start experiment with NEAT. Below is an example XOR experiment.
package main
import (
"log"
"math"
// Import NEAT package after installing the package through
// the instruction provided above.
"github.com/jinyeom/neat"
)
func main() {
// First, create a new instance of Config from the JSON file created above.
// If there's a file import error, the program will crash.
config, err := neat.NewConfigJSON("config.json")
if err != nil{
log.Fatal(err)
}
// Then, we can define the evaluation function, which is a type of function
// which takes a neural network, evaluates its performance, and returns some
// score that indicates its performance. This score is essentially a genome's
// fitness score. With the configuration and the evaluation function we
// defined, we can create a new instance of NEAT and start the evolution
// process.
neat.New(config, neat.XORTest()).Run()
}
License
This package is under GNU General Public License.
*Note that all licence references and agreements mentioned in the neat README section above
are relevant to that project's source code only.