You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Move doc sections to "guide" + "reference" (#2115)
* make guide + reference
* try moving Fitting a Line to tutorials
* restore Fitting a Line's place
* remove duplicate functor docstring (it's in Functors section)
* alter "Learning Flux" section to match
* reduce the amount of text on welcome page
* add a CUDA.jl reference page
* add Alternatives to Flux section
* fixup
* hide some tutorials, try moving Ecosystem up
* remove the 📚 marker
* tweaks to links page
* trim even more surplus text from welcome page
* rm CUDA page
* the welcome page is short enough not to need navigation, and using H3 saves space on the navigation panel so you can see more other things
* fix misleading sentence in nnlib intro
* remove CUDA
* Apply 2 suggestions
Co-authored-by: Brian Chen <[email protected]>
* Update docs/src/index.md
* add terminology note
* remove example functions from headings
* fixup
Co-authored-by: Brian Chen <[email protected]>
Copy file name to clipboardExpand all lines: docs/src/ecosystem.md
+28-3
Original file line number
Diff line number
Diff line change
@@ -5,11 +5,11 @@ globally providing a rich and consistent user experience.
5
5
6
6
This is a non-exhaustive list of Julia packages, nicely complementing `Flux` in typical
7
7
machine learning and deep learning workflows. To add your project please send a [PR](https://github.com/FluxML/Flux.jl/pulls).
8
-
See also academic work citing Flux or Zygote.
8
+
See also academic work [citing Flux](https://scholar.google.com/scholar?cites=9731162218836700005&hl=en) or [citing Zygote](https://scholar.google.com/scholar?cites=11943854577624257878&hl=en).
9
9
10
10
## Flux models
11
11
12
-
Packages that are actual `Flux` models but are not available directly through the `Flux` package.
12
+
-Flux's [model-zoo](https://github.com/FluxML/model-zoo) contains examples from many domains.
13
13
14
14
### Computer vision
15
15
@@ -38,6 +38,8 @@ Packages that are actual `Flux` models but are not available directly through th
38
38
39
39
-[FluxArchitectures.jl](https://github.com/sdobber/FluxArchitectures.jl) is a collection of advanced network architectures for time series forecasting.
40
40
41
+
---
42
+
41
43
## Tools closely associated with Flux
42
44
43
45
Utility tools you're unlikely to have met if you never used Flux!
@@ -64,9 +66,10 @@ Tools to put data into the right order for creating a model.
64
66
65
67
### Parameters
66
68
67
-
-[Parameters.jl](https://github.com/mauro3/Parameters.jl) types with default field values, keyword constructors and (un-)pack macros.
68
69
-[ParameterSchedulers.jl](https://github.com/darsnack/ParameterSchedulers.jl) standard scheduling policies for machine learning.
69
70
71
+
---
72
+
70
73
## Differentiable programming
71
74
72
75
Packages based on differentiable programming but not necessarily related to Machine Learning.
@@ -90,6 +93,7 @@ Packages based on differentiable programming but not necessarily related to Mach
90
93
91
94
-[OnlineStats.jl](https://github.com/joshday/OnlineStats.jl) provides single-pass algorithms for statistics.
92
95
96
+
---
93
97
94
98
## Useful miscellaneous packages
95
99
@@ -104,8 +108,29 @@ Some useful and random packages!
104
108
-[ProgressMeter.jl](https://github.com/timholy/ProgressMeter.jl) progress meters for long-running computations.
105
109
-[TensorBoardLogger.jl](https://github.com/PhilipVinc/TensorBoardLogger.jl) easy peasy logging to [tensorboard](https://www.tensorflow.org/tensorboard) in Julia
106
110
-[ArgParse.jl](https://github.com/carlobaldassi/ArgParse.jl) is a package for parsing command-line arguments to Julia programs.
111
+
-[Parameters.jl](https://github.com/mauro3/Parameters.jl) types with default field values, keyword constructors and (un-)pack macros.
107
112
-[BSON.jl](https://github.com/JuliaIO/BSON.jl) is a package for working with the Binary JSON serialisation format.
108
113
-[DataFrames.jl](https://github.com/JuliaData/DataFrames.jl) in-memory tabular data in Julia.
109
114
-[DrWatson.jl](https://github.com/JuliaDynamics/DrWatson.jl) is a scientific project assistant software.
110
115
111
116
This tight integration among Julia packages is shown in some of the examples in the [model-zoo](https://github.com/FluxML/model-zoo) repository.
117
+
118
+
---
119
+
120
+
## Alternatives to Flux
121
+
122
+
Julia has several other libraries for making neural networks.
123
+
124
+
*[SimpleChains.jl](https://github.com/PumasAI/SimpleChains.jl) is focused on making small, simple, CPU-based, neural networks fast. Uses [LoopVectorization.jl](https://github.com/JuliaSIMD/LoopVectorization.jl). (Was `FastChain` in DiffEqFlux.jl)
125
+
126
+
*[Knet.jl](https://github.com/denizyuret/Knet.jl) is a neural network library built around [AutoGrad.jl](https://github.com/denizyuret/AutoGrad.jl).
127
+
128
+
*[Lux.jl](https://github.com/avik-pal/Lux.jl) (earlier ExplicitFluxLayers.jl) shares much of the design, use-case, and NNlib.jl / Optimisers.jl back-end of Flux. But instead of encapsulating all parameters within the model structure, it separates this into 3 components: a model, a tree of parameters, and a tree of model states.
129
+
130
+
!!! compat "Explicit or explicit?"
131
+
Flux's [training docs](@ref man-training) talk about changes from Zygote's implicit to
132
+
explicit gradients, dictionary-like to tree-like structures.
133
+
(See also [Zygote's description](https://fluxml.ai/Zygote.jl/dev/#Explicit-and-Implicit-Parameters-1) of these.)
134
+
Lux also uses Zygote, but uses the word "explicit" to mean something unrelated,
135
+
namely storing the tree of parameters (and of state) separately from the model.
More information for conditional use of GPUs in CUDA.jl can be found in its [documentation](https://cuda.juliagpu.org/stable/installation/conditional/#Conditional-use), and information about the specific use of the variable is described in the [Nvidia CUDA blog post](https://developer.nvidia.com/blog/cuda-pro-tip-control-gpu-visibility-cuda_visible_devices/).
Copy file name to clipboardExpand all lines: docs/src/index.md
+11-15
Original file line number
Diff line number
Diff line change
@@ -2,29 +2,25 @@
2
2
3
3
Flux is a library for machine learning. It comes "batteries-included" with many useful tools built in, but also lets you use the full power of the Julia language where you need it. We follow a few key principles:
4
4
5
-
***Doing the obvious thing**. Flux has relatively few explicit APIs for features like regularisation or embeddings. Instead, writing down the mathematical form will work – and be fast.
6
-
***Extensible by default**. Flux is written to be highly extensible and flexible while being performant. Extending Flux is as simple as using your own code as part of the model you want - it is all [high-level Julia code](https://github.com/FluxML/Flux.jl/blob/ec16a2c77dbf6ab8b92b0eecd11661be7a62feef/src/layers/recurrent.jl#L131). When in doubt, it’s well worth looking at [the source](https://github.com/FluxML/Flux.jl/tree/master/src). If you need something different, you can easily roll your own.
7
-
***Play nicely with others**. Flux works well with Julia libraries from [images](https://github.com/JuliaImages/Images.jl) to [differential equation solvers](https://github.com/SciML/DifferentialEquations.jl), so you can easily build complex data processing pipelines that integrate Flux models.
5
+
***Doing the obvious thing**. Flux has relatively few explicit APIs. Instead, writing down the mathematical form will work – and be fast.
6
+
***Extensible by default**. Flux is written to be highly flexible while being performant. Extending Flux is as simple as using your own code as part of the model you want - it is all [high-level Julia code](https://github.com/FluxML/Flux.jl/tree/master/src).
7
+
***Play nicely with others**. Flux works well with unrelated Julia libraries from [images](https://github.com/JuliaImages/Images.jl) to [differential equation solvers](https://github.com/SciML/DifferentialEquations.jl), rather than duplicating them.
8
8
9
-
## Installation
9
+
###Installation
10
10
11
-
Download [Julia 1.6](https://julialang.org/downloads/) or later, preferably the current stable release. You can add Flux using Julia's package manager, by typing `] add Flux` in the Julia prompt.
11
+
Download [Julia 1.6](https://julialang.org/downloads/) or later, preferably the current stable release. You can add Flux using Julia's package manager, by typing `] add Flux` in the Julia prompt. This will automatically install several other packages, including [CUDA.jl](https://github.com/JuliaGPU/CUDA.jl) for Nvidia GPU support.
12
12
13
-
This will automatically install several other packages, including [CUDA.jl](https://github.com/JuliaGPU/CUDA.jl) which supports Nvidia GPUs. To directly access some of its functionality, you may want to add `] add CUDA` too. The page on [GPU support](gpu.md) has more details.
13
+
### Learning Flux
14
14
15
-
Other closely associated packages, also installed automatically, include [Zygote](https://github.com/FluxML/Zygote.jl), [Optimisers](https://github.com/FluxML/Optimisers.jl), [NNlib](https://github.com/FluxML/NNlib.jl), [Functors](https://github.com/FluxML/Functors.jl) and [MLUtils](https://github.com/JuliaML/MLUtils.jl).
15
+
The **[quick start](@ref man-quickstart)** page trains a simple neural network.
16
16
17
-
## Learning Flux
17
+
This rest of the **guide** provides a from-scratch introduction to Flux's take on models and how they work, starting with [fitting a line](@ref man-overview). Once you understand these docs, congratulations, you also understand [Flux's source code](https://github.com/FluxML/Flux.jl), which is intended to be concise, legible and a good reference for more advanced concepts.
18
18
19
-
The [quick start](@ref man-quickstart) page trains a simple neural network.
19
+
There are some **tutorials** about building particular models. The **[model zoo](https://github.com/FluxML/model-zoo/)** has starting points for many other common ones. And finally, the **[ecosystem page](ecosystem.md)** lists packages which define Flux models.
20
20
21
-
This rest of this documentation provides a from-scratch introduction to Flux's take on models and how they work, starting with [fitting a line](@ref man-overview). Once you understand these docs, congratulations, you also understand [Flux's source code](https://github.com/FluxML/Flux.jl), which is intended to be concise, legible and a good reference for more advanced concepts.
21
+
The **reference** section includes, beside Flux's own functions, those of some companion packages: [Zygote.jl](https://github.com/FluxML/Zygote.jl) (automatic differentiation), [Optimisers.jl](https://github.com/FluxML/Optimisers.jl) (training) and others.
22
22
23
-
Sections with 📚 contain API listings. The same text is avalable at the Julia prompt, by typing for example `?gpu`.
24
-
25
-
If you just want to get started writing models, the [model zoo](https://github.com/FluxML/model-zoo/) gives good starting points for many common ones.
26
-
27
-
## Community
23
+
### Community
28
24
29
25
Everyone is welcome to join our community on the [Julia discourse forum](https://discourse.julialang.org/), or the [slack chat](https://discourse.julialang.org/t/announcing-a-julia-slack/4866) (channel #machine-learning). If you have questions or issues we'll try to help you out.
Copy file name to clipboardExpand all lines: docs/src/models/basics.md
+2-6
Original file line number
Diff line number
Diff line change
@@ -213,13 +213,13 @@ m(5) # => 26
213
213
214
214
## Layer Helpers
215
215
216
-
There is still one problem with this `Affine` layer, that Flux does not know to look inside it. This means that [`Flux.train!`](@ref) won't see its parameters, nor will [`gpu`](@ref) be able to move them to your GPU. These features are enabled by the `@functor` macro:
216
+
There is still one problem with this `Affine` layer, that Flux does not know to look inside it. This means that [`Flux.train!`](@ref) won't see its parameters, nor will [`gpu`](@ref) be able to move them to your GPU. These features are enabled by the [`@functor`](@ref Functors.@functor) macro:
217
217
218
218
```
219
219
Flux.@functor Affine
220
220
```
221
221
222
-
Finally, most Flux layers make bias optional, and allow you to supply the function used for generating random weights. We can easily add these refinements to the `Affine` layer as follows:
222
+
Finally, most Flux layers make bias optional, and allow you to supply the function used for generating random weights. We can easily add these refinements to the `Affine` layer as follows, using the helper function [`create_bias`](@ref Flux.create_bias):
223
223
224
224
```
225
225
function Affine((in, out)::Pair; bias=true, init=Flux.randn32)
Copy file name to clipboardExpand all lines: docs/src/models/nnlib.md
+1-1
Original file line number
Diff line number
Diff line change
@@ -1,6 +1,6 @@
1
1
# Neural Network primitives from NNlib.jl
2
2
3
-
Flux re-exports all of the functions exported by the [NNlib](https://github.com/FluxML/NNlib.jl) package. This includes activation functions, described on the next page. Many of the functions on this page exist primarily as the internal implementation of Flux layer, but can also be used independently.
3
+
Flux re-exports all of the functions exported by the [NNlib](https://github.com/FluxML/NNlib.jl) package. This includes activation functions, described on [their own page](@ref man-activations). Many of the functions on this page exist primarily as the internal implementation of Flux layer, but can also be used independently.
0 commit comments