Releases: FluxML/Flux.jl
Releases ยท FluxML/Flux.jl
v0.16.9
v0.16.8
Flux v0.16.8
Merged pull requests:
- Enable other reactant tests (#2600) (@wsmoses)
- fix code in docs too wide (#2641) (@CarloLucibello)
- Set
early_stoppingdefault value forinit_scoretoInf(#2642) (@andreuvall) - extend
gradientto take an ADType argument (#2645) (@CarloLucibello) - allow passing an ADType to train! (#2650) (@CarloLucibello)
Closed issues:
- Local Response Normalisation (#192)
- generic_matmul! hit in
back!because type-promotion in activation function (#613) - Layer Transpose (#635)
- add
trainstep!(#666) - Hypernetwork API (#797)
- Optimizer handling of infinite loss (#821)
- Flux Optimizers should define equality (#823)
- Model optimization fails (NaNs) with Zygote.pullback but works with Tracker.forward (#876)
- more issue labels? (#879)
- Diagonal does not return same size due to broadcast (#890)
- Integrate epochs within Flux.train! (#1058)
- Simplest linear model on housing data w/ Flux (#1122)
- OneHotVector(i, n) when i > n (#1300)
- Support DirectML (#1347)
- Feature request: Modifying Dense Layer to accommodate kernel/bias constraints and kernel/bias regularisation (#1389)
- Flux.softmax returns wrong result with CuArray (#1425)
- Rethink train design and better callbacks support (#1461)
- tied weights (by transposition) are not tied when sent to gpu (#1504)
- Regularization example from docs can't be differentiated (#1588)
- Tied weights using Flux layers (#1592)
- Import Flux on worker crashes (#1625)
- Unclear wording in "Composing Optimizers" section of docs (#1627)
- Recurrent network interface updates/design (#1678)
- Please do not deprecate Dense(...;initb=...) (#1684)
- Triage Meetings (#1709)
- Differentiating a Model While Resetting the Parameters at every Epoch (#1766)
- BatchNorm on GPU without affine or tracking statistics (#1810)
- Float32 parameters in structs unsupported? (#1817)
- using Flux -> InitError: IOError: mkdir("/pbs/software/centos-7-x86_64/julia/1.7.0/share/julia/packages/Flux/BPPNj/src/data/../../deps"; mode=0o777): read-only file system (EROFS) (#1839)
- Easy to make mistake with
gpu()(#1887) - Inline printing for OneHotArray is not GPU-friendly (#1905)
- Freezing layers at model construction time (#1931)
- Unable to precompile -- "allequal not defined" (#1934)
- Weird Side Effects of
loadparams!(#1979) - Taking serialization seriously (#1988)
- Issue with logitcrossentropy on onehotencoded input on GPU (#2002)
- Add default inner constructor to Dense (#2158)
- docs on freezing layers should be ported to the explicit syntax (#2216)
- Loading Flux 0.13.15 for the first time results in error (#2232)
- Default for
init_scoreinearly_stopping(#2639) - Does not compile anymore in conjunction with CUDA due to dependency on
MLDataDevices(#2647)
v0.16.7
Flux v0.16.7
Merged pull requests:
- fix cuda ext for real (#2637) (@CarloLucibello)
Closed issues:
- Docker images for Floydhub and similar (#148)
- Implement einsum function/macro ร la PyTorch and TF (#297)
- Flux and Images (#326)
- "Tracing" memory pre-allocator (#349)
- make Juno dependency conditional (#454)
- Encoding array dimensions in flux type system? (#614)
- Gradient Interface Design (#628)
- New New Optimisers (#637)
- Clipping (#672)
- CUDA Programming Model (#706)
- LBFGS Optimizer (#719)
- Flux plots (#729)
- "ADAM" and friends should be called "Adam" (#795)
- Add lookahead optimizer (#838)
- ADAM does not accept keyword arguments (#871)
- Compatibility with Tracker (#883)
- Numerical issues for (logit)binarycrossentropy (#914)
- Change abstract argument names to meaningful ASCII (#915)
- Roadmap to Flux 1.0 (#961)
- Zygote gives extra gradient entries for BatchNorm (#1018)
- Helper methods for extracting RNN final state in a GPU compatible way (#1043)
- helper function for selecting a gpu in multi-gpu setting (#1074)
- Provide iper-simple examples directly in readme.md (#1115)
- gpu function does nothing, but only on first run (#1119)
- Behavior of chunk (#1120)
- ArrayFire (#1126)
- MethodError: no method matching zero(::Type{Array{Float32,2}}) In Flux Loss function (#1134)
- Parameter collection and GPU movement fail on models defined via functions (#1201)
- Derivative in loss function error (#1464)
- Document OneHotArray (#1519)
- Second order derivative (#1582)
Convis not working for Complex when using CUDA (#1655)- Flux installation errors in julia 1.7.0-rc1, WSL2 (#1757)
- Two-arg update!(x, d) is never used (#1860)
- cpu() type stability (#1878)
v0.16.6
Flux v0.16.6
Merged pull requests:
- Bump actions/checkout from 4 to 5 (#2615) (@dependabot[bot])
- Fix Typos in Docs (#2619) (@Solanasol233)
- Fix reference to inner gradient structure in docs basics.md (#2622) (@AbelHo)
- Fix the "Gradients and Layers" documentation (#2628) (#2629) (@gomurin0428)
- Bump actions/checkout from 5 to 6 (#2630) (@dependabot[bot])
- fix cuda extension on julia 1.12 (#2632) (@CarloLucibello)
- fix CI (#2634) (@CarloLucibello)
Closed issues:
- Enzyme gradient example broken (#2554)
- [enzyme] broken Bilinear gradient (#2565)
- [enzyme] broken MultiHeadAttention gradient (#2567)
- Regression:
cpufunction is incompatible with DataFrames since Functors became opt-in (#2617) - Flux on GPU is incompatible with NVIDIA driver version 13 (#2618)
- Views built within MLUtils.kfolds scalar indexing error on gpu (#2620)
- Simple single layer line fit does not converge for most sets of sample points (#2623)
- Julia 1.12: Deadlock detected in loading Flux ext (#2625)
- Enzyme Documentation example does not work on CPU (#2627)
- Minor documentation issue ("Gradients and Layers") (#2628)
- Deadlock when loading
FluxCUDAcuDNNExton Julia 1.12.2 (works on 1.11.7) (#2631) - [Metal] error in forward pass with tanh activation (#2633)
v0.16.5
Flux v0.16.5
Merged pull requests:
- Fix Typos in Old Tutorials Documentation (#2610) (@leopardracer)
- CompatHelper: bump compat for AMDGPU in [weakdeps] to 2, (keep existing compat) (#2613) (@github-actions[bot])
- Bump to 0.16.5 (#2614) (@pxl-th)
Closed issues:
unsafe_free!from MLDataDevices fails for CuArray{CartesianIndex{4}, 1, CUDA.DeviceMemory}) (#2612)
v0.16.4
Flux v0.16.4
Merged pull requests:
- Fix missing imports in FluxMPIExt (#2589) (@Alexander-Barth)
- Added shape validation for Conv weight tensor (#2590) (@mikymatt01)
- Disable broken Reactant tests for now (#2595) (@ToucheSir)
- Test re enable reactant test to identify failures (#2596) (@wsmoses)
- fix recurrent docstrings (#2597) (@CarloLucibello)
- Fix loading of pooling layers (#2598) (@adrhill)
- Fix typos (#2601) (@omahs)
- small updates to tests (#2602) (@CarloLucibello)
- reinstate enzyme tests (#2603) (@CarloLucibello)
- fix docs CI (#2604) (@CarloLucibello)
- fix piracy with DataLoader (#2608) (@CarloLucibello)
Closed issues:
v0.16.3
v0.16.2
v0.16.1
Flux v0.16.1
Merged pull requests:
- Adding RecurrentLayers to ecosystem.md (#2555) (@MartinuzziFrancesco)
- Fixed typo in recurrence documentation (#2556) (@MartinuzziFrancesco)
- Adding return state option to recurrent layers (#2557) (@MartinuzziFrancesco)
- update Schedulers docs (#2560) (@CarloLucibello)
- collapse doc string in layers docs (#2562) (@CarloLucibello)
- fix test enzyme (#2563) (@CarloLucibello)
- Remove 2 items from public, to fix 1.12 (#2569) (@mcabbott)
- Add reactant forward and reverse pass tests (#2576) (@wsmoses)
- cleanup Reactant and Enzyme tests (#2578) (@CarloLucibello)
Closed issues:
- cell output is not clearly distinguishable from the state (#2548)
- Flux.cpu and Flux.gpu no longer move data on views (#2553)
- remove usage example of old optimiser (#2558)
- Optimizing over
AbstractMatrixsubtypes (#2559) - introduce a FlattenLayer (#2561)
- [enzyme] broken MeanPool gradient (#2564)
- [enzyme] broken BatchNorm gradient (#2566)
- [enyzme] broken recurrent cell loss (#2568)
v0.16.0
Flux v0.16.0
Highlights
This release has a single breaking change:
- The recurrent cells
RNNCell,LSTMCell, andGRUCellforward has been changed to
$y_t, state_t = cell(x_t, state_{t-1})$ in (#2551). Previously, it was$state_t = cell(x_t, state_{t-1})$ .
Other highlights include:
- Added
WeightNormnormalization layer. - Added
Recurrencelayer, turning a recurrent layer into a layer processing the entire sequence at once.
Merged pull requests:
- Recurrence layer (#2549) (@CarloLucibello)
- Add
WeightNormreparametrization (#2550) (@pxl-th) - Change cells' return to
out, state(#2551) (@CarloLucibello) - fix:
gpu_devicenot defined inFlux.DistributedUtils(#2552) (@AntonOresten)
Closed issues: