You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
One of the "tough luck" paths that go through the clustering section appear to say this is the case when there are >10k samples.
Suggest a potential alternative/fix
However, with modern computational hardware, and the optimized implementation of DBSCAN in Scikit-learn, it appears that it may be helpful to recommend DBSCAN as a possible solution for datasets containing <100K or even <1M datapoints for clustering in reasonable amounts of time on CPU.
The text was updated successfully, but these errors were encountered:
Indeed, we should update this map. We add IRL discussion with @ArturoAmorQ and @GaelVaroquaux regarding this topic. The map is missing new estimators. We could think also about more dynamic breakdown when zooming on the map. Anyway this a good suggestion, we need to come up with a plan to execute it properly.
Describe the issue linked to the documentation
As seen here:
https://scikit-learn.org/stable/tutorial/machine_learning_map/index.html
One of the "tough luck" paths that go through the clustering section appear to say this is the case when there are >10k samples.
Suggest a potential alternative/fix
However, with modern computational hardware, and the optimized implementation of DBSCAN in Scikit-learn, it appears that it may be helpful to recommend DBSCAN as a possible solution for datasets containing <100K or even <1M datapoints for clustering in reasonable amounts of time on CPU.
The text was updated successfully, but these errors were encountered: