# Version 0.12.1¶

**October 8, 2012**

The 0.12.1 release is a bug-fix release with no additional features, but is instead a set of bug fixes

## Changelog¶

- Improved numerical stability in spectral embedding by Gael Varoquaux
- Doctest under windows 64bit by Gael Varoquaux
- Documentation fixes for elastic net by Andreas Müller and Alexandre Gramfort
- Proper behavior with fortran-ordered NumPy arrays by Gael Varoquaux
- Make GridSearchCV work with non-CSR sparse matrix by Lars Buitinck
- Fix parallel computing in MDS by Gael Varoquaux
- Fix Unicode support in count vectorizer by Andreas Müller
- Fix MinCovDet breaking with X.shape = (3, 1) by :user:`Virgile Fritsch <VirgileFritsch>`
- Fix clone of SGD objects by Peter Prettenhofer
- Stabilize GMM by :user:`Virgile Fritsch <VirgileFritsch>`

# Version 0.12¶

**September 4, 2012**

## Changelog¶

- Various speed improvements of the decision trees module, by Gilles Louppe.
`ensemble.GradientBoostingRegressor`

and`ensemble.GradientBoostingClassifier`

now support feature subsampling via the`max_features`

argument, by Peter Prettenhofer.- Added Huber and Quantile loss functions to
`ensemble.GradientBoostingRegressor`

, by Peter Prettenhofer. - Decision trees and forests of randomized trees now support multi-output classification and regression problems, by Gilles Louppe.
- Added
`preprocessing.LabelEncoder`

, a simple utility class to normalize labels or transform non-numerical labels, by Mathieu Blondel. - Added the epsilon-insensitive loss and the ability to make probabilistic predictions with the modified huber loss in 随机梯度下降(Stochastic Gradient Descent), by Mathieu Blondel.
- Added 多维尺度分析(MDS), by Nelle Varoquaux.
- SVMlight file format loader now detects compressed (gzip/bzip2) files and decompresses them on the fly, by Lars Buitinck.
- SVMlight file format serializer now preserves double precision floating point values, by Olivier Grisel.
- A common testing framework for all estimators was added, by Andreas Müller.
- Understandable error messages for estimators that do not accept sparse input by Gael Varoquaux
- Speedups in hierarchical clustering by Gael Varoquaux. In particular building the tree now supports early stopping. This is useful when the number of clusters is not small compared to the number of samples.
- Add MultiTaskLasso and MultiTaskElasticNet for joint feature selection, by Alexandre Gramfort.
- Added
`metrics.auc_score`

and`metrics.average_precision_score`

convenience functions by Andreas Müller. - Improved sparse matrix support in the 特征选择(Feature selection) module by Andreas Müller.
- New word boundaries-aware character n-gram analyzer for the 文本特征提取 module by :user:`@kernc <kernc>`.
- Fixed bug in spectral clustering that led to single point clusters by Andreas Müller.
- In
`feature_extraction.text.CountVectorizer`

, added an option to ignore infrequent words,`min_df`

by Andreas Müller. - Add support for multiple targets in some linear models (ElasticNet, Lasso and OrthogonalMatchingPursuit) by Vlad Niculae and Alexandre Gramfort.
- Fixes in
`decomposition.ProbabilisticPCA`

score function by Wei Li. - Fixed feature importance computation in Gradient Tree Boosting.

## API changes summary¶

- The old
`scikits.learn`

package has disappeared; all code should import from`sklearn`

instead, which was introduced in 0.9. - In
`metrics.roc_curve`

, the`thresholds`

array is now returned with it’s order reversed, in order to keep it consistent with the order of the returned`fpr`

and`tpr`

. - In
`hmm`

objects, like`hmm.GaussianHMM`

,`hmm.MultinomialHMM`

, etc., all parameters must be passed to the object when initialising it and not through`fit`

. Now`fit`

will only accept the data as an input parameter. - For all SVM classes, a faulty behavior of
`gamma`

was fixed. Previously, the default gamma value was only computed the first time`fit`

was called and then stored. It is now recalculated on every call to`fit`

. - All
`Base`

classes are now abstract meta classes so that they can not be instantiated. `cluster.ward_tree`

now also returns the parent array. This is necessary for early-stopping in which case the tree is not completely built.- In
`feature_extraction.text.CountVectorizer`

the parameters`min_n`

and`max_n`

were joined to the parameter`n_gram_range`

to enable grid-searching both at once. - In
`feature_extraction.text.CountVectorizer`

, words that appear only in one document are now ignored by default. To reproduce the previous behavior, set`min_df=1`

. - Fixed API inconsistency:
`linear_model.SGDClassifier.predict_proba`

now returns 2d array when fit on two classes. - Fixed API inconsistency:
`discriminant_analysis.QuadraticDiscriminantAnalysis.decision_function`

and`discriminant_analysis.LinearDiscriminantAnalysis.decision_function`

now return 1d arrays when fit on two classes. - Grid of alphas used for fitting
`linear_model.LassoCV`

and`linear_model.ElasticNetCV`

is now stored in the attribute`alphas_`

rather than overriding the init parameter`alphas`

. - Linear models when alpha is estimated by cross-validation store
the estimated value in the
`alpha_`

attribute rather than just`alpha`

or`best_alpha`

. `ensemble.GradientBoostingClassifier`

now supports`ensemble.GradientBoostingClassifier.staged_predict_proba`

, and`ensemble.GradientBoostingClassifier.staged_predict`

.`svm.sparse.SVC`

and other sparse SVM classes are now deprecated. The all classes in the 支持向量机(Support Vector Machines) module now automatically select the sparse or dense representation base on the input.- All clustering algorithms now interpret the array
`X`

given to`fit`

as input data, in particular`cluster.SpectralClustering`

and`cluster.AffinityPropagation`

which previously expected affinity matrices. - For clustering algorithms that take the desired number of clusters as a parameter,
this parameter is now called
`n_clusters`

.

## People¶

- 267 Andreas Müller
- 94 Gilles Louppe
- 89 Gael Varoquaux
- 79 Peter Prettenhofer
- 60 Mathieu Blondel
- 57 Alexandre Gramfort
- 52 Vlad Niculae
- 45 Lars Buitinck
- 44 Nelle Varoquaux
- 37 Jaques Grobler
- 30 Alexis Mignon
- 30 Immanuel Bayer
- 27 Olivier Grisel
- 16 Subhodeep Moitra
- 13 Yannick Schwartz
- 12 :user:`@kernc <kernc>`
- 11 :user:`Virgile Fritsch <VirgileFritsch>`
- 9 Daniel Duckworth
- 9 Fabian Pedregosa
- 9 Robert Layton
- 8 John Benediktsson
- 7 Marko Burjek
- 5 Nicolas Pinto
- 4 Alexandre Abraham
- 4 Jake Vanderplas
- 3 Brian Holt
- 3 Edouard Duchesnay
- 3 Florian Hoenig
- 3 flyingimmidev
- 2 Francois Savard
- 2 Hannes Schulz
- 2 Peter Welinder
- 2 Yaroslav Halchenko
- 2 Wei Li
- 1 Alex Companioni
- 1 Brandyn A. White
- 1 Bussonnier Matthias
- 1 Charles-Pierre Astolfi
- 1 Dan O’Huiginn
- 1 David Cournapeau
- 1 Keith Goodman
- 1 Ludwig Schwardt
- 1 Olivier Hervieu
- 1 Sergio Medina
- 1 Shiqiao Du
- 1 Tim Sheerman-Chase
- 1 buguen

# Version 0.11¶

**May 7, 2012**

## Changelog¶

### Highlights¶

- Gradient boosted regression trees (Gradient Tree Boosting) for classification and regression by Peter Prettenhofer and Scott White .
- Simple dict-based feature loader with support for categorical variables
(
`feature_extraction.DictVectorizer`

) by Lars Buitinck. - Added Matthews correlation coefficient (
`metrics.matthews_corrcoef`

) and added macro and micro average options to`metrics.precision_score`

,`metrics.recall_score`

and`metrics.f1_score`

by Satrajit Ghosh. - Out of Bag Estimates of generalization error for 集成学习方法(Ensemble methods) by Andreas Müller.
- Randomized sparse linear models for feature selection, by Alexandre Gramfort and Gael Varoquaux
- 标签传播(Label Propagation) for semi-supervised learning, by Clay
Woolam.
**Note**the semi-supervised API is still work in progress, and may change. - Added BIC/AIC model selection to classical 高斯混合模型(Gaussian mixture models) and unified the API with the remainder of scikit-learn, by Bertrand Thirion
- Added
`sklearn.cross_validation.StratifiedShuffleSplit`

, which is a`sklearn.cross_validation.ShuffleSplit`

with balanced splits, by Yannick Schwartz. `sklearn.neighbors.NearestCentroid`

classifier added, along with a`shrink_threshold`

parameter, which implements**shrunken centroid classification**, by Robert Layton.

### Other changes¶

- Merged dense and sparse implementations of 随机梯度下降(Stochastic Gradient Descent) module and
exposed utility extension types for sequential
datasets
`seq_dataset`

and weight vectors`weight_vector`

by Peter Prettenhofer. - Added
`partial_fit`

(support for online/minibatch learning) and warm_start to the 随机梯度下降(Stochastic Gradient Descent) module by Mathieu Blondel. - Dense and sparse implementations of 支持向量机(Support Vector Machines) classes and
`linear_model.LogisticRegression`

merged by Lars Buitinck. - Regressors can now be used as base estimator in the 多类和多标签算法(Multiclass and multilabel algorithms) module by Mathieu Blondel.
- Added n_jobs option to
`metrics.pairwise.pairwise_distances`

and`metrics.pairwise.pairwise_kernels`

for parallel computation, by Mathieu Blondel. - K-均值 can now be run in parallel, using the
`n_jobs`

argument to either K-均值 or`KMeans`

, by Robert Layton. - Improved 交叉验证:评估估计器的性能 and 通过网格搜索调节估计器超参数 documentation
and introduced the new
`cross_validation.train_test_split`

helper function by Olivier Grisel `svm.SVC`

members`coef_`

and`intercept_`

changed sign for consistency with`decision_function`

; for`kernel==linear`

,`coef_`

was fixed in the one-vs-one case, by Andreas Müller.- Performance improvements to efficient leave-one-out cross-validated
Ridge regression, esp. for the
`n_samples > n_features`

case, in`linear_model.RidgeCV`

, by Reuben Fletcher-Costin. - Refactoring and simplification of the 文本特征提取 API and fixed a bug that caused possible negative IDF, by Olivier Grisel.
- Beam pruning option in
`_BaseHMM`

module has been removed since it is difficult to Cythonize. If you are interested in contributing a Cython version, you can use the python version in the git history as a reference. - Classes in 最近邻方法(Nearest Neighbors) now support arbitrary Minkowski metric for
nearest neighbors searches. The metric can be specified by argument
`p`

.

## API changes summary¶

`covariance.EllipticEnvelop`

is now deprecated - Please use`covariance.EllipticEnvelope`

instead.`NeighborsClassifier`

and`NeighborsRegressor`

are gone in the module 最近邻方法(Nearest Neighbors). Use the classes`KNeighborsClassifier`

,`RadiusNeighborsClassifier`

,`KNeighborsRegressor`

and/or`RadiusNeighborsRegressor`

instead.- Sparse classes in the 随机梯度下降(Stochastic Gradient Descent) module are now deprecated.
- In
`mixture.GMM`

,`mixture.DPGMM`

and`mixture.VBGMM`

, parameters must be passed to an object when initialising it and not through`fit`

. Now`fit`

will only accept the data as an input parameter. - methods
`rvs`

and`decode`

in`GMM`

module are now deprecated.`sample`

and`score`

or`predict`

should be used instead. - attribute
`_scores`

and`_pvalues`

in univariate feature selection objects are now deprecated.`scores_`

or`pvalues_`

should be used instead. - In
`LogisticRegression`

,`LinearSVC`

,`SVC`

and`NuSVC`

, the`class_weight`

parameter is now an initialization parameter, not a parameter to fit. This makes grid searches over this parameter possible. - LFW
`data`

is now always shape`(n_samples, n_features)`

to be consistent with the Olivetti faces dataset. Use`images`

and`pairs`

attribute to access the natural images shapes instead. - In
`svm.LinearSVC`

, the meaning of the`multi_class`

parameter changed. Options now are`'ovr'`

and`'crammer_singer'`

, with`'ovr'`

being the default. This does not change the default behavior but hopefully is less confusing. - Class
`feature_selection.text.Vectorizer`

is deprecated and replaced by`feature_selection.text.TfidfVectorizer`

. - The preprocessor / analyzer nested structure for text feature
extraction has been removed. All those features are
now directly passed as flat constructor arguments
to
`feature_selection.text.TfidfVectorizer`

and`feature_selection.text.CountVectorizer`

, in particular the following parameters are now used: `analyzer`

can be`'word'`

or`'char'`

to switch the default analysis scheme, or use a specific python callable (as previously).`tokenizer`

and`preprocessor`

have been introduced to make it still possible to customize those steps with the new API.`input`

explicitly control how to interpret the sequence passed to`fit`

and`predict`

: filenames, file objects or direct (byte or Unicode) strings.- charset decoding is explicit and strict by default.
- the
`vocabulary`

, fitted or not is now stored in the`vocabulary_`

attribute to be consistent with the project conventions. - Class
`feature_selection.text.TfidfVectorizer`

now derives directly from`feature_selection.text.CountVectorizer`

to make grid search trivial. - methods
`rvs`

in`_BaseHMM`

module are now deprecated.`sample`

should be used instead. - Beam pruning option in
`_BaseHMM`

module is removed since it is difficult to be Cythonized. If you are interested, you can look in the history codes by git. - The SVMlight format loader now supports files with both zero-based and one-based column indices, since both occur “in the wild”.
- Arguments in class
`ShuffleSplit`

are now consistent with`StratifiedShuffleSplit`

. Arguments`test_fraction`

and`train_fraction`

are deprecated and renamed to`test_size`

and`train_size`

and can accept both`float`

and`int`

. - Arguments in class
`Bootstrap`

are now consistent with`StratifiedShuffleSplit`

. Arguments`n_test`

and`n_train`

are deprecated and renamed to`test_size`

and`train_size`

and can accept both`float`

and`int`

. - Argument
`p`

added to classes in 最近邻方法(Nearest Neighbors) to specify an arbitrary Minkowski metric for nearest neighbors searches.

## People¶

- 282 Andreas Müller
- 239 Peter Prettenhofer
- 198 Gael Varoquaux
- 129 Olivier Grisel
- 114 Mathieu Blondel
- 103 Clay Woolam
- 96 Lars Buitinck
- 88 Jaques Grobler
- 82 Alexandre Gramfort
- 50 Bertrand Thirion
- 42 Robert Layton
- 28 flyingimmidev
- 26 Jake Vanderplas
- 26 Shiqiao Du
- 21 Satrajit Ghosh
- 17 David Marek
- 17 Gilles Louppe
- 14 Vlad Niculae
- 11 Yannick Schwartz
- 10 Fabian Pedregosa
- 9 fcostin
- 7 Nick Wilson
- 5 Adrien Gaidon
- 5 Nicolas Pinto
- 4 David Warde-Farley
- 5 Nelle Varoquaux
- 5 Emmanuelle Gouillart
- 3 Joonas Sillanpää
- 3 Paolo Losi
- 2 Charles McCarthy
- 2 Roy Hyunjin Han
- 2 Scott White
- 2 ibayer
- 1 Brandyn White
- 1 Carlos Scheidegger
- 1 Claire Revillet
- 1 Conrad Lee
- 1 Edouard Duchesnay
- 1 Jan Hendrik Metzen
- 1 Meng Xinfan
- 1 Rob Zinkov
- 1 Shiqiao
- 1 Udi Weinsberg
- 1 Virgile Fritsch
- 1 Xinfan Meng
- 1 Yaroslav Halchenko
- 1 jansoe
- 1 Leon Palafox

# Version 0.10¶

**January 11, 2012**

## Changelog¶

- Python 2.5 compatibility was dropped; the minimum Python version needed to use scikit-learn is now 2.6.
- 稀疏逆协方差 estimation using the graph Lasso, with associated cross-validated estimator, by Gael Varoquaux
- New Tree module by Brian Holt, Peter Prettenhofer, Satrajit Ghosh and Gilles Louppe. The module comes with complete documentation and examples.
- Fixed a bug in the RFE module by Gilles Louppe (issue #378).
- Fixed a memory leak in 支持向量机(Support Vector Machines) module by Brian Holt (issue #367).
- Faster tests by Fabian Pedregosa and others.
- Silhouette Coefficient cluster analysis evaluation metric added as
`sklearn.metrics.silhouette_score`

by Robert Layton. - Fixed a bug in K-均值 in the handling of the
`n_init`

parameter: the clustering algorithm used to be run`n_init`

times but the last solution was retained instead of the best solution by Olivier Grisel. - Minor refactoring in 随机梯度下降(Stochastic Gradient Descent) module; consolidated dense and sparse predict methods; Enhanced test time performance by converting model parameters to fortran-style arrays after fitting (only multi-class).
- Adjusted Mutual Information metric added as
`sklearn.metrics.adjusted_mutual_info_score`

by Robert Layton. - Models like SVC/SVR/LinearSVC/LogisticRegression from libsvm/liblinear now support scaling of C regularization parameter by the number of samples by Alexandre Gramfort.
- New Ensemble Methods module by Gilles Louppe and Brian Holt. The module comes with the random forest algorithm and the extra-trees method, along with documentation and examples.
- 新奇点和孤立点检测(Novelty and Outlier Detection): outlier and novelty detection, by :user:`Virgile Fritsch <VirgileFritsch>`.
- 核逼近(Kernel Approximation): a transform implementing kernel approximation for fast SGD on non-linear kernels by Andreas Müller.
- Fixed a bug due to atom swapping in 正交匹配追踪法 (OMP) by Vlad Niculae.
- 用预先计算好的字典进行稀疏编码 by Vlad Niculae.
- 小批量 K-均值 performance improvements by Olivier Grisel.
- K-均值 support for sparse matrices by Mathieu Blondel.
- Improved documentation for developers and for the
`sklearn.utils`

module, by Jake Vanderplas. - Vectorized 20newsgroups dataset loader
(
`sklearn.datasets.fetch_20newsgroups_vectorized`

) by Mathieu Blondel. - 多类和多标签算法(Multiclass and multilabel algorithms) by Lars Buitinck.
- Utilities for fast computation of mean and variance for sparse matrices by Mathieu Blondel.
- Make
`sklearn.preprocessing.scale`

and`sklearn.preprocessing.Scaler`

work on sparse matrices by Olivier Grisel - Feature importances using decision trees and/or forest of trees, by Gilles Louppe.
- Parallel implementation of forests of randomized trees by Gilles Louppe.
`sklearn.cross_validation.ShuffleSplit`

can subsample the train sets as well as the test sets by Olivier Grisel.- Errors in the build of the documentation fixed by Andreas Müller.

## API changes summary¶

Here are the code migration instructions when upgrading from scikit-learn version 0.9:

Some estimators that may overwrite their inputs to save memory previously had

`overwrite_`

parameters; these have been replaced with`copy_`

parameters with exactly the opposite meaning.This particularly affects some of the estimators in

`linear_model`

. The default behavior is still to copy everything passed in.The SVMlight dataset loader

`sklearn.datasets.load_svmlight_file`

no longer supports loading two files at once; use`load_svmlight_files`

instead. Also, the (unused)`buffer_mb`

parameter is gone.Sparse estimators in the 随机梯度下降(Stochastic Gradient Descent) module use dense parameter vector

`coef_`

instead of`sparse_coef_`

. This significantly improves test time performance.The 协方差估计(Covariance estimation) module now has a robust estimator of covariance, the Minimum Covariance Determinant estimator.

Cluster evaluation metrics in

`metrics.cluster`

have been refactored but the changes are backwards compatible. They have been moved to the`metrics.cluster.supervised`

, along with`metrics.cluster.unsupervised`

which contains the Silhouette Coefficient.The

`permutation_test_score`

function now behaves the same way as`cross_val_score`

(i.e. uses the mean score across the folds.)Cross Validation generators now use integer indices (

`indices=True`

) by default instead of boolean masks. This make it more intuitive to use with sparse matrix data.The functions used for sparse coding,

`sparse_encode`

and`sparse_encode_parallel`

have been combined into`sklearn.decomposition.sparse_encode`

, and the shapes of the arrays have been transposed for consistency with the matrix factorization setting, as opposed to the regression setting.Fixed an off-by-one error in the SVMlight/LibSVM file format handling; files generated using

`sklearn.datasets.dump_svmlight_file`

should be re-generated. (They should continue to work, but accidentally had one extra column of zeros prepended.)`BaseDictionaryLearning`

class replaced by`SparseCodingMixin`

.`sklearn.utils.extmath.fast_svd`

has been renamed`sklearn.utils.extmath.randomized_svd`

and the default oversampling is now fixed to 10 additional random vectors instead of doubling the number of components to extract. The new behavior follows the reference paper.

## People¶

The following people contributed to scikit-learn since last release:

- 246 Andreas Müller
- 242 Olivier Grisel
- 220 Gilles Louppe
- 183 Brian Holt
- 166 Gael Varoquaux
- 144 Lars Buitinck
- 73 Vlad Niculae
- 65 Peter Prettenhofer
- 64 Fabian Pedregosa
- 60 Robert Layton
- 55 Mathieu Blondel
- 52 Jake Vanderplas
- 44 Noel Dawe
- 38 Alexandre Gramfort
- 24 :user:`Virgile Fritsch <VirgileFritsch>`
- 23 Satrajit Ghosh
- 3 Jan Hendrik Metzen
- 3 Kenneth C. Arnold
- 3 Shiqiao Du
- 3 Tim Sheerman-Chase
- 3 Yaroslav Halchenko
- 2 Bala Subrahmanyam Varanasi
- 2 DraXus
- 2 Michael Eickenberg
- 1 Bogdan Trach
- 1 Félix-Antoine Fortin
- 1 Juan Manuel Caicedo Carvajal
- 1 Nelle Varoquaux
- 1 Nicolas Pinto
- 1 Tiziano Zito
- 1 Xinfan Meng

# Version 0.9¶

**September 21, 2011**

scikit-learn 0.9 was released on September 2011, three months after the 0.8 release and includes the new modules 流形学习(Manifold learning), 狄利克雷过程(The Dirichlet Process) as well as several new algorithms and documentation improvements.

This release also includes the dictionary-learning work developed by Vlad Niculae as part of the Google Summer of Code program.

## Changelog¶

- New 流形学习(Manifold learning) module by Jake Vanderplas and Fabian Pedregosa.
- New Dirichlet Process Gaussian Mixture Model by Alexandre Passos
- 最近邻方法(Nearest Neighbors) module refactoring by Jake Vanderplas : general refactoring, support for sparse matrices in input, speed and documentation improvements. See the next section for a full list of API changes.
- Improvements on the 特征选择(Feature selection) module by Gilles Louppe : refactoring of the RFE classes, documentation rewrite, increased efficiency and minor API changes.
- 稀疏主成分分析 (SparsePCA 和 MiniBatchSparsePCA) by Vlad Niculae, Gael Varoquaux and Alexandre Gramfort
- Printing an estimator now behaves independently of architectures and Python version thanks to :user:`Jean Kossaifi <JeanKossaifi>`.
- Loader for libsvm/svmlight format by Mathieu Blondel and Lars Buitinck
- Documentation improvements: thumbnails in example gallery by Fabian Pedregosa.
- Important bugfixes in 支持向量机(Support Vector Machines) module (segfaults, bad performance) by Fabian Pedregosa.
- Added 多项分布朴素贝叶斯 and 伯努利朴素贝叶斯 by Lars Buitinck
- Text feature extraction optimizations by Lars Buitinck
- Chi-Square feature selection
(
`feature_selection.univariate_selection.chi2`

) by Lars Buitinck. - 人工合成的数据集 module refactoring by Gilles Louppe
- 多类和多标签算法(Multiclass and multilabel algorithms) by Mathieu Blondel
- Ball tree rewrite by Jake Vanderplas
- Implementation of DBSCAN algorithm by Robert Layton
- Kmeans predict and transform by Robert Layton
- Preprocessing module refactoring by Olivier Grisel
- Faster mean shift by Conrad Lee
- New
`Bootstrap`

, 随机置换交叉验证,也就是: 洗牌 & 分割 and various other improvements in cross validation schemes by Olivier Grisel and Gael Varoquaux - Adjusted Rand index and V-Measure clustering evaluation metrics by Olivier Grisel
- Added
`Orthogonal Matching Pursuit`

by Vlad Niculae - Added 2D-patch extractor utilities in the 特征提取(Feature extraction) module by Vlad Niculae
- Implementation of
`linear_model.LassoLarsCV`

(cross-validated Lasso solver using the Lars algorithm) and`linear_model.LassoLarsIC`

(BIC/AIC model selection in Lars) by Gael Varoquaux and Alexandre Gramfort - Scalability improvements to
`metrics.roc_curve`

by Olivier Hervieu - Distance helper functions
`metrics.pairwise.pairwise_distances`

and`metrics.pairwise.pairwise_kernels`

by Robert Layton `Mini-Batch K-Means`

by Nelle Varoquaux and Peter Prettenhofer.- mldata utilities by Pietro Berkes.
- olivetti_faces by David Warde-Farley.

## API changes summary¶

Here are the code migration instructions when upgrading from scikit-learn version 0.8:

The

`scikits.learn`

package was renamed`sklearn`

. There is still a`scikits.learn`

package alias for backward compatibility.Third-party projects with a dependency on scikit-learn 0.9+ should upgrade their codebase. For instance, under Linux / MacOSX just run (make a backup first!):

find -name "*.py" | xargs sed -i 's/\bscikits.learn\b/sklearn/g'

Estimators no longer accept model parameters as

`fit`

arguments: instead all parameters must be only be passed as constructor arguments or using the now public`set_params`

method inherited from`base.BaseEstimator`

.Some estimators can still accept keyword arguments on the

`fit`

but this is restricted to data-dependent values (e.g. a Gram matrix or an affinity matrix that are precomputed from the`X`

data matrix.The

`cross_val`

package has been renamed to`cross_validation`

although there is also a`cross_val`

package alias in place for backward compatibility.Third-party projects with a dependency on scikit-learn 0.9+ should upgrade their codebase. For instance, under Linux / MacOSX just run (make a backup first!):

find -name "*.py" | xargs sed -i 's/\bcross_val\b/cross_validation/g'

The

`score_func`

argument of the`sklearn.cross_validation.cross_val_score`

function is now expected to accept`y_test`

and`y_predicted`

as only arguments for classification and regression tasks or`X_test`

for unsupervised estimators.`gamma`

parameter for support vector machine algorithms is set to`1 / n_features`

by default, instead of`1 / n_samples`

.The

`sklearn.hmm`

has been marked as orphaned: it will be removed from scikit-learn in version 0.11 unless someone steps up to contribute documentation, examples and fix lurking numerical stability issues.`sklearn.neighbors`

has been made into a submodule. The two previously available estimators,`NeighborsClassifier`

and`NeighborsRegressor`

have been marked as deprecated. Their functionality has been divided among five new classes:`NearestNeighbors`

for unsupervised neighbors searches,`KNeighborsClassifier`

&`RadiusNeighborsClassifier`

for supervised classification problems, and`KNeighborsRegressor`

&`RadiusNeighborsRegressor`

for supervised regression problems.`sklearn.ball_tree.BallTree`

has been moved to`sklearn.neighbors.BallTree`

. Using the former will generate a warning.`sklearn.linear_model.LARS()`

and related classes (LassoLARS, LassoLARSCV, etc.) have been renamed to`sklearn.linear_model.Lars()`

.All distance metrics and kernels in

`sklearn.metrics.pairwise`

now have a Y parameter, which by default is None. If not given, the result is the distance (or kernel similarity) between each sample in Y. If given, the result is the pairwise distance (or kernel similarity) between samples in X to Y.`sklearn.metrics.pairwise.l1_distance`

is now called`manhattan_distance`

, and by default returns the pairwise distance. For the component wise distance, set the parameter`sum_over_features`

to`False`

.

Backward compatibility package aliases and other deprecated classes and functions will be removed in version 0.11.

## People¶

38 people contributed to this release.

- 387 Vlad Niculae
- 320 Olivier Grisel
- 192 Lars Buitinck
- 179 Gael Varoquaux
- 168 Fabian Pedregosa (INRIA, Parietal Team)
- 127 Jake Vanderplas
- 120 Mathieu Blondel
- 85 Alexandre Passos
- 67 Alexandre Gramfort
- 57 Peter Prettenhofer
- 56 Gilles Louppe
- 42 Robert Layton
- 38 Nelle Varoquaux
- 32 :user:`Jean Kossaifi <JeanKossaifi>`
- 30 Conrad Lee
- 22 Pietro Berkes
- 18 andy
- 17 David Warde-Farley
- 12 Brian Holt
- 11 Robert
- 8 Amit Aides
- 8 :user:`Virgile Fritsch <VirgileFritsch>`
- 7 Yaroslav Halchenko
- 6 Salvatore Masecchia
- 5 Paolo Losi
- 4 Vincent Schut
- 3 Alexis Metaireau
- 3 Bryan Silverthorn
- 3 Andreas Müller
- 2 Minwoo Jake Lee
- 1 Emmanuelle Gouillart
- 1 Keith Goodman
- 1 Lucas Wiman
- 1 Nicolas Pinto
- 1 Thouis (Ray) Jones
- 1 Tim Sheerman-Chase

# Version 0.8¶

**May 11, 2011**

scikit-learn 0.8 was released on May 2011, one month after the first “international” scikit-learn coding sprint and is marked by the inclusion of important modules: 层次聚类(Hierarchical clustering), 交叉分解(Cross decomposition), 非负矩阵分解(NMF or NNMF), initial support for Python 3 and by important enhancements and bug fixes.

## Changelog¶

Several new modules where introduced during this release:

- New 层次聚类(Hierarchical clustering) module by Vincent Michel, Bertrand Thirion, Alexandre Gramfort and Gael Varoquaux.
- 核 PCA implementation by Mathieu Blondel
- labeled_faces_in_the_wild by Olivier Grisel.
- New 交叉分解(Cross decomposition) module by Edouard Duchesnay.
- 非负矩阵分解(NMF or NNMF) module Vlad Niculae
- Implementation of the Oracle Approximating Shrinkage algorithm by :user:`Virgile Fritsch <VirgileFritsch>` in the 协方差估计(Covariance estimation) module.

Some other modules benefited from significant improvements or cleanups.

- Initial support for Python 3: builds and imports cleanly, some modules are usable while others have failing tests by Fabian Pedregosa.
`decomposition.PCA`

is now usable from the Pipeline object by Olivier Grisel.- Guide How to optimize for speed by Olivier Grisel.
- Fixes for memory leaks in libsvm bindings, 64-bit safer BallTree by Lars Buitinck.
- bug and style fixing in K-均值 algorithm by Jan Schlüter.
- Add attribute converged to Gaussian Mixture Models by Vincent Schut.
- Implemented
`transform`

,`predict_log_proba`

in`discriminant_analysis.LinearDiscriminantAnalysis`

By Mathieu Blondel. - Refactoring in the 支持向量机(Support Vector Machines) module and bug fixes by Fabian Pedregosa, Gael Varoquaux and Amit Aides.
- Refactored SGD module (removed code duplication, better variable naming), added interface for sample weight by Peter Prettenhofer.
- Wrapped BallTree with Cython by Thouis (Ray) Jones.
- Added function
`svm.l1_min_c`

by Paolo Losi. - Typos, doc style, etc. by Yaroslav Halchenko, Gael Varoquaux, Olivier Grisel, Yann Malet, Nicolas Pinto, Lars Buitinck and Fabian Pedregosa.

## People¶

People that made this release possible preceded by number of commits:

- 159 Olivier Grisel
- 96 Gael Varoquaux
- 96 Vlad Niculae
- 94 Fabian Pedregosa
- 36 Alexandre Gramfort
- 32 Paolo Losi
- 31 Edouard Duchesnay
- 30 Mathieu Blondel
- 25 Peter Prettenhofer
- 22 Nicolas Pinto
- 11 :user:`Virgile Fritsch <VirgileFritsch>`
- 7 Lars Buitinck
- 6 Vincent Michel
- 5 Bertrand Thirion
- 4 Thouis (Ray) Jones
- 4 Vincent Schut
- 3 Jan Schlüter
- 2 Julien Miotte
- 2 Matthieu Perrot
- 2 Yann Malet
- 2 Yaroslav Halchenko
- 1 Amit Aides
- 1 Andreas Müller
- 1 Feth Arezki
- 1 Meng Xinfan

# Version 0.7¶

**March 2, 2011**

scikit-learn 0.7 was released in March 2011, roughly three months after the 0.6 release. This release is marked by the speed improvements in existing algorithms like k-Nearest Neighbors and K-Means algorithm and by the inclusion of an efficient algorithm for computing the Ridge Generalized Cross Validation solution. Unlike the preceding release, no new modules where added to this release.

## Changelog¶

- Performance improvements for Gaussian Mixture Model sampling [Jan Schlüter].
- Implementation of efficient leave-one-out cross-validated Ridge in
`linear_model.RidgeCV`

[Mathieu Blondel] - Better handling of collinearity and early stopping in
`linear_model.lars_path`

[Alexandre Gramfort and Fabian Pedregosa]. - Fixes for liblinear ordering of labels and sign of coefficients [Dan Yamins, Paolo Losi, Mathieu Blondel and Fabian Pedregosa].
- Performance improvements for Nearest Neighbors algorithm in high-dimensional spaces [Fabian Pedregosa].
- Performance improvements for
`cluster.KMeans`

[Gael Varoquaux and James Bergstra]. - Sanity checks for SVM-based classes [Mathieu Blondel].
- Refactoring of
`neighbors.NeighborsClassifier`

and`neighbors.kneighbors_graph`

: added different algorithms for the k-Nearest Neighbor Search and implemented a more stable algorithm for finding barycenter weights. Also added some developer documentation for this module, see notes_neighbors for more information [Fabian Pedregosa]. - Documentation improvements: Added
`pca.RandomizedPCA`

and`linear_model.LogisticRegression`

to the class reference. Also added references of matrices used for clustering and other fixes [Gael Varoquaux, Fabian Pedregosa, Mathieu Blondel, Olivier Grisel, Virgile Fritsch , Emmanuelle Gouillart] - Binded decision_function in classes that make use of liblinear,
dense and sparse variants, like
`svm.LinearSVC`

or`linear_model.LogisticRegression`

[Fabian Pedregosa]. - Performance and API improvements to
`metrics.euclidean_distances`

and to`pca.RandomizedPCA`

[James Bergstra]. - Fix compilation issues under NetBSD [Kamel Ibn Hassen Derouiche]
- Allow input sequences of different lengths in
`hmm.GaussianHMM`

[Ron Weiss]. - Fix bug in affinity propagation caused by incorrect indexing [Xinfan Meng]

## People¶

People that made this release possible preceded by number of commits:

- 85 Fabian Pedregosa
- 67 Mathieu Blondel
- 20 Alexandre Gramfort
- 19 James Bergstra
- 14 Dan Yamins
- 13 Olivier Grisel
- 12 Gael Varoquaux
- 4 Edouard Duchesnay
- 4 Ron Weiss
- 2 Satrajit Ghosh
- 2 Vincent Dubourg
- 1 Emmanuelle Gouillart
- 1 Kamel Ibn Hassen Derouiche
- 1 Paolo Losi
- 1 VirgileFritsch
- 1 Yaroslav Halchenko
- 1 Xinfan Meng

# Version 0.6¶

**December 21, 2010**

scikit-learn 0.6 was released on December 2010. It is marked by the inclusion of several new modules and a general renaming of old ones. It is also marked by the inclusion of new example, including applications to real-world datasets.

## Changelog¶

- New stochastic gradient descent module by Peter Prettenhofer. The module comes with complete documentation and examples.
- Improved svm module: memory consumption has been reduced by 50%, heuristic to automatically set class weights, possibility to assign weights to samples (see SVM: Weighted samples for an example).
- New 高斯过程(Gaussian Processes) module by Vincent Dubourg. This module also has great documentation and some very neat examples. See example_gaussian_process_plot_gp_regression.py or example_gaussian_process_plot_gp_probabilistic_classification_after_regression.py for a taste of what can be done.
- It is now possible to use liblinear’s Multi-class SVC (option
multi_class in
`svm.LinearSVC`

) - New features and performance improvements of text feature extraction.
- Improved sparse matrix support, both in main classes
(
`grid_search.GridSearchCV`

) as in modules sklearn.svm.sparse and sklearn.linear_model.sparse. - Lots of cool new examples and a new section that uses real-world datasets was created. These include: Faces recognition example using eigenfaces and SVMs, Species distribution modeling, Libsvm GUI, Wikipedia principal eigenvector and others.
- Faster 最小角回归 algorithm. It is now 2x faster than the R version on worst case and up to 10x times faster on some cases.
- Faster coordinate descent algorithm. In particular, the full path
version of lasso (
`linear_model.lasso_path`

) is more than 200x times faster than before. - It is now possible to get probability estimates from a
`linear_model.LogisticRegression`

model. - module renaming: the glm module has been renamed to linear_model, the gmm module has been included into the more general mixture model and the sgd module has been included in linear_model.
- Lots of bug fixes and documentation improvements.

## People¶

People that made this release possible preceded by number of commits:

- 207 Olivier Grisel
- 167 Fabian Pedregosa
- 97 Peter Prettenhofer
- 68 Alexandre Gramfort
- 59 Mathieu Blondel
- 55 Gael Varoquaux
- 33 Vincent Dubourg
- 21 Ron Weiss
- 9 Bertrand Thirion
- 3 Alexandre Passos
- 3 Anne-Laure Fouque
- 2 Ronan Amicel
- 1 Christian Osendorfer

# Version 0.5¶

**October 11, 2010**

## Changelog¶

## New classes¶

- Support for sparse matrices in some classifiers of modules
`svm`

and`linear_model`

(see`svm.sparse.SVC`

,`svm.sparse.SVR`

,`svm.sparse.LinearSVC`

,`linear_model.sparse.Lasso`

,`linear_model.sparse.ElasticNet`

) - New
`pipeline.Pipeline`

object to compose different estimators. - Recursive Feature Elimination routines in module 特征选择(Feature selection).
- Addition of various classes capable of cross validation in the
linear_model module (
`linear_model.LassoCV`

,`linear_model.ElasticNetCV`

, etc.). - New, more efficient LARS algorithm implementation. The Lasso
variant of the algorithm is also implemented. See
`linear_model.lars_path`

,`linear_model.Lars`

and`linear_model.LassoLars`

. - New Hidden Markov Models module (see classes
`hmm.GaussianHMM`

,`hmm.MultinomialHMM`

,`hmm.GMMHMM`

) - New module feature_extraction (see class reference)
- New FastICA algorithm in module sklearn.fastica

## Documentation¶

- Improved documentation for many modules, now separating narrative documentation from the class reference. As an example, see documentation for the SVM module and the complete class reference.

## Fixes¶

- API changes: adhere variable names to PEP-8, give more meaningful names.
- Fixes for svm module to run on a shared memory context (multiprocessing).
- It is again possible to generate latex (and thus PDF) from the sphinx docs.

## Examples¶

- new examples using some of the mlcomp datasets:
`sphx_glr_auto_examples_mlcomp_sparse_document_classification.py`

(since removed) and Classification of text documents using sparse features - Many more examples. See here the full list of examples.

## External dependencies¶

- Joblib is now a dependency of this package, although it is shipped with (sklearn.externals.joblib).

## Removed modules¶

- Module ann (Artificial Neural Networks) has been removed from the distribution. Users wanting this sort of algorithms should take a look into pybrain.

## Misc¶

- New sphinx theme for the web page.

## Authors¶

The following is a list of authors for this release, preceded by number of commits:

- 262 Fabian Pedregosa
- 240 Gael Varoquaux
- 149 Alexandre Gramfort
- 116 Olivier Grisel
- 40 Vincent Michel
- 38 Ron Weiss
- 23 Matthieu Perrot
- 10 Bertrand Thirion
- 7 Yaroslav Halchenko
- 9 VirgileFritsch
- 6 Edouard Duchesnay
- 4 Mathieu Blondel
- 1 Ariel Rokem
- 1 Matthieu Brucher

# Version 0.4¶

**August 26, 2010**

## Changelog¶

Major changes in this release include:

- Coordinate Descent algorithm (Lasso, ElasticNet) refactoring & speed improvements (roughly 100x times faster).
- Coordinate Descent Refactoring (and bug fixing) for consistency with R’s package GLMNET.
- New metrics module.
- New GMM module contributed by Ron Weiss.
- Implementation of the LARS algorithm (without Lasso variant for now).
- feature_selection module redesign.
- Migration to GIT as version control system.
- Removal of obsolete attrselect module.
- Rename of private compiled extensions (added underscore).
- Removal of legacy unmaintained code.
- Documentation improvements (both docstring and rst).
- Improvement of the build system to (optionally) link with MKL. Also, provide a lite BLAS implementation in case no system-wide BLAS is found.
- Lots of new examples.
- Many, many bug fixes …

## Authors¶

The committer list for this release is the following (preceded by number of commits):

- 143 Fabian Pedregosa
- 35 Alexandre Gramfort
- 34 Olivier Grisel
- 11 Gael Varoquaux
- 5 Yaroslav Halchenko
- 2 Vincent Michel
- 1 Chris Filo Gorgolewski

# Earlier versions¶

Earlier versions included contributions by Fred Mailhot, David Cooke, David Huard, Dave Morrill, Ed Schofield, Travis Oliphant, Pearu Peterson.