[scikit-learn] any interest in incorporating a new Transformer?
Michael Capizzi
mcapizzi at email.arizona.edu
Thu Aug 17 14:46:50 EDT 2017
Hi all -
Forgive me if this is the wrong place for posting this question, but I'd
like to inquire about the community's interest in incorporating a new
Transformer into the code base.
This paper ( https://nlp.stanford.edu/pubs/sidaw12_simple_sentiment.pdf )
is a "classic" in Natural Language Processing and is often times used as a
very competitive baseline. TL;DR it transforms a traditional count-based
feature space into the conditional probabilities of a `Naive Bayes`
classifier. These transformed features can then be used to train any
linear classifier. The paper focuses on `SVM`.
The attached notebook has an example of the custom `Transformer` I built
along with a custom `Classifier` to utilize this `Transformer` in a
`multiclass` case (as the feature space transformation differs depending on
the label).
If there is interest in the community for the inclusion of this
`Transformer` and `Classifier`, I'd happily go through the official process
of a `pull-request`, etc.
-Michael
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mail.python.org/pipermail/scikit-learn/attachments/20170817/2cdb076a/attachment-0002.html>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mail.python.org/pipermail/scikit-learn/attachments/20170817/2cdb076a/attachment-0003.html>
More information about the scikit-learn
mailing list