<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
<HTML><HEAD><TITLE></TITLE>
<META http-equiv=Content-Type content="text/html; charset=us-ascii">
<META content="MSHTML 6.00.6000.16544" name=GENERATOR></HEAD>
<BODY text=#333399 bgColor=#ffffff>
<DIV dir=ltr align=left><SPAN class=877234914-10122007><FONT face=Arial
color=#0000ff size=2>Sounds like there's a research paper in here somewhere,
should anyone want one. I suspect that I'm one of a large majority of contented
SpamBayes users, but that's hard to know. I'm quite happy with the filtering
obtained by training SpamBayes using what seems to me an easy and intuitive
approach: if SpamBayes puts a message in the wrong place, I move it where it
belongs.</FONT></SPAN></DIV>
<DIV dir=ltr align=left><SPAN class=877234914-10122007><FONT face=Arial
color=#0000ff size=2></FONT></SPAN> </DIV>
<DIV dir=ltr align=left><SPAN class=877234914-10122007><FONT face=Arial
color=#0000ff size=2>But there are some number of people who get results that
leave them frustrated and annoyed. It would be interesting to know why that is.
Different data? Different training methods? Different expectations? Something
else altogether?</FONT></SPAN></DIV>
<DIV dir=ltr align=left><SPAN class=877234914-10122007><FONT face=Arial
color=#0000ff size=2></FONT></SPAN> </DIV>
<DIV dir=ltr align=left><SPAN class=877234914-10122007><FONT face=Arial
color=#0000ff size=2>Maybe it's an indictment of the open source process that no
one has answered these questions. There's no one to commission the research,
just a few good souls who had a need, saw a possible solution, scratched their
own itch, and kindly made the resulting software available for the rest of us.
On the other hand, maybe it's just a matter of time, and Pete or Thomas or
someone else will have the interest and resources to puzzle out an explanation
for the wide range of experiences.</FONT></SPAN></DIV><BR>
<DIV class=OutlookMessageHeader lang=en-us dir=ltr align=left>
<HR tabIndex=-1>
<FONT face=Tahoma size=2><B>From:</B> spambayes-bounces@python.org
[mailto:spambayes-bounces@python.org] <B>On Behalf Of </B>Audiography
Support<BR><B>Sent:</B> Saturday, December 08, 2007 7:08 PM<BR><B>Cc:</B>
spambayes@python.org<BR><B>Subject:</B> Re: [Spambayes] Leaving for another
tool.<BR></FONT><BR></DIV>
<DIV></DIV>Skip, I understand your concern - too much data is sometimes worse
than not enough. And I'm not a mathematician by any stretch of the imagination.
Programmer, yes, hardware designer, yes, PC guru, yes (in another life), Geek of
the week 5 times running, yes, but not mathematician.<BR><BR>But if I start
using a tool that looks for certain words, patterns, phrases, and so on in
messages in order to identify similar messages in the future, then it's really
counter-intuitive to say "don't train it so much, you'll break it." Why
not?<BR><BR>The problem I've run into (every day now) is that no matter how much
or little training I used, SB/TB doesn't work the way everyone tells me a
bayesian filter is supposed to. And when I try to give examples, people say "oh,
that's not the way it works!". What way does it work, then? I'm either training
it too much, or not enough!<BR><BR>If I look at the scores given by Spambayes
and Thunderbayes to words like penis, viagra, etc, they are ridiculously low
compared to words like "sell" or "friend" or "the" - when those 'bad' words
don't occur - <I>ever - </I>in good messages! Never, ever, ever. So the
technology should be able to look at those words and say "hey, this word only
ever appears in bad messages, so I'm going to weight it like hell and mark it as
bad!", without me needing to train it. That didn't happen.<BR><BR>By way of
example (and this probably says as much about my lack of mathematical knowledge
as anything else!):<BR><BR>I had a "penis" list of messages. These messages all
contained the word "penis". I had 1,000 of these messages, which typically
consisted of 20-150 words each, with a lexical word base of about 150 words.
Then, I had a list of identical messages, with the word "penis" taken out. So I
reset SB (completely), then trained it on the bad messages as spam, and the
"edited" messages as ham. When I fed it a new message, identical to another of
the previously trained bad messages, it scored it as 8%. How in the name of the
Oxford Abridged Dictionary can that be calculated as right?<BR><BR>I admit
this isn't the way SB should be used. But if it can't distinguish such a
fundamentally simple concept - out of 150 words, 149 are common, and the one
"uncommon" word - which appeared 1,000 times only in bad messages - was weighted
at about 1 chance in 16 of being "bad"? And the word "yesterday" was weighted at
22%. That's the kind of number that makes me say "this tool doesn't work". More
training doesn't fix the problems, but I thought it would. That's my
mistake!<BR><BR>OK, there are other considerations, such as the mail headers,
character formatting, misspellings, character sets, and so on - but there is no
easy (or difficult, for that matter) to configure ANY bayesian filter I've EVER
seen to work the way I need it. I can't seem to tell it to ignore headers - and
I know for a fact that header data is taken into account when training - but
since so much of the header data (taken as individual words) is
<B>identical</B>, where's the tool to let me tell SB not to score
"x-Mozilla-Status" or "Envelope-to" in every single damn piece of email I get?
But if I look at the training dbs (assuming I can find them and see what's in
them), I find those same terms are weighted/scored just as high as "bad"
words!<BR><BR> So what I'm narked off about (and don't misunderstand me,
I'm absolutely frustrated as hell at wasting close to 100 hours with installing,
training, resetting uninstalling, reinstalling, retraining, and then rescuing
email from the junk pile and simultaneously manually putting spam into the
training folder or the junk folder, and repeat) is the fact that the
<B>technology </B>doesn't work the way it's supposed to. Not for me, anyway. And
I'm a poster boy for nerds, believe me.<BR><BR>Maybe the problem is SB is trying
to be all things to all users. But what I've learned from the last 5 months is
that the SB and TB tools are not easy or friendly or configurable. They're
difficult to get going, difficult to maintain, unable to be truly configured to
specific needs (that are actually common across every smtp/pop email client on
every OS), and they don't like not being trained almost as much as they hate
being over trained!<BR><BR>The one time I tried to "get into" a spam table
(after exporting it to XML and re-importing it with some words weighted more
heavily), it completely broke TB - but not because of the weighting I used, it
was because the XML import filter doesn't actually import valid XML, and the
export XML filter doesn't export valid XML.<BR><BR>Yep, I'm sticking to what
works for me until I have enough free time to try and add something to the bayes
community. I really do want to try and improve the bayes filter technology,
because it should work better.<BR><BR>Thomas Hruska wrote:
<BLOCKQUOTE cite=mid:475AD7B7.4070708@cubiclesoft.com type="cite"><A
class=moz-txt-link-abbreviated href="mailto:skip@pobox.com">skip@pobox.com</A>
wrote: <BR>
<BLOCKQUOTE type="cite"> Pete> Way too many false
negatives (still running at around 7%, after <BR> Pete>
13,000+ spam training messages and 50,000+ good training
<BR> Pete> messages), <BR>Way too large a
database. Train on just mistakes and unsures. If you've
<BR>trained on over 60,000 messages you must be training on everything you
<BR>receive. <BR><BR>Good luck with K9. Sounds like it's doing the
trick. <BR><BR>Skip <BR></BLOCKQUOTE><BR>This only proves that Spambayes needs
an autobalancing ham/spam feature built in by default. Users train on
everything in the hopes of eliminating all spam from the in-box. Also,
by _your_ logic, the default training mechanism in Spambayes should be to NOT
train on spam. VERY counter-intuitive. <BR><BR>In terms of usability,
Spambayes is clearly designed "by geeks, for geeks" but since this tool has
appeared in major computing magazines that _users_ read, the tool needs to
change to fit the mindset of those who will actually use the product.
What Pete said has crossed my mind quite frequently while using the
tool. Your focus is on "training database size" rather than the user's
actual complaint: That the product is not _usable_. I can use and
understand the product only because I'm a geek. However, it needs to be
significantly simplified so a user can use it. <BR><BR>Maybe your goal is only
to cater to geeks. If that's the case, you need to state it somewhere at
the top of your homepage and drop the support for the Outlook add-in - at
which point I too will probably stop using the tool because there is no hope
for it...ever, Users will not take the time to learn to use the tool how
you want them to use it. If they see spam, they are going to train it as
spam no matter how large their training database gets. That is how users
think. Developing software is more about psychology than code:
Study the user and code accordingly. <BR><BR>Sorry for the rant. I've
been feeling the same way as Pete and wanted to put what he said into a little
different perspective - perhaps one that you'd understand better. You
completely ignored Pete's very lengthy e-mail on what it means to be a user of
Spambayes from his perspective and instantly focused on the one sentence that
is useless to him but is "comfortable" for you. I'm hoping this helps
craft an improved tool rather than write Pete, me, and other users like us off
as "annoying". I know that you are going to be upset when you read this
but I don't care if you hate me as long as you end up going back and pondering
Pete's e-mail. His words, from just that one e-mail, are capable of
guiding how Spambayes should be developed for the next 5 years.
<BR><BR></BLOCKQUOTE><BR>
<DIV class=moz-signature>-- <BR>
<META content="OpenOffice.org 2.3 (Win32)" name=GENERATOR>
<META content="Peter Naus" name=AUTHOR>
<META content=20071202;16483703 name=CREATED>
<META content="Peter Naus" name=CHANGEDBY>
<META content=20071202;17003064 name=CHANGED>
<STYLE type=text/css>@page {size: 21cm 29.7cm; margin: 2cm; }
P {
        MARGIN-BOTTOM: 0.21cm
}
</STYLE>
<P style="MARGIN-BOTTOM: 0cm"><BR></P>
<P style="MARGIN-BOTTOM: 0cm"><BR></P>
<P style="MARGIN-BOTTOM: 0cm" align=left><FONT size=1><B>Peter
Naus</B></FONT></P>
<P style="MARGIN-BOTTOM: 0cm" align=left><FONT color=#008000><FONT
size=2><I>Audio Engineering Manager</I></FONT></FONT></P>
<P style="MARGIN-BOTTOM: 0cm" align=left><FONT color=#008080><FONT
face="Microsoft Sans Serif"><FONT
size=4><B>Audiography</B></FONT></FONT></FONT></P>
<P style="MARGIN-BOTTOM: 0cm" align=left><FONT face=Arial><FONT size=1>20
Churinga Avenue, Mitcham. Victoria. 3132. Australia.</FONT></FONT></P>
<P style="MARGIN-BOTTOM: 0cm" align=left><FONT color=#ff0000><FONT
face=Arial><FONT size=1><B>Freecall: 1300 78
4576</B></FONT></FONT></FONT></P>
<P style="MARGIN-BOTTOM: 0cm" align=left><B><FONT color=#ff0000><FONT
face=Arial><FONT size=1>Phone:</FONT></FONT></FONT><FONT color=#ff0000>
</FONT><FONT color=#ff0000><FONT face=Arial><FONT size=1>+613
8802-4562</FONT></FONT></FONT></B></P>
<P style="MARGIN-BOTTOM: 0cm" align=left><FONT color=#ff0000><FONT
face=ari><FONT size=1><B>e-mail:</B></FONT></FONT></FONT><FONT
color=#006666><FONT face="Courier New"><FONT
size=1> </FONT></FONT></FONT><A
href="mailto:support@audiography.com.au"><FONT face="Courier New"><FONT
size=1>support@audiography.com.au</FONT></FONT></A></P>
<P style="MARGIN-BOTTOM: 0cm" align=left><FONT color=#ff0000><FONT
face="Courier New"><FONT size=1><B>web:</B></FONT></FONT></FONT><FONT
face="Courier New"><FONT size=1> <A
href="http://www.audiography.com.au/">http://www.audiography.com.au</A></FONT></FONT></P>
<P style="MARGIN-BOTTOM: 0cm"><BR></P></DIV></BODY></HTML>