%!ps-adobe-3.0susandumais.com/sigir00.doc · web viewthe hierarchical model showed 2-4%...

14
Hierarchical Classification of Web Content Susan Dumais Microsoft Research One Microsoft Way Redmond, WA 99802 USA [email protected] Hao Chen Computer Science Division University of California at Berkeley Berkeley, CA 94720-1776 USA [email protected] [email protected] ABSTRACT This paper explores the use of hierarchical structure for classifying a large, heterogeneous collection of web content. The hierarchical structure is initially used to train different second-level classifiers. In the hierarchical case, a model is learned to distinguish a second-level category from other categories within the same top level. In the flat non-hierarchical case, a model distinguishes a second- level category from all other second- level categories. Scoring rules can further take advantage of the hierarchy by considering only second-level categories that exceed a threshold at the top level. We use support vector machine (SVM) classifiers, which have been shown to be efficient and effective for classification, but not previously explored in the context of hierarchical classification. We found small advantages in accuracy for hierarchical models over flat models. For the hierarchical approach, we found the same accuracy using a sequential Boolean decision rule and a multiplicative decision rule. Since the sequential approach is much more efficient, requiring only 14%-16% of the comparisons used in the other approaches, we find it to be a good choice for classifying text into large hierarchical structures. KEYWORDS Text classification, text categorization, classification, support vector machines, machine learning, hierarchical models, web hierarchies INTRODUCTION With the exponential growth of information on the internet and intranets, it is becoming increasingly difficult to find and organize relevant materials. More and more, simple text retrieval systems are being supplemented with structured organizations. Since the 19 th century, librarians have used classification systems like Dewey and Library of Congress subject headings to organize vast amounts of information. More recently, web directories such as Yahoo! and LookSmart have been used to classify web pages. Structured directories support browsing and search, but the manual nature of the directory compiling process makes it difficult to keep pace with the ever increasing amount of information. Our work looks at the use of automatic classification methods to supplement human effort in creating structured knowledge hierarchies.

Upload: others

Post on 27-Jan-2021

1 views

Category:

Documents


0 download

TRANSCRIPT

%!PS-Adobe-3.0

Hierarchical Classification of Web Content

Susan Dumais

Microsoft Research

One Microsoft Way

Redmond, WA 99802 [email protected]

Hao Chen

Computer Science DivisionUniversity of California at Berkeley Berkeley, CA 94720-1776 USA

[email protected]@sims.berkeley.edu

ABSTRACT

This paper explores the use of hierarchical structure for classifying a large, heterogeneous collection of web content. The hierarchical structure is initially used to train different second-level classifiers. In the hierarchical case, a model is learned to distinguish a second-level category from other categories within the same top level. In the flat non-hierarchical case, a model distinguishes a second-level category from all other second-level categories. Scoring rules can further take advantage of the hierarchy by considering only second-level categories that exceed a threshold at the top level.

We use support vector machine (SVM) classifiers, which have been shown to be efficient and effective for classification, but not previously explored in the context of hierarchical classification. We found small advantages in accuracy for hierarchical models over flat models. For the hierarchical approach, we found the same accuracy using a sequential Boolean decision rule and a multiplicative decision rule. Since the sequential approach is much more efficient, requiring only 14%-16% of the comparisons used in the other approaches, we find it to be a good choice for classifying text into large hierarchical structures.

KEYWORDS

Text classification, text categorization, classification, support vector machines, machine learning, hierarchical models, web hierarchies

INTRODUCTION

With the exponential growth of information on the internet and intranets, it is becoming increasingly difficult to find and organize relevant materials. More and more, simple text retrieval systems are being supplemented with structured organizations. Since the 19th century, librarians have used classification systems like Dewey and Library of Congress subject headings to organize vast amounts of information. More recently, web directories such as Yahoo! and LookSmart have been used to classify web pages. Structured directories support browsing and search, but the manual nature of the directory compiling process makes it difficult to keep pace with the ever increasing amount of information. Our work looks at the use of automatic classification methods to supplement human effort in creating structured knowledge hierarchies.

LEAVE BLANK THE LAST 2.5 cm (1”) OF THE LEFT COLUMN ON THE FIRST PAGE FOR THE COPYRIGHT NOTICE.

A wide range of statistical and machine learning techniques have been applied to text categorization, including multivariate regression models [8,22], nearest neighbor classifiers [26], probabilistic Bayesian models [13, 16], decision trees [16], neural networks [22, 25], symbolic rule learning [1, 4], and support vector machines [7,12]. These approaches all depend on having some initial labeled training data from which category models are learned. Once category models are trained, new items can be added with little or no additional human effort.

Although many real world classification systems have complex hierarchical structure (e.g., MeSH, U.S. Patents, Yahoo!, LookSmart), few learning methods capitalize on this structure. Most of the approaches mentioned above ignore hierarchical structure and treat each category or class separately, thus in effect “flattening” the class structure. A separate binary classifier is learned to distinguish each class from all other classes. The binary classifiers can be considered independently, so an item may fall into none, one, or more than one category. Or they can be considered as an m-ary problem, where the best matching category is chosen. Such simple approaches work rather well on small problems, but they are likely to be difficult to train when there are a large number of classes and a very large number of features. By utilizing known hierarchical structure, the classification problem can be decomposed into a set of smaller problems corresponding to hierarchical splits in the tree. Roughly speaking, one first learns to distinguish among classes at the top level, then lower level distinctions are learned only within the appropriate top level of the tree. Each of these sub-problems can be solved much more efficiently, and hopefully more accurately as well.

The use of a hierarchical decomposition of a classification problem allows for efficiencies in both learning and representation. Each sub-problem is smaller than the original problem, and it is sometimes possible to use a much smaller set of features for each [13]. The hierarchical structure can also be used to set the negative set for discriminative training and at classification time to combine information from different levels. In addition, there is some evidence that decomposing the problem can lead to more accurate specialized classifiers. Intuitively, many potentially good features are not useful discriminators in non-hierarchical representations. Imagine a hierarchy with two top-level categories (“Computers” and “Sports”), and three subcategories within each (“Computers/Hardware”, “Computers/Software”, “Computers/Chat”, “Sports/Chat”, “Sports/Soccer”, Sports/Football”). In a non-hierarchical model, a word like “computer” is not very discriminating since it is associated with items in “Computers/Software”, “Computers/Hardware”, and “Computers/Chat”. In a hierarchical model, the word “computer” would be very discriminating at the first level. At the second level more specialized words could be used as features within the top-level “Computer” category. And, the same features could be used at the second level for two different top-level classed (e.g., “chat” might be a useful feature for both the category “Sports/Chat”, and “Computers/Chat”). Informal failure analyses of classification errors for non-hierarchical models support this intuition. Many of the classification errors are for related categories (e.g., a page about “Sports/Soccer” might be confused with “Sports/Football”, thus category specific features should improve accuracy).

Recently several researchers have investigated the use of hierarchies for text classification, with promising results. Our work differs from earlier work in a couple of important respects. First, we test the approach on a large collection of very heterogeneous web content, which we believe is increasingly characteristic of information organization problems. Second, we use a learning model, support vector machine (SVM), that has not previously been explored in the context of hierarchical classification. SVMs have been found to be more accurate for text classification than popular approaches like naïve Bayes, neural nets, and Rocchio [7, 12, 27]. We use a reduced-dimension binary-feature version of the SVM model that is very efficient for both initial learning and real-time classification, thus making it applicable to large dynamic collections. We will briefly review the earlier work and contrast it with ours.

RELATED WORK

Reuters

Much of the previous work on hierarchical methods for text classification uses the Reuters-22173 or Reuters-21578 articles. This is a rather small and tidy collection, and this alone is problematic for understanding how the approaches generalize to larger more complex internet applications. In addition, the Reuters articles are organized into 135 topical categories with no hierarchical structure. To study hierarchical models, researchers have added one level of hierarchical structure manually. Kohler and Sahami [13] generated a small hierarchical subset of Reuters-22173 by identifying labels that tended to subsume other labels (e.g., corn and wheat are subsumed by grain). The largest of their hierarchies consisted of 939 documents organized into 3 top-level and 6 second-level categories. They compared naïve Bayes, and two limited dependency Bayes net classifiers on flat and hierarchical models. Test documents were classified into the hierarchy by first filtering them through the single best matching first level class and then sending them to the appropriate second level. Note that errors made at the first level are not recoverable, so the system has to make k correct classification for a k-level hierarchy. They found advantages for the hierarchical models when a very small number of features (10) were used per class. For larger numbers of features (which will be required in many complex domains) no advantages for the hierarchical models were found.

Weigend et al. [25] also used the Reuters-22173 collection. An exploratory cluster analysis was first used to suggest an implicit hierarchical structure, and this was then verified by human assignments. They created 4 top-level categories (agriculture, energy, foreign exchange, and metals) and a 5th miscellaneous category that was not used for evaluation. The final evaluations were on 37 categories with at least 16 positive training examples. They used a probabilistic approach that frames the learning problem as one of function approximation for the posterior probability of the topic vector given the input vector. They used a neural net architecture and explored several input representations. Information from each level of the hierarchy is combined in a multiplicative fashion, so no hard decisions have to be made except at the leaf nodes. They found a 5% advantage in average precision for the hierarchical representation when using words, but not when using latent semantic indexing (LSI) features.

D’Alessio et al. [6] used the Reuters-21578 articles. The hierarchy they use comes from Hayes and Weinstein’s original Reuters experiments [9]. It consists of 5 meta-category codes (economic indicator, currency, corporate, commodity, energy) that include all but three of the original categories. For their experiments they only considered articles that had a single category tag, and the 37 categories with more than 20 positive training examples. Their model requires hard assignment at each branch of the tree. The hierarchical model showed 2-4% improvements in precision and recall over the flat model, and modifications to the hierarchy led to advantages of 2-9%.

Ng et al. [19] also used a hierarchical version of Reuters that consisted of countries at the top level and two topical levels below that, but they did not compare their hierarchical model against a flat model. While all of this work is encouraging, the Reuters collection is small and very well organized compared with many realistic applications.

MeSH, U.S. patents

Some researchers have investigated text classification in domains that have rich hierarchical taxonomies (e.g., MeSH, IDC codes of diseases, U.S. patent codes). The size and complexity of the medical and patent hierarchies are like those used for web content. These hierarchies were designed for controlled vocabulary tagging and they have been extensively refined over the years. Ruiz and Srinivasan [21] used a hierarchical mixture of experts to classify abstracts within the MeSH sub-category Heart. They learned classifiers at different levels of granularity, with the top-level distinctions serving as “gates” to the lower level “experts”. They found small advantages for the hierarchical approach compared with a flat approach. Larkey [15] compared hierarchical and flat approaches for classifying patents in the Speech Signal Processing sub-category. She found no multilevel algorithms that performed significantly better than a flat one which chooses among all the speech classes.

Web

McCallum, et al. [17] describe some interesting experiments on three hierarchical collections -- Usenet news, the Science sub-category of Yahoo!, and company web pages. The Yahoo! sub-collection is most closely related to our experiments, although it is less diverse since all items come from the same top-level category. They used a probabilistic Bayesian framework (naïve Bayes), and a technique called “shrinkage” to improve parameter estimates for the probability of words given classes. The idea is to smooth the parameter estimate of a node by interpolation with all its parent nodes (given by the hierarchical organization of classes). This is especially useful for classes with small numbers of training examples. For the Usenet collection, the best performance and largest advantages of the hierarchical model over the flat model are observed for large numbers of features (10,000). For the Yahoo! sub-collection, accuracy is low and there is a much smaller difference between two approaches

Mladenic and Grobelnik [18] examined issues of feature selection for hierarchical classification of web content, but they did not compare hierarchical models with flat non-hierarchical models. Chakrabarti et al. [2] also experimented with web content as well as patent and Reuters data. They use hierarchical structure both to select features and to combine class information from multiple levels of the hierarchy. For the non-hierarchical case, they use a simple vector method with term weighting and no feature selection. They find no advantages for the hierarchical approach in Reuters, a small advantage in the patents collection, and did not test the difference for the Yahoo! sub-collection. But, because they used different representations, it is difficult to know whether differences are due to the number of features, feature selection, or the hierarchical classification algorithm.

Our work explores the use of hierarchies for classifying very heterogeneous web content. We use SVMs, which have been found to be efficient and effective for text classification, but not previously explored in the context of hierarchical classification. The efficiency of SVMs for both initial learning and real-time classification make them applicable to large dynamic collections like web content. We use the hierarchical structure for two purposes. First, we train second-level category models using different contrast sets (either within the same top-level category in the hierarchical case, or across all categories in the flat non-hierarchical case). Second, we combine scores from the top- and second-level models using different combination rules, some requiring a threshold to be exceeded at the top level before second level comparisons are made. The sequential approach is especially efficient for large problems, but how it compares in accuracy is not known. All of our models allow for each test item to be assigned to multiple categories.

WEB DATA SET AND APPLICATION

Application – Classifying web search results

We are interested in moving beyond today’s ranked lists of results by organizing web search results into hierarchical categories. HCI researchers have shown usability advantages of structuring content hierarchically [3, 10, 14]. For information as varied and voluminous as the web it is necessary to create these structures automatically. Several groups have explored methods for clustering search results. Zamir and Etzioni [29] grouped web search results using suffix tree clustering, and Hearst and Pedersen [11] used Scatter/Gather to organize and browse search results. While these methods show some promise, the computational complexity of clustering algorithms limits the number of documents that can be processed, and generating names for the resulting categories is a challenge.

The basic idea behind our work is to use classification techniques to automatically organize search results into existing hierarchical structures. Classification models are learned offline using a training set of human-labeled documents and web categories. Classification offers two advantages compared to clustering – run time classification is very efficient, and manually generated category names are easily understood. To support our goal of automatically classifying web search results two constraints are important in understanding the classification problems we studied. First, we used just the short summaries returned from web search engines since it takes too long to retrieve the full text of pages in a networked environment. These automatically generated summaries are much shorter than the texts used in most classification experiments, and they are much noisier than other document surrogates like abstracts that some have worked with. Second, we focused on the top levels of the hierarchy since we believe that many search results can be usefully disambiguated at this level. We also developed an interface that tightly couples search results and category structure, and found large preference and performance advantages for automatically classified search results (see Chen and Dumais [3] for details). For the experiments reported in this paper, we focused on the top two levels of the hierarchy, and used short automatically generated summaries of web pages.

LookSmart Web Directory

For our experiments, we used a large heterogeneous collection of pages from LookSmart’s web directory [30]. At the time we conducted our experiments in May 1999, the collection consisted of 370597 unique pages that had been manually classified into a hierarchy of categories by trained professional web editors. There were a total of 17173 categories organized into a 7-level hierarchy. We focused on the 13 top-level and 150 second-level categories.

Training/Test Selection

We selected a random 16% sample of the pages for our experiments. We picked this sampling proportion to ensure that even the smallest second-level categories would have some examples in our training set. We used 50078 pages for training, and 10024 for testing. Table 1 shows the number of pages in each top level category. Top level categories had between 578 and 11163 training examples, and second-level categories had between 3 and 3141 training examples. The wide range in number of pages in each category reflects the distribution in the LookSmart directory. Pages could be classified into more than one category. On average, pages were classified into 1.20 second-level categories and 1.07 first-level categories.

Category Name

Total

Train

Test

Automotive

4982

578

114

Business & Finance

31599

3508

703

Computers & Internet

46000

5718

1126

Entertainment & Media

88697

11163

2159

Health & Fitness

25380

3500

722

Hobbies & Interests

22959

3227

682

Home & Family

11484

1373

280

People & Chat

35157

3309

682

Reference & Education

58002

5574

1175

Shopping & Services

19667

2122

423

Society & Politics

38968

4855

946

Sports & Recreation

23559

3081

640

Travel & Vacations

43685

5409

1091

Total Unique

370597

50078

10024

Table 1- Training and Test Samples by Category

Pre-processing

A pre-processing module extracted plain text from each web page. In addition, the title, description and keywords fields from the META tag, and the ALT field from the IMG tag were also extracted if they existed because they provide useful descriptions of the web page. If a web page contained frames, text from each frame was extracted.

Since we were interested in classifying web search results, we needed to work with just short summary descriptions of web pages. We automatically generated summaries of each page and used this for evaluating our classification algorithms. Our summaries consisted of the title, the keywords, and either the description tag if it existed or the first 40 words of the body otherwise.

We did minimal additional processing on the text. We translated upper to lower case, used white space and punctuation to separate words, and used a small stop list to omit the most common words. No morphological or syntactic analyses were used.

After preprocessing, a binary vector was created for each page indicating whether each term appeared in the page summary or not. For each category 1000 terms were selected based on the mutual information between the term and category (details below). Unlike many text retrieval applications, term frequency and document length are not taken into account in this representation, because earlier experiments showed good performance with the binary representation for SVMs [7], and this representation improves efficiency.

TEXT CLASSIFICATION USING SVMs

Text classification involves a training phase and a testing phase. During the training phase, a large set of web pages with known category labels are used to train a classifier. An initial model is built using a subset of the labeled data, and a holdout set is used to identify optimal model parameters. During the testing or operational phase, the learned classifier is used to classify new web pages.

A support vector machine (SVM) algorithm was used as the classifier, because it has been shown in previous work to be both very fast and effective for text classification problems [7, 12, 27]. Vapnik introduced SVMs in his work on structural risk minimization [23, 24], and there has been a recent surge of interest in SVM classifiers in the learning community. In its simplest form, a linear SVM is a hyperplane that separates a set of positive examples from a set of negative examples with maximum margin. The margin is the distance from the hyperplane to the nearest of the positive and negative examples.

Figure 1 shows an example of a simple two-dimensional problem that is linearly separable.

Category Name

Best F1

0.841

Health & Fitness/Drugs & Medicines

0.797

Home & Family/Real Estate

0.781

Reference & Education/K-12 Education

0.750

Sports & Recreation/Fishing

0.741

Reference & Education/Higher & Cont. Ed.

Worst F1

0.034

Society & Politics/World Cultures

0.088

Home & Family/For Kids

0.122

Computers & Internet/News & Magazines

0.131

Computers & Internet/Internet & the Web

0.133

Business & Finance/Business Professions

Figure 1 - Graphical Representation of a Linear SVM

In the linearly separable case maximizing the margin can be expressed as an optimization problem:

minimize

2

2

1

w

r

subject to

i

b

x

w

y

i

i

"

³

-

×

,

1

)

(

r

r

where xi is the ith training example and yi is the correct output of the SVM for the ith training example.

In cases where points are not linearly separable, slack variables are introduced that permit, but penalize, points that fall on the wrong side of the decision boundary. For problems that are not linearly separable, kernel methods can be used to transform the input space so that some non-linear problems can be learned. We used the simplest linear form of the SVM because it provided good classification accuracy, and is fast to learn and apply.

Platt [20] developed a very efficient method for learning the SVM weights. His sequential minimal optimization (SMO) algorithm breaks the large quadratic programming (QP) problem down into a series of small QP problems that can be solved analytically. Additional efficiencies can be realized because the training sets used for text classification are sparse and binary. Thus the SMO algorithm is nicely applicable for large feature and training sets.

Once the weights are learned, new items are classified by computing

x

w

r

r

×

, where

w

r

is the vector of learned weights, and

x

r

is the input vector for a new document. With the binary representation we use, this amounts to taking the sum of the weights for features present in a document and this is very efficient.

After training the SVM, we fit a sigmoid to the output of the SVM using regularized maximum likelihood fitting, so that the SVM produces posterior probabilities that are directly comparable across categories.

Feature Selection

For reasons of both efficiency and efficacy, feature selection is often used when applying machine learning methods to text categorization. We reduce the feature space by eliminating words that appear in only a single document (hapax legomena), then selecting the 1000 words with highest mutual information with each category. Yang and Pedersen [28] compared a number of methods for feature selection. We used a mutual information measure (Cover and Thomas [5]) that is similar to their information gain measure. The mutual information MI(F, C) between a feature, F, and a category, C, is defined as:

{

}

{

}

å

å

Î

Î

=

f

f

F

c

c

C

C

P

F

P

C

F

P

C

F

P

C

F

MI

,

,

)

(

)

(

)

,

(

log

)

,

(

)

,

(

We compute the mutual information between every pair of features and categories.

For the non-hierarchical case, we select the 1000 features with the largest MI for each of the 150 categories (vs. all the other categories). For the hierarchical case, we select 1000 features with the largest MI for each of the 13 top-level categories, and also select the 1000 features for each of the 150 second-level categories (vs. the categories that share the same top-level category). We did not rigorously explore the optimum number of features for this problem, but these numbers provided good results on a training validation set so they were used for testing. In all cases, the selected features are used as input to the SVM algorithm that learns the optimum weights for the features,

w

r

. SVM Parameters

In addition to varying the number of features, SVM performance is governed by two parameters, C (the penalty imposed on training examples that fall on the wrong side of the decision boundary), and p (the decision threshold). The default C parameter value (0.01) was used.

The decision threshold, p, can be set to control precision and recall for different tasks. Increasing p, results in fewer test items meeting the criterion, and this usually increases precision but decreases recall. Conversely, decreasing p typically decreases precision but increases recall. We chose p so as to optimize performance on the F1 measure on a training validation set. For the flat non-hierarchical models, p=0.5. For the hierarchical models, p=0.2 for the multiplicative decision rule, and p1=0.2 (first level) and p2=0.5 (second level) for the Boolean decision rule.

RESULTS

As just described decision thresholds were established on a training validation set. For each category, if a test item exceeds the decision threshold, it is judged to be in the category. A test item can be in zero, one, or more than one categories. From this we compute precision (P) and recall (R). These are micro-averaged to weight the contribution of each category by the number of test examples in it. We used the F measure to summarize the effects of both precision and recall. With equal weights assigned to precision and recall, F1 = 2*P*R/(P+R).

For each test example, we compute the probability of it being in each of the 13 top-level categories and each of the 150 second-level categories. Recall that for the non-hierarchical case, models were learned (and features chosen) to distinguish each category from all other categories, and that for the hierarchical case, models were learned (and features chosen) to distinguish each category from only those categories within the same top-level category. We explored two general ways to combine probabilities from the first and second level for the hierarchical approach. In one case we first set a threshold at the top level and only match second-level categories that pass this test – that is we compute a Boolean function P(L1)&&P(L2). In order to be correctly classified, a test instance must satisfy both constraints. This method is quite efficient, since large numbers of second level categories do not need to be tested. In the other case, we compute P(L1)*P(L2). This approach allows matches even though the scores at one level fall below threshold. Although the non-hierarchical models are not trained to use top-level information, we can compute the same probabilities for this case as well.

Learned Features

There are many differences in the learned features for the hierarchical and non-hierarchical models. For example, the feature “sports” is useful in distinguishing the top-level category Sports & Recreation from the other top-level categories. The feature “sports” is also useful for distinguishing between some second-level categories -- it has a high weight for News & Magazines (under Society & Politics), SportingGoods (under Shopping & Services), and Collecting (under Hobbies & Interests). But, the feature “sports” is not selected for any categories using the flat non-hierarchical approach. It is difficult to study this systematically, but the different learning approaches are serving an important role in feature selection and weighting.

F1 Accuracy

Top Level

The overall F1 value for the 13 top-level categories is .572. Classifying short summaries of very heterogeneous web pages is a difficult task, so we did not expect to have exceptionally high accuracy. Performance on the original training set is only .649, so this is a difficult learning task and generalization to the test set is quite reasonable. This level of accuracy is sufficient to support some web search tasks [3], and could also be used as an aid to human indexers. In addition, we know that we can improve the absolute level of performance by 15%-20% using the full text of pages, and by optimizing the C parameter. What is of more interest to us here is the comparative performance of hierarchical and non-hierarchical approaches.

Second Level, Non-Hierarchical (Baseline)

The overall F1 value for the 150 second-level categories, treated as a flat non-hierarchical problem, is .476. There is a drop in performance in going from 13 to 150 categories. Performance for the 150 categories is better than the .364 value reported by McCallum et al. [17], but it is difficult to compare precisely because they use average precision not F1, the categories themselves are different, and they use the full text of pages. The .476 non-hierarchal value will serve as the baseline for looking at the use of hierarchical models.

Performance varies widely across categories. The five categories with highest and lowest F1 scores (based on at least 30 test instances) are shown in Table 2.

The most difficult categories to learn models for are those based, at least in part, on non-content distinctions. It might be useful to learn models for genre (e.g., for kids, magazines, web) in addition to our current use of content-based models.

Second Level, Hierarchical

As described above, we examined both a multiplicative scoring function, P(L1)*P(L2), and a Boolean scoring function, P(L1)&&P(L2). The former allows for matches that fall below individual thresholds, and the latter requires that the threshold be exceeded at every level. Both scoring rules allow items to be classified into multiple classes.

The overall F1 value for the P(L1)*P(L2) scoring function is .495, at the threshold of p=0.20 established on the validation set. This represents a 4% improvement over the baseline flat models, and is statistically significant using a sign test, p<.01. The overall F1 value for the P(L1)&&P(L2) scoring function is .497, at the thresholds of p1=0.20 and p2=0.50 established on the validation set. This again is a small improvement over the non-hierarchical model, and is significantly different than the baseline using a sign test, p<0.01. Somewhat surprisingly, there is no difference between the sequential Boolean rule and the multiplicative scoring rule. The added efficiencies that can be obtained with the hierarchical model and the Boolean rule make it a good overall choice, as we discuss in more detail below.

An upper bound on the performance of these hierarchical models can be obtained by assuming that the top-level classification is correct (i.e., P(L1)=1). In this case, performance is quite good with an F1 value of .711, so there is considerable room for improvement.

Category Name

Total

Train

Test

Automotive

4982

578

114

Business & Finance

31599

3508

703

Computers & Internet

46000

5718

1126

Entertainment & Media

88697

11163

2159

Health & Fitness

25380

3500

722

Hobbies & Interests

22959

3227

682

Home & Family

11484

1373

280

People & Chat

35157

3309

682

Reference & Education

58002

5574

1175

Shopping & Services

19667

2122

423

Society & Politics

38968

4855

946

Sports & Recreation

23559

3081

640

Travel & Vacations

43685

5409

1091

Total Unique

370597

50078

10024

Table 2 - Best and worst F1 score by category

Efficiency

Offline training

As noted earlier, linear SVM models can be learned quite efficiently with Platt’s SMO algorithm. The total training time for all the models described in the paper was only about 30 minutes. Training time for all the non-hierarchical models on 50078 examples in each of 150 categories was 729 CPU seconds. Training time for all 150 hierarchical second-level models was 128 CPU seconds. Training is faster here because the negative set is smaller, including only items in the same top-level category. Training time for the 13 top-level categories was 1258 CPU seconds. All times were computed on a standard 266MHz Pentium II PC running Windows NT.

Online evaluation

At run time we need to take the dot product of the learned weight vector for each category and the feature vector for the test item. The cost of this operation depends on the number of features and the number of categories. Efficiency can be improved if some category comparisons can be omitted without hurting accuracy. As we have seen above, the Boolean decision function accomplishes this with no decrease in classification accuracy. For the non-hierarchical model, each test instance must be compared with all 150 second-level categories. For the hierarchical model with the multiplicative scoring rule, each test instance must be compared to all 13 first-level categories and all 150 second-level categories. For the Boolean scoring rule, each test instance must be compared to all 13 first-level categories but only to second-level categories that pass the first-level criterion. For the top-level threshold value of p=0.20, only 7.4% of the second-level categories need to be examined. Top-level comparisons are still required for the Boolean decision rule, so overall 14.8% of the comparisons used for the multiplicative rule and 16.1% of the comparisons used for the non-hierarchical model are required, both representing a considerable savings at evaluation time.

CONCLUSION

The research described in this paper explores the use of hierarchical structure for classifying a large, heterogeneous collection of web content to support classification of search results. We used SVMs, which have been found to be an efficient and effective learning method for text classification, but not previously explored for hierarchical problems. We use the hierarchical structure for two purposes. First, we train second-level category models using different contrast sets (either within the same top-level category in the hierarchical case, or across all categories in the non-hierarchical case). Second, we combine scores from the top- and second-level models using different combination rules, some requiring a threshold to be exceeded at the top level before comparisons at the second level are made.

We found small advantages in the F1 accuracy score for the hierarchical models, compared with a baseline flat non-hierarchical model. We found no difference between a multiplicative scoring function and a sequential Boolean function for combining scores from the two levels of the hierarchy. Since the sequential Boolean approach is much more efficient, requiring only 14%-16% of the number of comparisons, we find it to be a good choice.

There are a number of interesting directions for future research. We should be able to obtain further advantages in efficiency in the hierarchical approach by reducing the number of features needed to discriminate between categories within the same top-level category. Koller and Sahami [13] have shown that one can dramatically reduce the number of features for simple Reuters categories without decreasing accuracy too much, and this is definitely worth trying with our content and learning algorithm. There are some preliminary indications that this will work in our application. The SVM model assigns a weight of 0 to features that are not predictive (and could thus be omitted). We find a larger number of features with 0 weights in the hierarchical case (137 per 1000) than the non-hierarchical case (85 per 1000). There are also many near-zero weights that contribute little to the overall score. Varying the number of features for further efficiency gains seems promising to try with our content and learning model.

The sequential decision model provides large efficiency gains, so it will be important to see how it generalizes to more than two levels, where error cascading may be more of an issue. There are at least two ways to address the problem. One involves improved classifiers and methods for setting appropriate decisions thresholds at multiple levels. Another approach is through the use of interactive interfaces so that people are involved in making some of the critical decisions. This is a rich space for further exploration.

Our research adds to a growing body of work exploring how hierarchical structures can be used to improve the efficiency and efficacy of text classification. We worked with a large, heterogeneous collection of web content -- a scenario that is increasingly characteristic of the information management tasks that individuals and organizations face. And, we successfully extended SVM models to an application that takes advantage of hierarchical structure, for both category learning and run time efficiencies.

ACKNOWLEDGMENTS

We are grateful to John Platt for help with the Support Vector Machine code, and to four anonymous reviewers for their comments.

REFERENCES

1. Apte, C., Damerau, F. and Weiss, S. Automated learning of decision rules for text categorization. ACM Transactions on Information Systems, 12(3), 233-251, 1994.

2. Chakrabarti, S., Dom, B., Agrawal, R. and Raghavan, P. Scalable feature selection, classification and signature generation for organizing large text databases into hierarchical topic taxonomies. The VLDB Journal 7, 163-178, 1998.

3. Chen, H. and Dumais, S. Bringing order to the web: Automatically categorizing search results. Proceedings of the ACM SIGCHI Conference on Human Factors in Computing Systems (CHI’00), 145-152, 2000.

4. Cohen, W.W. and Singer, Y. Context-sensitive learning methods for text categorization Proceedings of the 19th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’96), 307-315, 1996.

5. Cover, T. and Thomas, J. Elements of Information Theory. Wiley, 1991.

6. D'Alessio, S., Murray, M., Schiaffino, R. and Kershenbaum, A. Category levels in hierarchical text categorization. Proceedings of EMNLP-3, 3rd Conference on Empirical Methods in Natural Language Processing, 1998.

7. Dumais, S. T., Platt, J., Heckerman, D. and Sahami, M. Inductive learning algorithms and representations for text categorization. Proceedings of the Seventh International Conference on Information and Knowledge Management (CIKM’98), 148-155, 1998.

8. Fuhr, N., Hartmanna, S., Lustig, G., Schwantner, M., and Tzeras, K. Air/X – A rule-based multi-stage indexing system for lage subject fields. Proceedings of RIAO’91, 606-623, 1991.

9. Hayes, P.J. and Weinstein, S.P. CONSTRUE: A System for Content-Based Indexing of a Database of News Stories. Second Annual Conference on Innovative Applications of Artificial Intelligence, 1990.

10. Hearst, M., and Karadi, C. Searching and browsing text collections with large category hierarchies. Proceedings of the ACM SIGCHI Conference on Human Factors in Computing Systems (CHI’97), Conference Companion, 1997.

11. Hearst, M. and Pedersen, J. Reexamining the cluster hypothesis: Scatter/Gather on retrieval results. Proceedings of 19th Annual International ACM/SIGIR Conference on Research and Development in Information Retrieval (SIGIR’96), 1996.

12. Joachims, T. Text categorization with support vector machines: Learning with many relevant features. Proceedings of European Conference on Machine Learning (ECML’98), 1998

13. Koller, D. and Sahami, M. 1997. Hierarchically classifying documents using very few words. Proceedings of the Fourteenth International Conference on Machine Learning (ICML’97), 170-178, 1997.

14. Landauer, T., Egan, D., Remde, J., Lesk, M., Lochbaum, C., and Ketchum, D. Enhancing the usability of text through computer delivery and formative evaluation: The SuperBook project. Hypertext – A Psychological Perspective. Ellis Horwood, 1993.

15. Larkey, L. Some issues in the automatic classification of U.S. patents. In Working Notes for the AAAI-98 Workshop on Learning for Text Categorization, 1998.

16. Lewis, D.D. and Ringuette, M.. A comparison of two learning algorithms for text categorization. Third Annual Symposium on Document Analysis and Information Retrieval (SDAIR’94), 81-93, 1994.

17. McCallum, A., Rosenfeld, R., Mitchell, T. and Ng, A. Improving text classification by shrinkage in a hierarchy of classes. Proceedings of the Fifteenth International Conference on Machine Learning, (ICML-98), 359-367, 1998.

18. Mladenic, D. and Grobelnik, M. Feature selection for classification based on text hierarchy. Proceedings of the Workshop on Learning from Text and the Web, 1998.

19. Ng, H.T., Goh, W.B. and Low, K.L, Proceedings of 20th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’97), 67-73, 1997.

20. Platt, J. Fast training of support vector machines using sequential minimal optimization. In Advances in Kernel Methods –Support Vector Learning. B. Schölkopf, C. Burges, and A. Smola, eds., MIT Press, 1999.

21. Ruiz, M.E. and Srinivasan, P. Hierarchical neural networks for text categorization. Proceedings of the 22nd International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’99), 281-282, 1999.

22. Schütze, H., Hull, D. and Pedersen, J.O. A comparison of classifiers and document representations for the routing problem. Proceedings of the 18th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’95), 229-237, 1995.

23. Vapnik, V., Estimation of Dependencies Based on Data [in Russian], Nauka, Moscow, 1979. (English translation: Springer Verlag, 1982.)

24. Vapnik, V., The Nature of Statistical Learning Theory, Springer-Verlag, 1995.

25. Weigend, A.S., Wiener, E.D. and Pedersen, J.O. Exploiting hierarchy in text categorization. Information Retrieval, 1(3), 193-216, 1999.

26. Yang, Y. Expert network: Effective and efficient learning from human decisions in text categorization and retrieval. Proceedings of the 17th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’94), 13-22, 1994.

27. Yang, Y. and Lui, Y. A re-examination of text categorization methods. Proceedings of the 22nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’99), 42-49, 1999.

28. Yang, Y. and Pedersen, J.O. A comparative study on feature selection in text categorization. Proceedings of the Fourteenth International Conference on Machine Learning (ICML’97), 412-420, 1997.

29. Zamir, O. and Etzioni, O. Web document clustering: A feasibility demonstration. Proceedings of the 21st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’98), 46-54, 1998.

30. http://www.looksmart.com

� EMBED Excel.Sheet.8 ���

w

� EMBED Excel.Sheet.8 ���

Space of possible inputs

Maximize distances to nearest points

Negative Examples

Positive Examples

Category Name

Best F1

0.841

Health & Fitness/Drugs & Medicines

0.797

Home & Family/Real Estate

0.781

Reference & Education/K-12 Education

0.750

Sports & Recreation/Fishing

0.741

Reference & Education/Higher & Cont. Ed.

Worst F1

0.034

Society & Politics/World Cultures

0.088

Home & Family/For Kids

0.122

Computers & Internet/News & Magazines

0.131

Computers & Internet/Internet & the Web

0.133

Business & Finance/Business Professions

_956404784.unknown
_1018252115.unknown
_1018262858.unknown
_1018264478.xls

LookSmartStats

Two-Level Analysis

13 top level

150 second level

missed items in 0.7 and 0.8 - redone 7/15/99

sum2 data 01/14/2000

MSLookSmart>250

TopLevelCategory NameTotalTrainTest

NumberNameNtotaltraintestAutomotive4982578114

0.0automotive49825781140.11601766360.0228823766Business & Finance315993508703

0.1business&finance3159935087030.11101617140.0222475395Computers & Internet4600057181126

0.2computers&internet46000571811260.12430434780.0244782609Entertainment & Media88697111632159

0.3entertainment&media886971116321590.12585544040.0243412968Health & Fitness253803500722

0.4health&fitness2538035007220.13790386130.0284475965Hobbies & Interests229593227682

0.5hobbies&interests2295932276820.14055490220.0297051265Home & Family114841373280

0.6home&family1148413732800.11955764540.0243817485People & Chat351573309682

0.7people&chat3515733096820.09412065880.0193986973Reference & Education5800255741175

0.8reference&education58002557411750.09610013450.0202579221Shopping & Services196672122423

0.9shopping&services1966721224230.10789647630.02150811Society & Politics389684855946

0.10society&politics3896848559460.12458940670.0242763293Sports & Recreation235593081640

0.11sports&recreation2355930816400.13077804660.027165839Travel & Vacations4368554091091

0.12travel&vacations43685540910910.12381824420.0249742475Total Unique3705975007810024

total - l14501395341710743

unique3705975007810024

1.06667598551.0717278532

total - l25584911257

1.11524022521.1230047885

SecondLevel0.0.131

0.0.08890.0.240

0.0.1310.0.740

0.0.2400.0.1150

0.0.3810.0.1062

0.0.4156380.0.381

0.0.577260.4.10102

0.0.6162240.7.4144

0.0.7400.3.10204

0.0.82230.4.7203

0.0.92140.7.7205

0.0.10620.7.11204

0.0.11500.0.9214

0.0.1236100.0.8223

0.1.0116300.1.7234

0.1.17561590.7.2258

0.1.23410.8.6259

0.1.3450900.4.9316

0.1.414722950.2.0336

0.1.5146290.6.33313

0.1.6125260.1.2341

0.1.72340.4.1349

0.1.892190.0.123610

0.1.9289550.7.5384

0.1.1096190.7.83914

0.2.03360.11.8394

0.2.113402510.7.04312

0.2.29221740.5.8449

0.2.36051210.12.95113

0.2.48011730.6.25411

0.2.5394770.7.35811

0.2.6481980.12.56016

0.2.7120260.9.106314

0.2.8235510.4.126414

0.2.9142330.6.7647

0.2.10112210.12.76414

0.2.119881960.0.57726

0.3.011312290.6.07812

0.3.17131440.11.11829

0.3.222254020.6.58421

0.3.3285540.3.58519

0.3.4195440.0.0889

0.3.585190.7.18818

0.3.616683150.4.38915

0.3.731416170.1.89219

0.3.8434770.10.59517

0.3.916773370.1.109619

0.3.102040.4.119723

0.4.0275620.10.110323

0.4.13490.9.610515

0.4.219914220.6.610724

0.4.389150.9.211025

0.4.4326580.11.111017

0.4.5244440.5.411125

0.4.6259570.2.1011221

0.4.72030.1.011630

0.4.8468820.2.712026

0.4.93160.9.912121

0.4.101020.9.012219

0.4.1197230.9.412221

0.4.1264140.11.612333

0.5.0152390.1.612526

0.5.1269670.9.312621

0.5.2147380.5.912716

0.5.37461450.11.413933

0.5.4111250.2.914233

0.5.5341790.5.614223

0.5.6142230.1.514629

0.5.7311710.5.214738

0.5.84490.5.015239

0.5.9127160.8.115426

0.5.107541700.0.415638

0.5.11171280.7.615729

0.6.078120.10.216139

0.6.1241530.0.616224

0.6.254110.9.816224

0.6.333130.11.716436

0.6.4395760.5.1117128

0.6.584210.10.917637

0.6.6107240.12.117639

0.6.76470.9.1218029

0.6.8353700.10.1118242

0.7.043120.8.518839

0.7.188180.3.419544

0.7.22580.9.719844

0.7.358110.11.921342

0.7.41440.12.821657

0.7.53840.2.823551

0.7.6157290.9.523561

0.7.72050.12.323944

0.7.839140.6.124153

0.7.925545200.11.024162

0.7.10266530.4.524444

0.7.112040.4.625957

0.8.012362510.10.326256

0.8.1154260.7.1026653

0.8.216623460.5.126967

0.8.3484940.8.727364

0.8.49592070.4.027562

0.8.5188390.3.328554

0.8.62590.1.928955

0.8.7273640.5.731171

0.8.84591100.10.631353

0.8.9321680.8.932168

0.9.0122190.9.132365

0.9.1323650.4.432658

0.9.2110250.12.232850

0.9.3126210.5.534179

0.9.4122210.11.534469

0.9.5235610.11.234770

0.9.6105150.6.835370

0.9.7198440.11.336162

0.9.8162240.10.737075

0.9.9121210.2.539477

0.9.1063140.6.439576

0.9.11424900.9.1142490

0.9.12180290.3.843477

0.10.0465920.1.345090

0.10.1103230.8.8459110

0.10.2161390.10.046592

0.10.3262560.4.846882

0.10.414012560.2.648198

0.10.595170.8.348494

0.10.6313530.2.3605121

0.10.7370750.3.1713144

0.10.87631440.10.10741156

0.10.9176370.5.3746145

0.10.107411560.5.10754170

0.10.11182420.1.1756159

0.11.0241620.10.8763144

0.11.1110170.2.4801173

0.11.2347700.2.2922174

0.11.3361620.8.4959207

0.11.4139330.12.4968204

0.11.5344690.2.11988196

0.11.6123330.11.101075233

0.11.7164360.3.01131229

0.11.83940.8.01236251

0.11.9213420.2.11340251

0.11.1010752330.10.41401256

0.11.118290.1.41472295

0.12.016123320.12.01612332

0.12.1176390.8.21662346

0.12.2328500.3.61668315

0.12.3239440.3.91677337

0.12.49682040.12.61921375

0.12.560160.4.21991422

0.12.619213750.3.22225402

0.12.764140.7.92554520

0.12.8216570.3.73141617

0.12.95113

5584911257

Top-m1000-Sum2.Sum2

SUM2-SUM2

Hier-TopLevel

53k training

>256-1000

default c=0.01

CategTrainTestSMOX 0.5

05007810024TrainTest

Net1Targ1Net1Targ0Net0Targ1Net0Targ0precrecallbeFNet1Targ1Net1Targ0Net0Targ1Net0Targ0precrecallbeF

0.0automotive5781140.0.p0.5.smox.train.sum.sum.out32270256494300.82142857140.55709342560.68926099850.66391752580.0.p0.5.smox.test.sum.sum.out74144098960.84090909090.6491228070.7450159490.7326732673

0.1business&finance35087030.1.p0.5.smox.train.sum.sum.out11844762324460940.7132530120.33751425310.52538363260.45820433440.1.p0.5.smox.test.sum.sum.out2049549992260.68227424750.29018492180.48622958460.4071856287

0.2computers&internet571811260.2.p0.5.smox.train.sum.sum.out28697032849436570.80319148940.50174886320.65247017630.61765339070.2.p0.5.smox.test.sum.sum.out55016257687360.77247191010.48845470690.63046330850.598476605

0.3entertainment&media1116321590.3.p0.5.smox.train.sum.sum.out674913534414375620.83300419650.60458658070.71879538860.70064884510.3.p0.5.smox.test.sum.sum.out125829490175710.81056701030.58267716540.69662208780.6779843708

0.4health&fitness35007220.4.p0.5.smox.train.sum.sum.out20483701452462080.8469809760.58514285710.71606191660.69212571810.4.p0.5.smox.test.sum.sum.out4087931492230.83778234090.56509695290.70143964690.6749379653

0.5hobbies&interests32276820.5.p0.5.smox.train.sum.sum.out9342642293465870.77963272120.28943290980.53453281550.42214689270.5.p0.5.smox.test.sum.sum.out1786750492750.72653061220.26099706740.49376383980.38403452

0.6home&family13732800.6.p0.5.smox.train.sum.sum.out407158966485470.72035398230.29643117260.50839257750.42002063980.6.p0.5.smox.test.sum.sum.out764320497010.63865546220.27142857140.45504201680.380952381

0.7people&chat33096820.7.p0.5.smox.train.sum.sum.out11063532203464160.75805346130.33423995160.54614670650.46392617450.7.p0.5.smox.test.sum.sum.out1998648392560.6982456140.29178885630.49501723520.411582213

0.8reference&education557411750.8.p0.5.smox.train.sum.sum.out25386353036438690.79987393630.455328310.62760112320.58031325030.8.p0.5.smox.test.sum.sum.out51513766087120.78987730060.43829787230.61408758650.5637657362

0.9shopping&services21224230.9.p0.5.smox.train.sum.sum.out4692131653477430.68768328450.22101790760.4543505960.33452211130.9.p0.5.smox.test.sum.sum.out894933495520.64492753620.21040189130.42766471370.3172905526

0.1society&politics48559460.10.p0.5.smox.train.sum.sum.out17255233130447000.76734875440.35530381050.56132628250.48571026330.10.p0.5.smox.test.sum.sum.out27912566789530.69059405940.29492600420.49276003180.4133333333

0.11sports&recreation30816400.11.p0.5.smox.train.sum.sum.out16952911386467060.85347432020.55014605650.70181018840.66903493190.11.p0.5.smox.test.sum.sum.out3048033693040.79166666670.4750.63333333330.59375

0.12travel&vacations540910910.12.p0.5.smox.train.sum.sum.out37354311674442380.8965434470.6905158070.7935296270.7801566580.12.p0.5.smox.test.sum.sum.out7019039088430.88621997470.64252978920.76437488190.7449521785

257815840276365917570.81531260870.48263661380.64897461130.60634069474835132159081182480.78541260560.45006050450.61773655510.5722232085

0.0automotive5780.66391752580.7326732673correlation=

0.6home&family13730.42002063980.3809523810.3466701657

0.9shopping&services21220.33452211130.3172905526

0.11sports&recreation30810.66903493190.59375

0.5hobbies&interests32270.42214689270.38403452

0.7people&chat33090.46392617450.411582213

0.4health&fitness35000.69212571810.6749379653

0.1business&finance35080.45820433440.4071856287

0.1society&politics48550.48571026330.4133333333

0.12travel&vacations54090.7801566580.7449521785

0.8reference&education55740.58031325030.5637657362

0.2computers&internet57180.61765339070.598476605

0.3entertainment&media111630.70064884510.6779843708

F-trainF-test

Top-m1000-Sum2.Sum2

training
test
Number of Training Examples
F1
F1 Perf as Fcn of Training

Hier012.02-m1000.Sum2.Sum2

Compute using netscore.pl

SUM2-SUM2

53k training

>256-1000

copied from first sheet

SUM2-SUM2 (top level only)

default c=0.01

CategTrainTestSMOX 0.5

05007810024TrainTest

Net1Targ1Net1Targ0Net0Targ1Net0Targ0precrecallbeNet1Targ1Net1Targ0Net0Targ1Net0Targ0precrecallbe

0.05781140.0.p0.5.smox.train.sum.sum.out32270256494300.82142857140.55709342560.68926099850.0.p0.5.smox.test.sum.sum.out74144098960.84090909090.6491228070.745015949

0.135087030.1.p0.5.smox.train.sum.sum.out11844762324460940.7132530120.33751425310.52538363260.1.p0.5.smox.test.sum.sum.out2049549992260.68227424750.29018492180.4862295846

0.2571811260.2.p0.5.smox.train.sum.sum.out28697032849436570.80319148940.50174886320.65247017630.2.p0.5.smox.test.sum.sum.out55016257687360.77247191010.48845470690.6304633085

0.31116321590.3.p0.5.smox.train.sum.sum.out674913534414375620.83300419650.60458658070.71879538860.3.p0.5.smox.test.sum.sum.out125829490175710.81056701030.58267716540.6966220878

0.435007220.4.p0.5.smox.train.sum.sum.out20483701452462080.8469809760.58514285710.71606191660.4.p0.5.smox.test.sum.sum.out4087931492230.83778234090.56509695290.7014396469

0.532276820.5.p0.5.smox.train.sum.sum.out9342642293465870.77963272120.28943290980.53453281550.5.p0.5.smox.test.sum.sum.out1786750492750.72653061220.26099706740.4937638398

0.613732800.6.p0.5.smox.train.sum.sum.out407158966485470.72035398230.29643117260.50839257750.6.p0.5.smox.test.sum.sum.out764320497010.63865546220.27142857140.4550420168

0.733096820.7.p0.5.smox.train.sum.sum.out11063532203464160.75805346130.33423995160.54614670650.7.p0.5.smox.test.sum.sum.out1998648392560.6982456140.29178885630.4950172352

0.8557411750.8.p0.5.smox.train.sum.sum.out25386353036438690.79987393630.455328310.62760112320.8.p0.5.smox.test.sum.sum.out51513766087120.78987730060.43829787230.6140875865

0.921224230.9.p0.5.smox.train.sum.sum.out4692131653477430.68768328450.22101790760.4543505960.9.p0.5.smox.test.sum.sum.out894933495520.64492753620.21040189130.4276647137

0.148559460.10.p0.5.smox.train.sum.sum.out17255233130447000.76734875440.35530381050.56132628250.10.p0.5.smox.test.sum.sum.out27912566789530.69059405940.29492600420.4927600318

0.1130816400.11.p0.5.smox.train.sum.sum.out16952911386467060.85347432020.55014605650.70181018840.11.p0.5.smox.test.sum.sum.out3048033693040.79166666670.4750.6333333333

0.12540910910.12.p0.5.smox.train.sum.sum.out37354311674442380.8965434470.6905158070.7935296270.12.p0.5.smox.test.sum.sum.out7019039088430.88621997470.64252978920.7643748819

5341710743257815840276365917570.81531260870.48263661380.64897461134835132159081182480.78541260560.45006050450.6177365551

Second Level, n=150

Rollup up (netscore.pl)L2 score (p=0.5)

Does not take into account duplication of items -- e.g., item in 0.0.0 and 0.0.1 gets counted twice

822593361276010.03065420560.69491525420.3627847299

41229533151065840.12243684990.56671251720.3445746836

60525066221165550.19447123110.49307253460.3437718829

134321428991058800.38536585370.59901873330.4921922935

52155772761239380.08543784850.65370138020.3695696143

49721442131174340.18818629310.70.4440931465

191298796869420.06010069230.66550522650.3628029594

57777681051118380.06914319950.84604105570.4575921276

8133151401958750.20509586280.66968698520.437391424

26231001871267630.07792980370.5835189310.3307243673

53324884571168100.17643164520.53838383840.3574077418

45343272171152910.09476987450.6761194030.3854446387

7191933425971630.27111613880.62849650350.4498063211

700843669424914486740.13828758610.62254597140.3804167787

Second Level, n=150

Rollup up (netscore.pl)L2 score = 0.9

Does not take into account duplication of items -- e.g., item in 0.0.0 and 0.0.1 gets counted twice

61340571298540.15211970070.51694915250.3345344266

2114655161090720.31213017750.29023383770.3011820076

3825658451184960.40337909190.31132844340.3573537676

94682612961071960.53386004510.42194469220.4779023687

3635304341289850.40649496080.45545796740.4309764641

42111152891184630.27408854170.59295774650.4335231441

1471021140889080.12585616440.5121951220.3190256432

45627242261168820.14339622640.66862170090.4060089636

619920595981060.40220922680.50988467870.4560469528

19110432581288200.15478119940.4253897550.2900854772

3278216631184770.28484320560.33030303030.3075731179

2985293721190890.36033857320.44477611940.4025573463

449366695987300.55092024540.39248251750.4717013814

487111265638614810780.30187159150.43270853690.3672900642

Second Level, n=150

Rollup up (netscore.pl)L2 score = 0.975

Does not take into account duplication of items -- e.g., item in 0.0.0 and 0.0.1 gets counted twice

4598731300960.31468531470.38135593220.3480206234

1181756091093620.40273037540.16231086660.282520621

2892469381188150.54018691590.23553382230.3878603691

66148615811075360.57628596340.29482604820.4355560058

2172035801293120.51666666670.27227101630.3944688415

3687433421188350.33123312330.51830985920.4247714912

116589171893400.16453900710.40418118470.2843600959

3228703601187360.27013422820.47214076250.3711374953

493424721986020.53762268270.40609555190.4718591173

1415573081293060.20200573070.31403118040.2580184555

2334467571188520.34315169370.23535353540.2892526145

2021604681194580.55801104970.30149253730.4297517935

269159875989370.62850467290.23513986010.4318222665

34745156778314871870.40254924680.30860797730.355578612

Second Level, n=150

Rollup up (netscore.pl)L1 && L2 score

Does not take into account duplication of items -- e.g., item in 0.0.0 and 0.0.1 gets counted twice

6431541301630.67368421050.54237288140.6080285459

1511245761094130.54909090910.20770288860.3783968988

4012848261187770.58540145990.32681336590.4561074129

100146812411075540.68141592920.44647636040.5639461448

3421704551293450.667968750.42910915930.5485389547

177895331194890.66541353380.24929577460.4573546542

7358214898710.55725190840.25435540070.4058036545

189944931195120.6678445230.27712609970.4724853113

477180737988460.72602739730.39291598020.5594716887

80813691297820.49689440990.17817371940.3375340647

2252027651190960.52693208430.22727272730.3771024058

2341364361194820.63243243240.34925373130.4908430819

541164603989320.76737588650.47290209790.6201389922

39552081730214902620.65523525510.35133694590.5032861005

Second Level, n=150

Rollup up (netscore.pl)L1 && L2 score (0.50, 0.25)

Does not take into account duplication of items -- e.g., item in 0.0.0 and 0.0.1 gets counted twice

6641521301530.61682242990.55932203390.5880722319

1802165471093210.45454545450.24759284730.3510691509

4524747751185870.48812095030.36837815810.4282495542

108167111611073510.61700913240.48215878680.5495839596

3692444281292710.60195758560.46298619820.5324718919

177955331194830.65073529410.24929577460.4500155344

7567212898620.52816901410.26132404180.3947465279

1911004911195060.65635738830.2800586510.4682080197

494261720987650.65430463580.40691927510.5306119554

851053641297580.44736842110.18930957680.3183389989

2422757481190230.46808510640.24444444440.3562647754

2571954131194230.56858407080.38358208960.4760830802

593296551988000.66704161980.51835664340.5926991316

42623040699514893030.58367570530.37860886560.4811422854

Second Level, n=150

Rollup up (netscore.pl)L1 && L2 score (0.25, 0.25)

Does not take into account duplication of items -- e.g., item in 0.0.0 and 0.0.1 gets counted twice

7055481301390.560.5932203390.5766101695

2734054541091320.40265486730.37551581840.3890853428

5517036761183580.43939393940.44906275470.444228347

128410269581069960.55584415580.57270294380.5642735498

4093173881291980.56336088150.5131744040.5382676428

2231364871194420.62116991640.3140845070.4676272117

8495203898340.4692737430.29268292680.3809783349

2902973921193090.49403747870.42521994130.45962871

569436645985900.56616915420.46869851730.5174338358

1021383471297250.4250.22717149220.3260857461

3475256431187730.39793577980.35050505050.3742204152

3082673621193510.53565217390.45970149250.4976768332

666456478986400.59358288770.58216783220.5878753599

51764856608114874870.51594896330.45980278940.4878758763

Second Level, n=150

Rollup up (netscore.pl)L1 && L2 score (0.25, 0.50)

Does not take into account duplication of items -- e.g., item in 0.0.0 and 0.0.1 gets counted twice

6639521301550.62857142860.55932203390.5939467312

2292344981093030.4946004320.31499312240.4047967772

4704057571186560.53714285710.38304808480.460095471

115167310911073490.63103070180.51338090990.5722058058

3762214211292940.62981574540.47176913430.5507924398

2231274871194510.63714285710.3140845070.4756136821

8279205898500.50931677020.28571428570.397515528

2842653981193410.51730418940.41642228740.4668632384

544309670987170.63774912080.44810543660.5429272787

931033561297600.47448979590.20712694880.3408083723

3163556741189430.47093889720.31919191920.3950654082

2811853891194330.60300429180.41940298510.5112036385

600251544988450.7050528790.52447552450.6147642017

47153246654214890970.59226227860.4188504930.5055563858

Second Level, n=150

Rollup up (netscore.pl)L1 && L2 score (0.20, 0.50)

Does not take into account duplication of items -- e.g., item in 0.0.0 and 0.0.1 gets counted twice

6640521301540.62264150940.55932203390.59098177170.5892857143

2472884801092490.4616822430.33975240720.40071732510.3914421553

4884457391186160.52304394430.39771801140.46038097780.4518518519

117776710651072550.60545267490.52497769850.56521518670.5623506928

3922484051292670.61250.49184441660.55217220830.5455810717

2431464671194320.62467866320.34225352110.48346609220.4422202002

8483203898460.5029940120.29268292680.39783846940.3700440529

3203503621192560.47761194030.46920821110.47341007570.4733727811

572359642986670.61439312570.4711696870.54278140630.5333333333

1031113461297520.48130841120.22939866370.35535353750.3107088989

3414306491188680.44228274970.34444444440.39336359710.3872799546

3022013681194170.60039761430.45074626870.52557194150.5149190111

620277524988190.69119286510.5419580420.61657545350.6075453209

49553745630214885980.56954022990.44017056050.50485539520.4965676204

Second Level, n=150

Rollup up (netscore.pl)L1 * L2 score, .50 = $p

Does not take into account duplication of items -- e.g., item in 0.0.0 and 0.0.1 gets counted twice

0006330551301640.67741935480.53389830510.60565883

000131975961094400.57456140350.18019257220.3773769879

0003702288571188330.6187290970.30154849230.4601387946

00092738713151076350.70547945210.4134701160.559474784

0003321314651293840.7170626350.41656210790.5668123714

000176845341194940.67692307690.24788732390.4624052004

0007349214898800.59836065570.25435540070.4263580282

000184814981195250.69433962260.26979472140.482067172

000466127748988990.78583473860.38385502470.5848448817

00078803711297830.49367088610.17371937640.3336951312

0002061597841191390.56438356160.20808080810.3862321849

0002251224451194960.64841498560.33582089550.4921179406

000522143622989530.7849624060.45629370630.6206280562

000000037531718750414906250.68598062510.33339255570.5096865904

Second Level, n=150

Rollup up (netscore.pl)L1 * L2 score, .35

Does not take into account duplication of items -- e.g., item in 0.0.0 and 0.0.1 gets counted twice

6746511301480.5929203540.56779661020.5803584821

2242575031092800.46569646570.30811554330.3869060045

4894847381185770.50256937310.39853300730.4505511902

117074010721072820.6125654450.52185548620.5672104656

3902434071292720.61611374410.48933500630.5527243752

2201234901194550.64139941690.30985915490.4756292859

8280205898490.50617283950.28571428570.3959435626

2742344081193720.53937007870.40175953080.4705648048

543339671986870.61564625850.44728171330.5314639859

951193541297440.44392523360.21158129180.3277532627

3093466811189520.47175572520.31212121210.3919384687

2942243761193940.56756756760.43880597010.5031867689

633313511987830.66913319240.55332167830.6112274354

47903548646714887950.57447829220.42551301410.4999956531

Second Level, n=150

Rollup up (netscore.pl)L1 * L2 score, .25 = $p*$p

Does not take into account duplication of items -- e.g., item in 0.0.0 and 0.0.1 gets counted twice

0006746511301480.5929203540.56779661020.5803584821

0002242575031092800.46569646570.30811554330.3869060045

0004894847381185770.50256937310.39853300730.4505511902

000117074010721072820.6125654450.52185548620.5672104656

0003902434071292720.61611374410.48933500630.5527243752

0002201234901194550.64139941690.30985915490.4756292859

0008280205898490.50617283950.28571428570.3959435626

0002742344081193720.53937007870.40175953080.4705648048

000543339671986870.61564625850.44728171330.5314639859

000951193541297440.44392523360.21158129180.3277532627

0003093466811189520.47175572520.31212121210.3919384687

0002942243761193940.56756756760.43880597010.5031867689

000633313511987830.66913319240.55332167830.6112274354

000000047903548646714887950.57447829220.42551301410.4999956531

Second Level, n=150

Rollup up (netscore.pl)L1 * L2 score, .20 = $p*$p

Does not take into account duplication of items -- e.g., item in 0.0.0 and 0.0.1 gets counted twice6951491301430.5750.58474576270.57987288140.5798319328

2553294721092080.43664383560.35075653370.39370018470.3890160183

5215757061184860.47536496350.42461287690.44998892020.4485578993

122689510161071270.57802923150.54683318470.56243120810.5619986248

4083023891292130.57464788730.51191969890.54328379310.5414731254

2381414721194370.62796833770.33521126760.48158980270.4370982553

8591202898380.48295454550.29616724740.38956089640.3671706263

3103103721192960.50.45454545450.47727272730.4761904762

580405634986210.58883248730.47775947280.53329598010.5275125057

1061373431297260.43621399180.23608017820.3361470850.3063583815

3424316481188670.44243208280.34545454550.39394331410.3879750425

3182553521193630.5549738220.47462686570.51480034380.5116653258

663402481986940.62253521130.57954545450.60104033290.6002716161

51214324613614880190.54219163580.45491694060.49855428820.4947348082

1024286481227229760380.54219163580.45491694060.49855428820.4947348082

Second Level, n=150

Rollup up (netscore.pl)L1 * L2 score, .75

Does not take into account duplication of items -- e.g., item in 0.0.0 and 0.0.1 gets counted twice

5315651301790.77941176470.44915254240.6142821535

80406471094970.66666666670.11004126550.3883539661

271989561189630.73441734420.22086389570.4776406199

70818315341078390.79461279460.31578947370.5552011341

288735091294420.79778393350.36135508160.5795695075

145545651195240.72864321610.20422535210.4664342841

6124226899050.71764705880.2125435540.4650953064

128255541195810.83660130720.18768328450.5121422958

36068854989580.84112149530.29654036240.5688309289

61463881298170.57009345790.1358574610.3529754595

141798491192190.64090909090.14242424240.3916666667

183624871195560.74693877550.27313432840.5100365519

41784727990120.83233532930.36451048950.5984229094

2896851836114914920.77288497460.2572621480.5150735613

Second Level, n=150

Rollup up (netscore.pl)L1 * L2 score, .10

Does not take into account duplication of items -- e.g., item in 0.0.0 and 0.0.1 gets counted twice

7477441301170.49006622520.62711864410.5585924346

3436893841088480.33236434110.47180192570.4020831334

65610945711179670.37485714290.53463732680.4547472348

145816747841063480.46551724140.65031222120.5579147313

4674783301290370.49417989420.58594730240.5400635983

3202463901193320.56537102470.45070422540.508037625

95125192898040.43181818180.3310104530.3814143174

3987792841188270.33814783350.58357771260.460862773

676721538983050.48389405870.55683690280.5203654807

1442273051296360.38814016170.32071269490.3544264283

4579185331183800.33236363640.46161616160.396989899

3854272851191910.4741379310.57462686570.5243823984

770826374982700.48245614040.67307692310.5777665317

62438281501414840620.42984026440.55458825620.4922142603

SecondLevel (0-12) w/ test on 0-2 (full 50078/10024) files

0.0.08890-12.0-2-sum2/0.0.0.p0.5.smox.test.sum2.sum2710229913

0.0.1310-12.0-2-sum2/0.0.1.p0.5.smox.test.sum2.sum200110023

0.0.2400-12.0-2-sum2/0.0.2.p0.5.smox.test.sum2.sum200010024

0.0.3810-12.0-2-sum2/0.0.3.p0.5.smox.test.sum2.sum200110023

0.0.4156380-12.0-2-sum2/0.0.4.p0.5.smox.test.sum2.sum23440649580

0.0.577260-12.0-2-sum2/0.0.5.p0.5.smox.test.sum2.sum213263139735

0.0.6162240-12.0-2-sum2/0.0.6.p0.5.smox.test.sum2.sum221176638234

0.0.7400-12.0-2-sum2/0.0.7.p0.5.smox.test.sum2.sum204010020

0.0.82230-12.0-2-sum2/0.0.8.p0.5.smox.test.sum2.sum225110016

0.0.92140-12.0-2-sum2/0.0.9.p0.5.smox.test.sum2.sum2115310005

0.0.10620-12.0-2-sum2/0.0.10.p0.5.smox.test.sum2.sum201210021

0.0.11500-12.0-2-sum2/0.0.11.p0.5.smox.test.sum2.sum200010024

0.0.1236100-12.0-2-sum2/0.0.12.p0.5.smox.test.sum2.sum243169983822593361276010.03065420560.69491525420.3627847299

0.1.0116300-12.0-2-sum2/0.1.0.p0.5.smox.test.sum2.sum27187239807

0.1.17561590-12.0-2-sum2/0.1.1.p0.5.smox.test.sum2.sum272254879611

0.1.23410-12.0-2-sum2/0.1.2.p0.5.smox.test.sum2.sum2023110000

0.1.3450900-12.0-2-sum2/0.1.3.p0.5.smox.test.sum2.sum24839429895

0.1.414722950-12.0-2-sum2/0.1.4.p0.5.smox.test.sum2.sum22131879827850

0.1.5146290-12.0-2-sum2/0.1.5.p0.5.smox.test.sum2.sum216126139869

0.1.6125260-12.0-2-sum2/0.1.6.p0.5.smox.test.sum2.sum27121199877

0.1.72340-12.0-2-sum2/0.1.7.p0.5.smox.test.sum2.sum201410019

0.1.892190-12.0-2-sum2/0.1.8.p0.5.smox.test.sum2.sum2777129928

0.1.9289550-12.0-2-sum2/0.1.9.p0.5.smox.test.sum2.sum233157229812

0.1.1096190-12.0-2-sum2/0.1.10.p0.5.smox.test.sum2.sum298910991641229533151065840.12243684990.56671251720.3445746836

0.2.03360-12.0-2-sum2/0.2.0.p0.5.smox.test.sum2.sum204610014

0.2.113402510-12.0-2-sum2/0.2.1.p0.5.smox.test.sum2.sum2172269799504

0.2.29221740-12.0-2-sum2/0.2.2.p0.5.smox.test.sum2.sum2117373579477

0.2.36051210-12.0-2-sum2/0.2.3.p0.5.smox.test.sum2.sum255309669594

0.2.48011730-12.0-2-sum2/0.2.4.p0.5.smox.test.sum2.sum289217849634

0.2.5394770-12.0-2-sum2/0.2.5.p0.5.smox.test.sum2.sum233129449818

0.2.6481980-12.0-2-sum2/0.2.6.p0.5.smox.test.sum2.sum220673789253

0.2.7120260-12.0-2-sum2/0.2.7.p0.5.smox.test.sum2.sum2591219907

0.2.8235510-12.0-2-sum2/0.2.8.p0.5.smox.test.sum2.sum21562369911

0.2.9142330-12.0-2-sum2/0.2.9.p0.5.smox.test.sum2.sum27156269835

0.2.10112210-12.0-2-sum2/0.2.10.p0.5.smox.test.sum2.sum2386189917

0.2.119881960-12.0-2-sum2/0.2.11.p0.5.smox.test.sum2.sum289137107969160525066221165550.19447123110.49307253460.3437718829

0.3.011312290-12.0-2-sum2/0.3.0.p0.5.smox.test.sum2.sum21293951009400

0.3.17131440-12.0-2-sum2/0.3.1.p0.5.smox.test.sum2.sum270160749720

0.3.222254020-12.0-2-sum2/0.3.2.p0.5.smox.test.sum2.sum22835081199114

0.3.3285540-12.0-2-sum2/0.3.3.p0.5.smox.test.sum2.sum215193399777

0.3.4195440-12.0-2-sum2/0.3.4.p0.5.smox.test.sum2.sum221373239607

0.3.585190-12.0-2-sum2/0.3.5.p0.5.smox.test.sum2.sum2799129906

0.3.616683150-12.0-2-sum2/0.3.6.p0.5.smox.test.sum2.sum2153711629638

0.3.731416170-12.0-2-sum2/0.3.7.p0.5.smox.test.sum2.sum24771881409219

0.3.8434770-12.0-2-sum2/0.3.8.p0.5.smox.test.sum2.sum24440339907

0.3.916773370-12.0-2-sum2/0.3.9.p0.5.smox.test.sum2.sum21441131939574

0.3.102040-12.0-2-sum2/0.3.10.p0.5.smox.test.sum2.sum202410018134321428991058800.38536585370.59901873330.4921922935

0.4.0275620-12.0-2-sum2/0.4.0.p0.5.smox.test.sum2.sum22449389913

0.4.13490-12.0-2-sum2/0.4.1.p0.5.smox.test.sum2.sum214810011

0.4.219914220-12.0-2-sum2/0.4.2.p0.5.smox.test.sum2.sum23614028615574

0.4.389150-12.0-2-sum2/0.4.3.p0.5.smox.test.sum2.sum269099919

0.4.4326580-12.0-2-sum2/0.4.4.p0.5.smox.test.sum2.sum24724119942

0.4.5244440-12.0-2-sum2/0.4.5.p0.5.smox.test.sum2.sum26218389762

0.4.6259570-12.0-2-sum2/0.4.6.p0.5.smox.test.sum2.sum237327209640

0.4.72030-12.0-2-sum2/0.4.7.p0.5.smox.test.sum2.sum2013310008

0.4.8468820-12.0-2-sum2/0.4.8.p0.5.smox.test.sum2.sum230609529333

0.4.93160-12.0-2-sum2/0.4.9.p0.5.smox.test.sum2.sum2114510004

0.4.101020-12.0-2-sum2/0.4.10.p0.5.smox.test.sum2.sum201210021

0.4.1197230-12.0-2-sum2/0.4.11.p0.5.smox.test.sum2.sum21184229817

0.4.1264140-12.0-2-sum2/0.4.12.p0.5.smox.test.sum2.sum27167999452155772761239380.08543784850.65370138020.3695696143

0.5.0152390-12.0-2-sum2/0.5.0.p0.5.smox.test.sum2.sum21929209956

0.5.1269670-12.0-2-sum2/0.5.1.p0.5.smox.test.sum2.sum252192159765

0.5.2147380-12.0-2-sum2/0.5.2.p0.5.smox.test.sum2.sum21999199887

0.5.37461450-12.0-2-sum2/0.5.3.p0.5.smox.test.sum2.sum2117255289624

0.5.4111250-12.0-2-sum2/0.5.4.p0.5.smox.test.sum2.sum21624399756

0.5.5341790-12.0-2-sum2/0.5.5.p0.5.smox.test.sum2.sum262110179835

0.5.6142230-12.0-2-sum2/0.5.6.p0.5.smox.test.sum2.sum2204839953

0.5.7311710-12.0-2-sum2/0.5.7.p0.5.smox.test.sum2.sum229483429470

0.5.84490-12.0-2-sum2/0.5.8.p0.5.smox.test.sum2.sum233469981

0.5.9127160-12.0-2-sum2/0.5.9.p0.5.smox.test.sum2.sum21033769671

0.5.107541700-12.0-2-sum2/0.5.10.p0.5.smox.test.sum2.sum2135249359605

0.5.11171280-12.0-2-sum2/0.5.11.p0.5.smox.test.sum2.sum2156513993149721442131174340.18818629310.70.4440931465

0.6.078120-12.0-2-sum2/0.6.0.p0.5.smox.test.sum2.sum2412389889

0.6.1241530-12.0-2-sum2/0.6.1.p0.5.smox.test.sum2.sum2291466248505

0.6.254110-12.0-2-sum2/0.6.2.p0.5.smox.test.sum2.sum2613359880

0.6.333130-12.0-2-sum2/0.6.3.p0.5.smox.test.sum2.sum2076139935

0.6.4395760-12.0-2-sum2/0.6.4.p0.5.smox.test.sum2.sum255500219448

0.6.584210-12.0-2-sum2/0.6.5.p0.5.smox.test.sum2.sum2156869935

0.6.6107240-12.0-2-sum2/0.6.6.p0.5.smox.test.sum2.sum214231109769

0.6.76470-12.0-2-sum2/0.6.7.p0.5.smox.test.sum2.sum2424739770

0.6.8353700-12.0-2-sum2/0.6.8.p0.5.smox.test.sum2.sum26414369811

0.7.043120-12.0-2-sum2/0.7.0.p0.5.smox.test.sum2.sum271510011

0.7.188180-12.0-2-sum2/0.7.1.p0.5.smox.test.sum2.sum21125679750

0.7.22580-12.0-2-sum2/0.7.2.p0.5.smox.test.sum2.sum203789979

0.7.358110-12.0-2-sum2/0.7.3.p0.5.smox.test.sum2.sum257169942

0.7.41440-12.0-2-sum2/0.7.4.p0.5.smox.test.sum2.sum209410011

0.7.53840-12.0-2-sum2/0.7.5.p0.5.smox.test.sum2.sum212139999

0.7.6157290-12.0-2-sum2/0.7.6.p0.5.smox.test.sum2.sum218132119863

0.7.72050-12.0-2-sum2/0.7.7.p0.5.smox.test.sum2.sum218410011

0.7.839140-12.0-2-sum2/0.7.8.p0.5.smox.test.sum2.sum295159959

0.7.925545200-12.0-2-sum2/0.7.9.p0.5.smox.test.sum2.sum24956445253059

0.7.10266530-12.0-2-sum2/0.7.10.p0.5.smox.test.sum2.sum230734239237

0.7.112040-12.0-2-sum2/0.7.11.p0.5.smox.test.sum2.sum203410017

0.8.012362510-12.0-2-sum2/0.8.0.p0.5.smox.test.sum2.sum21761599758174

0.8.1154260-12.0-2-sum2/0.8.1.p0.5.smox.test.sum2.sum29197179801

0.8.216623460-12.0-2-sum2/0.8.2.p0.5.smox.test.sum2.sum2286235609443

0.8.3484940-12.0-2-sum2/0.8.3.p0.5.smox.test.sum2.sum246240489690

0.8.49592070-12.0-2-sum2/0.8.4.p0.5.smox.test.sum2.sum2168113399704

0.8.5188390-12.0-2-sum2/0.8.5.p0.5.smox.test.sum2.sum22852119933

0.8.62590-12.0-2-sum2/0.8.6.p0.5.smox.test.sum2.sum201910014

0.8.7273640-12.0-2-sum2/0.8.7.p0.5.smox.test.sum2.sum233301319659

0.8.84591100-12.0-2-sum2/0.8.8.p0.5.smox.test.sum2.sum250282609632

0.8.9321680-12.0-2-sum2/0.8.9.p0.5.smox.test.sum2.sum217131519825

0.9.0122190-12.0-2-sum2/0.9.0.p0.5.smox.test.sum2.sum2998109907

0.9.1323650-12.0-2-sum2/0.9.1.p0.5.smox.test.sum2.sum23470319889

0.9.2110250-12.0-2-sum2/0.9.2.p0.5.smox.test.sum2.sum2164499955

0.9.3126210-12.0-2-sum2/0.9.3.p0.5.smox.test.sum2.sum2148879915

0.9.4122210-12.0-2-sum2/0.9.4.p0.5.smox.test.sum2.sum211139109864

0.9.5235610-12.0-2-sum2/0.9.5.p0.5.smox.test.sum2.sum235105269858

0.9.6105150-12.0-2-sum2/0.9.6.p0.5.smox.test.sum2.sum2104459965

0.9.7198440-12.0-2-sum2/0.9.7.p0.5.smox.test.sum2.sum221578239402

0.9.8162240-12.0-2-sum2/0.9.8.p0.5.smox.test.sum2.sum212325129675

0.9.9121210-12.0-2-sum2/0.9.9.p0.5.smox.test.sum2.sum21413579868

0.9.1063140-12.0-2-sum2/0.9.10.p0.5.smox.test.sum2.sum23145119865

0.9.11424900-12.0-2-sum2/0.9.11.p0.5.smox.test.sum2.sum2591126318808

0.9.12180290-12.0-2-sum2/0.9.12.p0.5.smox.test.sum2.sum22420359792

0.10.0465920-12.0-2-sum2/0.10.0.p0.5.smox.test.sum2.sum238283549649

0.10.1103230-12.0-2-sum2/0.10.1.p0.5.smox.test.sum2.sum258189993

0.10.2161390-12.0-2-sum2/0.10.2.p0.5.smox.test.sum2.sum2953309932

0.10.3262560-12.0-2-sum2/0.10.3.p0.5.smox.test.sum2.sum226249309719

0.10.414012560-12.0-2-sum2/0.10.4.p0.5.smox.test.sum2.sum2175693819075

0.10.595170-12.0-2-sum2/0.10.5.p0.5.smox.test.sum2.sum2568129939

0.10.6313530-12.0-2-sum2/0.10.6.p0.5.smox.test.sum2.sum212165419806

0.10.7370750-12.0-2-sum2/0.10.7.p0.5.smox.test.sum2.sum23668399881

0.10.87631440-12.0-2-sum2/0.10.8.p0.5.smox.test.sum2.sum291507539373

0.10.9176370-12.0-2-sum2/0.10.9.p0.5.smox.test.sum2.sum21426239961

0.10.107411560-12.0-2-sum2/0.10.10.p0.5.smox.test.sum2.sum2111156459712

0.10.11182420-12.0-2-sum2/0.10.11.p0.5.smox.test.sum2.sum211212319770

0.11.0241620-12.0-2-sum2/0.11.0.p0.5.smox.test.sum2.sum223125399837

0.11.1110170-12.0-2-sum2/0.11.1.p0.5.smox.test.sum2.sum2102579982

0.11.2347700-12.0-2-sum2/0.11.2.p0.5.smox.test.sum2.sum24420269934

0.11.3361620-12.0-2-sum2/0.11.3.p0.5.smox.test.sum2.sum24218209944

0.11.4139330-12.0-2-sum2/0.11.4.p0.5.smox.test.sum2.sum2303139960

0.11.5344690-12.0-2-sum2/0.11.5.p0.5.smox.test.sum2.sum24417259938

0.11.6123330-12.0-2-sum2/0.11.6.p0.5.smox.test.sum2.sum2263679955

0.11.7164360-12.0-2-sum2/0.11.7.p0.5.smox.test.sum2.sum220121169867

0.11.83940-12.0-2-sum2/0.11.8.p0.5.smox.test.sum2.sum224210016

0.11.9213420-12.0-2-sum2/0.11.9.p0.5.smox.test.sum2.sum22717159965

0.11.1010752330-12.0-2-sum2/0.11.10.p0.5.smox.test.sum2.sum21813901525890

0.11.118290-12.0-2-sum2/0.11.11.p0.5.smox.test.sum2.sum2412510003

0.12.016123320-12.0-2-sum2/0.12.0.p0.5.smox.test.sum2.sum22255251079167

0.12.1176390-12.0-2-sum2/0.12.1.p0.5.smox.test.sum2.sum21825219960

0.12.2328500-12.0-2-sum2/0.12.2.p0.5.smox.test.sum2.sum22988219886

0.12.3239440-12.0-2-sum2/0.12.3.p0.5.smox.test.sum2.sum22075249905

0.12.49682040-12.0-2-sum2/0.12.4.p0.5.smox.test.sum2.sum2124220809600

0.12.560160-12.0-2-sum2/0.12.5.p0.5.smox.test.sum2.sum2031610005

0.12.619213750-12.0-2-sum2/0.12.6.p0.5.smox.test.sum2.sum22735601029089

0.12.764140-12.0-2-sum2/0.12.7.p0.5.smox.test.sum2.sum267810003

0.12.8216570-12.0-2-sum2/0.12.8.p0.5.smox.test.sum2.sum219329389638

0.12.951130-12.0-2-sum2/0.12.9.p0.5.smox.test.sum2.sum2510189910

55849112570000000700843669424914486740.13828758610.62254597140.3804167787

BAD-Hier012.02-m1000.Sum2.Sum2

Compute using netscore.pl

SUM2-SUM2

53k training

>256-1000

copied from first sheet

SUM2-SUM2 (top level only)

default c=0.01

CategTrainTestSMOX 0.5

05007810024TrainTest

Net1Targ1Net1Targ0Net0Targ1Net0Targ0precrecallbeNet1Targ1Net1Targ0Net0Targ1Net0Targ0precrecallbe

0.05781140.0.p0.5.smox.train.sum.sum.out32270256494300.82142857140.55709342560.68926099850.0.p0.5.smox.test.sum.sum.out74144098960.84090909090.6491228070.745015949

0.135087030.1.p0.5.smox.train.sum.sum.out11844762324460940.7132530120.33751425310.52538363260.1.p0.5.smox.test.sum.sum.out2049549992260.68227424750.29018492180.4862295846

0.2571811260.2.p0.5.smox.train.sum.sum.out28697032849436570.80319148940.50174886320.65247017630.2.p0.5.smox.test.sum.sum.out55016257687360.77247191010.48845470690.6304633085

0.31116321590.3.p0.5.smox.train.sum.sum.out674913534414375620.83300419650.60458658070.71879538860.3.p0.5.smox.test.sum.sum.out125829490175710.81056701030.58267716540.6966220878

0.435007220.4.p0.5.smox.train.sum.sum.out20483701452462080.8469809760.58514285710.71606191660.4.p0.5.smox.test.sum.sum.out4087931492230.83778234090.56509695290.7014396469

0.532276820.5.p0.5.smox.train.sum.sum.out9342642293465870.77963272120.28943290980.53453281550.5.p0.5.smox.test.sum.sum.out1786750492750.72653061220.26099706740.4937638398

0.613732800.6.p0.5.smox.train.sum.sum.out407158966485470.72035398230.29643117260.50839257750.6.p0.5.smox.test.sum.sum.out764320497010.63865546220.27142857140.4550420168

0.733096820.7.p0.5.smox.train.sum.sum.out11063532203464160.75805346130.33423995160.54614670650.7.p0.5.smox.test.sum.sum.out1998648392560.6982456140.29178885630.4950172352

0.8557411750.8.p0.5.smox.train.sum.sum.out25386353036438690.79987393630.455328310.62760112320.8.p0.5.smox.test.sum.sum.out51513766087120.78987730060.43829787230.6140875865

0.921224230.9.p0.5.smox.train.sum.sum.out4692131653477430.68768328450.22101790760.4543505960.9.p0.5.smox.test.sum.sum.out894933495520.64492753620.21040189130.4276647137

0.148559460.10.p0.5.smox.train.sum.sum.out17255233130447000.76734875440.35530381050.56132628250.10.p0.5.smox.test.sum.sum.out27912566789530.69059405940.29492600420.4927600318

0.1130816400.11.p0.5.smox.train.sum.sum.out16952911386467060.85347432020.55014605650.70181018840.11.p0.5.smox.test.sum.sum.out3048033693040.79166666670.4750.6333333333

0.12540910910.12.p0.5.smox.train.sum.sum.out37354311674442380.8965434470.6905158070.7935296270.12.p0.5.smox.test.sum.sum.out7019039088430.88621997470.64252978920.7643748819

5341710743257815840276365917570.81531260870.48263661380.64897461134835132159081182480.78541260560.45006050450.6177365551

Second Level, n=150

Rollup up (netscore.pl)L2 score (p=0.5)

Does not take into account duplication of items -- e.g., item in 0.0.0 and 0.0.1 gets counted twice

69112491300820.38121546960.58474576270.4829806162

1641185631094190.58156028370.22558459420.403572439

3751658521188960.69444444440.30562347190.5000339582

105640411861076180.72328767120.47100802850.5971478499

42616543711278610.20480769230.53450439150.3696560419

3713153391192630.54081632650.52253521130.5316757689

110196177897330.35947712420.38327526130.3713761928

50744111751151950.10309068730.74340175950.4232462234

610392604986340.60878243510.50247116970.5556268024

124983251297650.55855855860.2761692650.4173639118

2321507581191480.60732984290.23434343430.4208366386

4044152661192030.49328449330.60298507460.548134784

613423531986730.59169884170.53583916080.5637690013

50618853619614834900.36373436830.44958692370.406660646

Second Level, n=150

Rollup up (netscore.pl)L2 score = 0.975

Does not take into account duplication of items -- e.g., item in 0.0.0 and 0.0.1 gets counted twice

389801301850.80851063830.32203389830.5652722683

50106771095270.83333333330.06877579090.4510545621

1871110401190500.94444444440.1524042380.5484243412

5125817301079640.8982456140.2283675290.5633065715

171256261294900.87244897960.21455457970.5435017796

2711224391194560.689567430.38169014080.5356287854

7273215898560.49655172410.25087108010.3737114021

153625291195440.7116279070.2243401760.4679840415

34642868989840.89175257730.28500823720.5883804073

43214061