Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

labels for columns in benchmark #1

Open
amueller opened this issue Oct 23, 2020 · 3 comments
Open

labels for columns in benchmark #1

amueller opened this issue Oct 23, 2020 · 3 comments

Comments

@amueller
Copy link

amueller commented Oct 23, 2020

hey! I was trying to run the benchmark, but it's not clear to me where the ground truth labels are stored.
I assume meta_data.csv maps the column id to the csv file it came from, right?

Thanks!

@pvn25
Copy link
Owner

pvn25 commented Oct 23, 2020

Hi,

The y_act column in the data_train and data_test files denote the ground truth label. The label vocabulary is coded as follows:

Numeric : 0
Categorical: 1
Datetime:2
Sentence:3
URL: 4
Numbers: 5
List: 6
Not-Generalizable: 7
Custom Object (or Context-Specific): 8

I've updated the readme to reflect this. I'm really sorry for this confusion.

Yes, the meta_data.csv maps every column to the source CSV file. So, a join on Record_id in data_train and meta_data would give the source file of that column. The raw CSV files can be downloaded from here.

Thank you for pointing this out.

@amueller
Copy link
Author

thank you for the quick reply, I'll give it a go next week!

@amueller
Copy link
Author

Took a bit longer, results on the training set are here:

                   precision    recall  f1-score   support

          Numeric       0.78      0.86      0.82      2909
      Categorical       0.51      0.64      0.57      1854
         Datetime       0.00      0.00      0.00       549
         Sentence       0.17      0.69      0.27       293
              URL       0.00      0.00      0.00       120
          Numbers       0.00      0.00      0.00       469
             List       0.00      0.00      0.00       188
Not-Generalizable       0.44      0.60      0.51       848
    Custom Object       0.00      0.00      0.00       706

         accuracy                           0.56      7936
        macro avg       0.21      0.31      0.24      7936
     weighted avg       0.46      0.56      0.50      7936

image

Some categories are not produced by dabl right now, but I'm also a bit confused by some of the mistakes, see #4

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants