What is the best redundant labeling strategy to maximize the performance of…
I have 50,000 sentences I need labeled for a custom NER model, with triple redudancy. I can have as many workers as I want do the labeling. I will combine their labels using Snorkel's LabelModel. What is the best strategy to employ to give the LabelModel the best performance…
Strategies for finding the best n-grams for keyword functions?
I'm working like hell to find enough LFs for the GENERAL label (as in general purpose open source library) compared to API (as in API-specific open source library) for Amazon's Github repositories as my dataset, to create labels for a discriminative classifier for Amazon Github…
Learning to denoise rules to obtain more accurate classifiers!
Hi Everyone, In case you have an application where rules are much noisy and labeled data is very limited, please consider checking out ICLR 2020 spotlight paper Learning from Rules Generalizing Labeled Exemplars …
What are best practices to combine noisy and "golden" labels?
Hello, I went through tutorials and so much amazed by Snorkel. I train the model on Snorkel labels and then run another training (just few epochs) with hand-labelled data. Are there any better ways to combine gold and noisy labels? Since I already have gold labels think they…