Comments (2)
Hi @logan-markewich thanks for your question, I am always happy to discuss these things.
When evaluating Relation Extraction, especially when referring to End-to-End, the comparison is always with work that extracts both entities and relations.
Of course an interesting "ablation" is to see what happens when NER is solved and see how the model performs on RE, to study how much of the error rate comes from that. With REBEL there is the particularity that both tasks are decoded simultaneously and evaluating RE separately can be tricky, and wasn't the point of our work.
Nevertheless there is Relation Classification (RC) which is more akin to what you describe. Benchmarks such as DocRED on it's original setting, or Re-Tacred, look at RC, and we also evaluate REBEL on it, by marking the spans of the entities in the input, however since pre-training wasn't designed for this, there is not that much of a contribution from REBEL but rather just competitive results that serve as comparison. Basically if I needed a system for Relation Classification between pairs of given entities I would not use REBEL.
Many systems perform RE in a pipeline fashion, where you first perform NER and then RE. Therefore I think it would be fair to compare what you describe with a system on let's say CONLL-04, by providing the gold entity spans to the RE module of such system. The paper from Table-sequence actually has a very nice ablation section where they perform this experiment on ACE-05, at Table 3, where they get around 3 points improvement on RE when the gold entities are provided. And you are in luck because someone had the same question in their github and you can run the same experiment on other datasets (like CONLL04): LorrinWWW/two-are-better-than-one#18
Sorry if I went into a tangent. To sum up, imo, directly comparing a system that considers NER solved with REBEL or other benchmarks in the paper wouldn't be entirely fair, but you could simply take some previous work and use their RE module once the entities are identified, so that the comparison would be fair. If you still want to compare with REBEL, you could follow the setup we had for Re-Tacred and mark the entities spans in the input and re-train for a dataset like CONLL-04. I recall trying that myself at some point and results didn't differ much (ie. NER is usually not the bottleneck) but can't remember more details.
Finally I can't recommend enough to check the following paper which we also cite in our work:
Letβs Stop Incorrect Comparisons in End-to-end Relation Extraction!. I think it is a crucial paper to understand the current state of RE (despite being a bit old now) and how many times comparisons have not been done properly.
Hope my opinion helps, best of luck in your work and let me know if I can help with anything else.
from rebel.
Wow, thanks for the details and links! Super helpful! ππ» Definitely answers my questions and should help a ton with my work
from rebel.
Related Issues (20)
- Documentation request for complete spacy setup HOT 1
- SpaCy component mapping to docs HOT 1
- error in your demo code HOT 2
- Issue while loading the trained checkpoint HOT 3
- How do I know which original span a predicted entity refers to? HOT 2
- DocRED dataset HOT 1
- Replicating REBEL from BART and some issues HOT 2
- Role of shift_tokens_left HOT 1
- Guide to Fine-tuning on Spacy HOT 2
- is it possible to specify which word i want to generate relations for? HOT 1
- Can't find factory for 'rebel' for language English (en). HOT 1
- Explainability of REBEL HOT 1
- Fine tuning for person to person entity relationship extraction HOT 1
- Problem with negative samples HOT 3
- Extraction of non-existant relation HOT 1
- Issue while running the default_model on training with conl dataset HOT 2
- Error while executing conl dataset HOT 1
- Dataset generation error HOT 1
- Pred file and gold file issues
- version incompatible HOT 2
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
π Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. πππ
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google β€οΈ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from rebel.