Code Monkey home page Code Monkey logo

Comments (2)

LittlePea13 avatar LittlePea13 commented on May 30, 2024 4

Hi @logan-markewich thanks for your question, I am always happy to discuss these things.

When evaluating Relation Extraction, especially when referring to End-to-End, the comparison is always with work that extracts both entities and relations.

Of course an interesting "ablation" is to see what happens when NER is solved and see how the model performs on RE, to study how much of the error rate comes from that. With REBEL there is the particularity that both tasks are decoded simultaneously and evaluating RE separately can be tricky, and wasn't the point of our work.

Nevertheless there is Relation Classification (RC) which is more akin to what you describe. Benchmarks such as DocRED on it's original setting, or Re-Tacred, look at RC, and we also evaluate REBEL on it, by marking the spans of the entities in the input, however since pre-training wasn't designed for this, there is not that much of a contribution from REBEL but rather just competitive results that serve as comparison. Basically if I needed a system for Relation Classification between pairs of given entities I would not use REBEL.

Many systems perform RE in a pipeline fashion, where you first perform NER and then RE. Therefore I think it would be fair to compare what you describe with a system on let's say CONLL-04, by providing the gold entity spans to the RE module of such system. The paper from Table-sequence actually has a very nice ablation section where they perform this experiment on ACE-05, at Table 3, where they get around 3 points improvement on RE when the gold entities are provided. And you are in luck because someone had the same question in their github and you can run the same experiment on other datasets (like CONLL04): LorrinWWW/two-are-better-than-one#18

Sorry if I went into a tangent. To sum up, imo, directly comparing a system that considers NER solved with REBEL or other benchmarks in the paper wouldn't be entirely fair, but you could simply take some previous work and use their RE module once the entities are identified, so that the comparison would be fair. If you still want to compare with REBEL, you could follow the setup we had for Re-Tacred and mark the entities spans in the input and re-train for a dataset like CONLL-04. I recall trying that myself at some point and results didn't differ much (ie. NER is usually not the bottleneck) but can't remember more details.

Finally I can't recommend enough to check the following paper which we also cite in our work:
Let’s Stop Incorrect Comparisons in End-to-end Relation Extraction!. I think it is a crucial paper to understand the current state of RE (despite being a bit old now) and how many times comparisons have not been done properly.

Hope my opinion helps, best of luck in your work and let me know if I can help with anything else.

from rebel.

logan-markewich avatar logan-markewich commented on May 30, 2024

Wow, thanks for the details and links! Super helpful! πŸ‘πŸ» Definitely answers my questions and should help a ton with my work

from rebel.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    πŸ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. πŸ“ŠπŸ“ˆπŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❀️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.