Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why would freezing entities/predicates most of the epochs work well? #2

Open
xuanmingcui opened this issue Sep 23, 2024 · 0 comments
Open

Comments

@xuanmingcui
Copy link

Hello,

Thank you very much for the wonderful work and the codebase! I'm also interested in NTP and am trying to build works on top of it. However I'm confused by why does GNTP work well when we freeze entities/relations for the most of the epochs (e.g. 95/100 for FB). In the appendix the paper mentioned:

"On FB122, we found it useful to pre-train rules first (95 epochs), without updating any entity or relation embeddings, and then training the entity embeddings jointly with the rules (5 epochs). This forces GNTPs to learn a good rule-based model of the domain before fine-tuning its representations."

I could imagine it may be possible to learn rule-templates with randomly-initialized entity/predicate embeddings. However, since the unification scores are calculated based on embeddings, I don't understand why freezing entity/relations 95% of the time can benefit the model. Is there an explanation of this?

Thank you very much!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant