# Attention-Based-BiLSTM-relation-extraction **Repository Path**: basilwang/Attention-Based-BiLSTM-relation-extraction ## Basic Information - **Project Name**: Attention-Based-BiLSTM-relation-extraction - **Description**: https://github.com/SeoSangwoo/Attention-Based-BiLSTM-relation-extraction.git - **Primary Language**: Unknown - **License**: Apache-2.0 - **Default Branch**: master - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 1 - **Created**: 2020-09-06 - **Last Updated**: 2020-12-19 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # Attention-Based Bidirectional Long Short-Term Memory Networks for Relation Classification Tensorflow Implementation of Deep Learning Approach for Relation Extraction Challenge([**SemEval-2010 Task #8**: *Multi-Way Classification of Semantic Relations Between Pairs of Nominals*](https://docs.google.com/document/d/1QO_CnmvNRnYwNWu1-QCAeR5ToQYkXUqFeAJbdEhsq7w/preview)) via Attention-based BiLSTM.

## Usage ### Train * Train data is located in "*SemEval2010_task8_all_data/SemEval2010_task8_training/TRAIN_FILE.TXT*". * "[glove.6B.100d](https://nlp.stanford.edu/projects/glove/)" is used as pre-trained glove model. * Performance (accuracy and f1-socre) outputs during training are **NOT OFFICIAL SCORE** of *SemEval 2010 Task 8*. To compute the official performance, you should proceed the follow [Evaluation](#evaluation) step with checkpoints obtained by training. ##### Display help message: ```bash $ python train.py --help ``` ##### Train Example: ```bash $ python train.py --embedding_path "glove.6B.100d.txt" ``` ### Evaluation * You can get an **OFFICIAL SCORE** of *SemEval 2010 Task 8* for test data by following this step. [README](SemEval2010_task8_all_data/SemEval2010_task8_scorer-v1.2/README.txt) describes how to evaluate the official score. * Test data is located in "*SemEval2010_task8_all_data/SemEval2010_task8_testing_keys/TEST_FILE_FULL.TXT*". * **MUST GIVE `--checkpoint_dir` ARGUMENT**, checkpoint directory from training run, like below example. ##### Evaluation Example: ```bash $ python eval.py --checkpoint_dir "runs/1523902663/checkpoints/" ``` ## SemEval-2010 Task #8 * Given: a pair of *nominals* * Goal: recognize the semantic relation between these nominals. * Example: * "There were apples, **pears** and oranges in the **bowl**."
→ *CONTENT-CONTAINER(pears, bowl)* * “The cup contained **tea** from dried **ginseng**.”
→ *ENTITY-ORIGIN(tea, ginseng)* ### The Inventory of Semantic Relations 1. *Cause-Effect(CE)*: An event or object leads to an effect(those cancers were caused by radiation exposures) 2. *Instrument-Agency(IA)*: An agent uses an instrument(phone operator) 3. *Product-Producer(PP)*: A producer causes a product to exist (a factory manufactures suits) 4. *Content-Container(CC)*: An object is physically stored in a delineated area of space (a bottle full of honey was weighed) Hendrickx, Kim, Kozareva, Nakov, O S´ eaghdha, Pad ´ o,´ Pennacchiotti, Romano, Szpakowicz Task Overview Data Creation Competition Results and Discussion The Inventory of Semantic Relations (III) 5. *Entity-Origin(EO)*: An entity is coming or is derived from an origin, e.g., position or material (letters from foreign countries) 6. *Entity-Destination(ED)*: An entity is moving towards a destination (the boy went to bed) 7. *Component-Whole(CW)*: An object is a component of a larger whole (my apartment has a large kitchen) 8. *Member-Collection(MC)*: A member forms a nonfunctional part of a collection (there are many trees in the forest) 9. *Message-Topic(CT)*: An act of communication, written or spoken, is about a topic (the lecture was about semantics) 10. *OTHER*: If none of the above nine relations appears to be suitable. ### Distribution for Dataset * **SemEval-2010 Task #8 Dataset [[Download](https://drive.google.com/file/d/0B_jQiLugGTAkMDQ5ZjZiMTUtMzQ1Yy00YWNmLWJlZDYtOWY1ZDMwY2U4YjFk/view?layout=list&ddrp=1&sort=name&num=50#)]** | Relation | Train Data | Test Data | Total Data | |--------------------|:-------------------:|:-------------------:|:--------------------:| | Cause-Effect | 1,003 (12.54%) | 328 (12.07%) | 1331 (12.42%) | | Instrument-Agency | 504 (6.30%) | 156 (5.74%) | 660 (6.16%) | | Product-Producer | 717 (8.96%) | 231 (8.50%) | 948 (8.85%) | | Content-Container | 540 (6.75%) | 192 (7.07%) | 732 (6.83%) | | Entity-Origin | 716 (8.95%) | 258 (9.50%) | 974 (9.09%) | | Entity-Destination | 845 (10.56%) | 292 (10.75%) | 1137 (10.61%) | | Component-Whole | 941 (11.76%) | 312 (11.48%) | 1253 (11.69%) | | Member-Collection | 690 (8.63%) | 233 (8.58%) | 923 (8.61%) | | Message-Topic | 634 (7.92%) | 261 (9.61%) | 895 (8.35%) | | Other | 1,410 (17.63%) | 454 (16.71%) | 1864 (17.39%) | | **Total** | **8,000 (100.00%)** | **2,717 (100.00%)** | **10,717 (100.00%)** | ## Reference * **Attention-Based Bidirectional Long Short-Term Memory Networks for Relation Classification** (ACL 2016), P Zhou et al. [[paper]](http://www.aclweb.org/anthology/P16-2034) * roomylee's cnn-relation-extraction repository [[github]](https://github.com/roomylee/cnn-relation-extraction)