Are you sure you want to delete this task? Once this task is deleted, it cannot be recovered.
Kevin Clark d441560021 | 5 years ago | |
---|---|---|
.. | ||
base | 5 years ago | |
corpus_processing | 5 years ago | |
model | 5 years ago | |
task_specific | 5 years ago | |
training | 5 years ago | |
README.md | 5 years ago | |
__init__.py | 5 years ago | |
cvt.py | 5 years ago | |
fetch_data.sh | 5 years ago | |
preprocessing.py | 5 years ago |
This repository contains code for Semi-Supervised Sequence Modeling with Cross-View Training. Currently sequence tagging and dependency parsing tasks are supported.
This code has been run with TensorFlow 1.10.1 and Numpy 1.14.5; other versions may work, but have not been tested.
Run fetch_data.sh
to download and extract pretrained GloVe vectors, the 1 Billion Word Language Model Benchmark corpus of unlabeled data, and the CoNLL-2000 text chunking dataset. Unfortunately the other datasets from our paper are not freely available and so can't be included in this repository.
To apply CVT to other datasets, the data should be placed in data/raw_data/<task_name>/(train|dev|test).txt
. For sequence tagging data, each line should contain a word followed by a space followed by that word's tag. Sentences should be separated by empty lines. For dependency parsing, each tag should be of the form <index_of_head>-<relation>
(e.g., 0-root
).
After all of the data has been downloaded, run preprocessing.py
.
Run python cvt.py --mode=train --model_name=chunking_model
. By default this trains a model on the chunking data downloaded with fetch_data.sh
. To change which task(s) are trained on or model hyperparameters, modify base/configure.py. Models are automatically checkpointed every 1000 steps; training will continue from the latest checkpoint if training is interrupted and restarted. Model checkpoints and other data such as dev set accuracy over time are stored in data/models/<model_name>
.
Run python cvt.py --mode=eval --model_name=chunking_model
. A CVT model trained on the chunking data for 200k steps should get at least 97.1 F1 on the dev set and 96.6 F1 on the test set.
If you use this code for your publication, please cite the original paper:
@inproceedings{clark2018semi,
title = {Semi-Supervised Sequence Modeling with Cross-View Training},
author = {Kevin Clark and Minh-Thang Luong and Christopher D. Manning and Quoc V. Le},
booktitle = {EMNLP},
year = {2018}
}
No Description
https://readpaper.com/paper/2804047946
Python Jupyter Notebook Unity3D Asset Text C++ other
Dear OpenI User
Thank you for your continuous support to the Openl Qizhi Community AI Collaboration Platform. In order to protect your usage rights and ensure network security, we updated the Openl Qizhi Community AI Collaboration Platform Usage Agreement in January 2024. The updated agreement specifies that users are prohibited from using intranet penetration tools. After you click "Agree and continue", you can continue to use our services. Thank you for your cooperation and understanding.
For more agreement content, please refer to the《Openl Qizhi Community AI Collaboration Platform Usage Agreement》