Welcome to mirror list, hosted at ThFree Co, Russian Federation.

github.com/stanfordnlp/stanza.git - Unnamed repository; edit this file 'description' to name the repository.
summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorYuhui Zhang <zyh.thu@gmail.com>2020-03-06 10:45:58 +0300
committerYuhui Zhang <zyh.thu@gmail.com>2020-03-06 10:45:58 +0300
commit451a9237415a88b07aaab0ce907de1e7bc3674c4 (patch)
tree01a132f1213051b7234f7ce360d06c95b7b0a14f /README.md
parent289f148dc2de232a2a62cbca85d9c6b9dc8e3f95 (diff)
renaming stanfordnlp to stanza!!
Diffstat (limited to 'README.md')
-rw-r--r--README.md50
1 files changed, 25 insertions, 25 deletions
diff --git a/README.md b/README.md
index d2279213..89dbc7f0 100644
--- a/README.md
+++ b/README.md
@@ -1,10 +1,10 @@
-# StanfordNLP: A Python NLP Library for Many Human Languages
+# Stanza: A Python NLP Library for Many Human Languages
[![Travis Status](https://travis-ci.com/stanfordnlp/stanfordnlp.svg?token=RPNzRzNDQRoq2x3J2juj&branch=master)](https://travis-ci.com/stanfordnlp/stanfordnlp)
[![PyPI Version](https://img.shields.io/pypi/v/stanfordnlp.svg?colorB=blue)](https://pypi.org/project/stanfordnlp/)
![Python Versions](https://img.shields.io/pypi/pyversions/stanfordnlp.svg?colorB=blue)
-The Stanford NLP Group's official Python NLP library. It contains packages for running our latest fully neural pipeline from the CoNLL 2018 Shared Task and for accessing the Java Stanford CoreNLP server. For detailed information please visit our [official website](https://stanfordnlp.github.io/stanfordnlp/).
+The Stanford NLP Group's official Python NLP library. It contains packages for running our latest fully neural pipeline from the CoNLL 2018 Shared Task and for accessing the Java Stanford CoreNLP server. For detailed information please visit our [official website](https://stanfordnlp.github.io/stanza/).
### References
@@ -31,45 +31,45 @@ If you use the CoreNLP server, please cite the CoreNLP software package and the
## Issues and Usage Q&A
-To ask questions, report issues or request features, please use the [GitHub Issue Tracker](https://github.com/stanfordnlp/stanfordnlp/issues).
+To ask questions, report issues or request features, please use the [GitHub Issue Tracker](https://github.com/stanfordnlp/stanza/issues).
## Setup
-StanfordNLP supports Python 3.6 or later. We strongly recommend that you install StanfordNLP from PyPI. If you already have [pip installed](https://pip.pypa.io/en/stable/installing/), simply run:
+Stanza supports Python 3.6 or later. We strongly recommend that you install Stanza from PyPI. If you already have [pip installed](https://pip.pypa.io/en/stable/installing/), simply run:
```bash
-pip install stanfordnlp
+pip install stanza
```
-this should also help resolve all of the dependencies of StanfordNLP, for instance [PyTorch](https://pytorch.org/) 1.0.0 or above.
+this should also help resolve all of the dependencies of Stanza, for instance [PyTorch](https://pytorch.org/) 1.0.0 or above.
-If you currently have a previous version of `stanfordnlp` installed, use:
+If you currently have a previous version of `stanza` installed, use:
```bash
-pip install stanfordnlp -U
+pip install stanza -U
```
-Alternatively, you can also install from source of this git repository, which will give you more flexibility in developing on top of StanfordNLP and training your own models. For this option, run
+Alternatively, you can also install from source of this git repository, which will give you more flexibility in developing on top of Stanza and training your own models. For this option, run
```bash
-git clone https://github.com/stanfordnlp/stanfordnlp.git
-cd stanfordnlp
+git clone https://github.com/stanfordnlp/stanza.git
+cd stanza
pip install -e .
```
-## Running StanfordNLP
+## Running Stanza
### Getting Started with the neural pipeline
-To run your first StanfordNLP pipeline, simply following these steps in your Python interactive interpreter:
+To run your first Stanza pipeline, simply following these steps in your Python interactive interpreter:
```python
->>> import stanfordnlp
->>> stanfordnlp.download('en') # This downloads the English models for the neural pipeline
+>>> import stanza
+>>> stanza.download('en') # This downloads the English models for the neural pipeline
# IMPORTANT: The above line prompts you before downloading, which doesn't work well in a Jupyter notebook.
-# To avoid a prompt when using notebooks, instead use: >>> stanfordnlp.download('en', force=True)
->>> nlp = stanfordnlp.Pipeline() # This sets up a default neural pipeline in English
+# To avoid a prompt when using notebooks, instead use: >>> stanza.download('en', force=True)
+>>> nlp = stanza.Pipeline() # This sets up a default neural pipeline in English
>>> doc = nlp("Barack Obama was born in Hawaii. He was elected president in 2008.")
>>> doc.sentences[0].print_dependencies()
```
-The last command will print out the words in the first sentence in the input string (or `Document`, as it is represented in StanfordNLP), as well as the indices for the word that governs it in the Universal Dependencies parse of that sentence (its "head"), along with the dependency relation between the words. The output should look like:
+The last command will print out the words in the first sentence in the input string (or `Document`, as it is represented in Stanza), as well as the indices for the word that governs it in the Universal Dependencies parse of that sentence (its "head"), along with the dependency relation between the words. The output should look like:
```
('Barack', '4', 'nsubj:pass')
@@ -83,13 +83,13 @@ The last command will print out the words in the first sentence in the input str
**Note:** If you are running into issues like `OSError: [Errno 22] Invalid argument`, it's very likely that you are affected by a [known Python issue](https://bugs.python.org/issue24658), and we would recommend Python 3.6.8 or later and Python 3.7.2 or later.
-We also provide a multilingual [demo script](https://github.com/stanfordnlp/stanfordnlp/blob/master/demo/pipeline_demo.py) that demonstrates how one uses StanfordNLP in other languages than English, for example Chinese (traditional)
+We also provide a multilingual [demo script](https://github.com/stanfordnlp/stanza/blob/master/demo/pipeline_demo.py) that demonstrates how one uses Stanza in other languages than English, for example Chinese (traditional)
```bash
python demo/pipeline_demo.py -l zh
```
-See [our getting started guide](https://stanfordnlp.github.io/stanfordnlp/installation_usage.html#getting-started) for more details.
+See [our getting started guide](https://stanfordnlp.github.io/stanza/installation_usage.html#getting-started) for more details.
### Access to Java Stanford CoreNLP Server
@@ -101,7 +101,7 @@ There are a few initial setup steps.
* Put the model jars in the distribution folder
* Tell the python code where Stanford CoreNLP is located: `export CORENLP_HOME=/path/to/stanford-corenlp-full-2018-10-05`
-We provide another [demo script](https://github.com/stanfordnlp/stanfordnlp/blob/master/demo/corenlp.py) that shows how one can use the CoreNLP client and extract various annotations from it.
+We provide another [demo script](https://github.com/stanfordnlp/stanza/blob/master/demo/corenlp.py) that shows how one can use the CoreNLP client and extract various annotations from it.
### Online Colab Notebooks
@@ -110,11 +110,11 @@ To get your started, we also provide interactive Jupyter notebooks in the `demo`
* Go to the [Google Colab website](https://colab.research.google.com)
* Navigate to `File` -> `Open notebook`, and choose `GitHub` in the pop-up menu
* Note that you do **not** need to give Colab access permission to your github account
-* Type `stanfordnlp/stanfordnlp` in the search bar, and click enter
+* Type `stanfordnlp/stanza` in the search bar, and click enter
### Trained Models for the Neural Pipeline
-We currently provide models for all of the treebanks in the CoNLL 2018 Shared Task. You can find instructions for downloading and using these models [here](https://stanfordnlp.github.io/stanfordnlp/models.html).
+We currently provide models for all of the treebanks in the CoNLL 2018 Shared Task. You can find instructions for downloading and using these models [here](https://stanfordnlp.github.io/stanza/models.html).
### Batching To Maximize Pipeline Speed
@@ -127,8 +127,8 @@ We are actively working on improving multi-document processing.
All neural modules in this library, including the tokenizer, the multi-word token (MWT) expander, the POS/morphological features tagger, the lemmatizer and the dependency parser, can be trained with your own [CoNLL-U](https://universaldependencies.org/format.html) format data. Currently, we do not support model training via the `Pipeline` interface. Therefore, to train your own models, you need to clone this git repository and set up from source.
-For detailed step-by-step guidance on how to train and evaluate your own models, please visit our [training documentation](https://stanfordnlp.github.io/stanfordnlp/training.html).
+For detailed step-by-step guidance on how to train and evaluate your own models, please visit our [training documentation](https://stanfordnlp.github.io/stanza/training.html).
## LICENSE
-StanfordNLP is released under the Apache License, Version 2.0. See the [LICENSE](https://github.com/stanfordnlp/stanfordnlp/blob/master/LICENSE) file for more details.
+Stanza is released under the Apache License, Version 2.0. See the [LICENSE](https://github.com/stanfordnlp/stanza/blob/master/LICENSE) file for more details.