Skip to content

Commit

Permalink
- Fixed image URL in Entity Linking doc (NVIDIA#5528)
Browse files Browse the repository at this point in the history
- Fixed title format in Question Answering doc

Signed-off-by: Vitaly Lavrukhin <[email protected]>

Signed-off-by: Vitaly Lavrukhin <[email protected]>
Co-authored-by: George <[email protected]>
Signed-off-by: andrusenkoau <[email protected]>
  • Loading branch information
2 people authored and andrusenkoau committed Jan 5, 2023
1 parent 23908d7 commit ccbd4dd
Show file tree
Hide file tree
Showing 2 changed files with 9 additions and 9 deletions.
2 changes: 1 addition & 1 deletion docs/source/nlp/entity_linking.rst
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,7 @@ The main idea behind this approach is to reshape an initial concept embedding sp
pulled closer together and unrelated concepts are pushed further apart. The concept embeddings from this reshaped space can then
be used to build a knowledge base embedding index.

.. image:: https://github.com/NVIDIA/NeMo/blob/entity-linking-documentation/docs/source/nlp/entity_linking_overview.jpg
.. image:: entity_linking_overview.jpg
:alt: Entity-Linking-Overview
:width: 800px

Expand Down
16 changes: 8 additions & 8 deletions docs/source/nlp/question_answering.rst
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@ For every word in our training dataset the model predicts:
Given a question and a context, both in natural language, generate an answer for the question. Unlike the BERT-like models, there is no constraint that the answer should be a span within the context.

Supported Tasks
===============
---------------

+----------------------------------+-----------------+----------------------------------------------------------------------+------------------------------------------+
| **Task** | **Models** | **Supported Options for model.language_model.pretrained_model_name** | **Supported options for model.library** |
Expand All @@ -35,7 +35,7 @@ Supported Tasks
+----------------------------------+-----------------+----------------------------------------------------------------------+------------------------------------------+

Available models
^^^^^^^^^^^^^^^^
----------------

Following BERT-like models are available for Extractive Question-Answering

Expand Down Expand Up @@ -63,7 +63,7 @@ Following BERT-like models are available for Extractive Question-Answering
- https://ngc.nvidia.com/catalog/models/nvidia:nemo:qa_squadv2_0_megatron_uncased

Module Design
=============
-------------

The module is decouple data and model components to support idependent integration of various model achitectures and datasets.
QAProcessor, QAExample, and the base QADataset modules are responsible for model-independent data handling utilites like loading SQuAD format dataset files and parsing examples.
Expand All @@ -75,7 +75,7 @@ Similarly, the BaseQAModel module handles common model tasks like creating datal
:width: 800px

Configuration
=============
-------------

The default sample model training configuration can be found at: `NeMo/examples/nlp/question_answering/conf/qa_conf.yaml`

Expand Down Expand Up @@ -108,7 +108,7 @@ Arguments that very commonly need to be edited for all models and datasets
- :code:`model.tokens_to_generate`: maximum answer tokens to be generated for the generative models

Data Format
===========
-----------

The QA models expect datasets to be present in the SQuAD format. For using datasets other than the standard SQuAD v1.1 and v2.0, the datasets should be first converted into the SQuAD format.

Expand Down Expand Up @@ -154,7 +154,7 @@ The following is an example of the expected SQuAD data format (JSON file):
For datasets of generative nature where the answer might not be an exact span within the context, the :code:`answer_start` field can be set to -1.

Downloading Datasets
====================
--------------------

Following sections describes how to download the SQuAD datasets, along with an example of converting a non-SQuAD dataset (MS-MARCO) into the SQuAD format for the QA models.

Expand Down Expand Up @@ -232,7 +232,7 @@ The conversion to SQuAD format can be performed using the following script:
- setting :code:`keep_only_relevant_passages` to ``True`` will exclude passages that have ``is_selected=0`` in the MS-MARCO dataset

Training, Validation, Testing
=============================
-----------------------------

A step-by-step guide to training and testing QA models, as well as running inference can be found at `NeMo/tutorials/nlp/Question_Answering.ipynb`. Following is an example of training a QA model using the example script provided at `NeMo/examples/nlp/question_answering/question_answering.py`:

Expand Down Expand Up @@ -283,4 +283,4 @@ During evaluation of the :code:`validation_ds` and :code:`test_ds`, the script g
- :code:`Exact Match (EM)`
- :code:`F1`

More details about these metrics can be found `here <https://en.wikipedia.org/wiki/F-score>`__.
More details about these metrics can be found `here <https://en.wikipedia.org/wiki/F-score>`__.

0 comments on commit ccbd4dd

Please sign in to comment.